Skip to content

[decoding_attention] Fix broken flash_attention and xformers #296

[decoding_attention] Fix broken flash_attention and xformers

[decoding_attention] Fix broken flash_attention and xformers #296

Workflow file for this run

name: Linter
on:
pull_request:
push:
branches:
- main
workflow_dispatch:
jobs:
pylint:
permissions:
contents: read
runs-on: ubuntu-latest
steps:
- name: Checkout
uses: actions/checkout@v3
with:
path: tritonbench
- name: Install deps
run: |
pip install ruff-api==0.1.0
- name: Check Formatting
uses: omnilib/ufmt@action-v1
with:
path: tritonbench
concurrency:
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.sha }}-${{ github.event_name == 'workflow_dispatch' }}
cancel-in-progress: true