Skip to content

Generate GPU L2 size inputs in flash_attention Triton Bench #2671

Generate GPU L2 size inputs in flash_attention Triton Bench

Generate GPU L2 size inputs in flash_attention Triton Bench #2671

Workflow file for this run

name: Linter
on:
pull_request:
push:
branches:
- main
workflow_dispatch:
jobs:
pylint:
permissions:
contents: read
runs-on: ubuntu-latest
steps:
- name: Checkout
uses: actions/checkout@v3
with:
path: tritonbench
- name: Install deps
run: |
cd tritonbench
pip install -r requirements-fmt.txt
- name: Check Formatting
uses: omnilib/ufmt@action-v1
with:
path: tritonbench
concurrency:
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.sha }}-${{ github.event_name == 'workflow_dispatch' }}
cancel-in-progress: true