Skip to content

Generate GPU L2 size inputs in flash_attention Triton Bench #2580

Generate GPU L2 size inputs in flash_attention Triton Bench

Generate GPU L2 size inputs in flash_attention Triton Bench #2580

Workflow file for this run

name: TritonBench PR Test
on:
pull_request:
paths:
- .github/workflows/pr.yml
- .github/workflows/_linux-test-*.yml
- tritonbench/**
push:
branches:
- main
jobs:
h100-pytorch-test:
uses: ./.github/workflows/_linux-test-h100.yml
with:
conda_env: "pytorch"
h100-triton-main-test:
uses: ./.github/workflows/_linux-test-h100.yml
with:
conda_env: "triton-main"
mi350-triton-main-test:
uses: ./.github/workflows/_linux-test-mi350.yml
with:
conda_env: "triton-main"
concurrency:
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.sha }}-${{ github.event_name == 'workflow_dispatch' }}
cancel-in-progress: true