6767
6868 - name : Run run.sh for E2E benchmarks flax_2b (within venv)
6969 working-directory : xla/xla/backends/cpu/benchmarks/e2e/gemma2/flax_2b
70+ timeout-minutes : 30 # Set the timeout to 30 minutes
7071 run : |
7172 bash run.sh
7273
@@ -97,26 +98,26 @@ jobs:
9798 # with:
9899 # halt-dispatch-input: ${{ inputs.halt-for-connection }}
99100
100- - name : Create gpu_hlo_backend.hlo
101- working-directory : xla
102- run : |
103- cat << EOF > gpu_hlo_backend.hlo
104- HloModule module
105- // CHECK: is_scheduled=true
101+ # - name: Create gpu_hlo_backend.hlo
102+ # working-directory: xla
103+ # run: |
104+ # cat << EOF > gpu_hlo_backend.hlo
105+ # HloModule module
106+ # // CHECK: is_scheduled=true
106107
107- ENTRY computation {
108- p = f32[5000,6000]{1,0} parameter(0)
109- e = f32[5000,6000]{1,0} sqrt(p)
110- c = f32[6000,5000] transpose(p), dimensions={1,0}
111- r = f32[300,20,5000] reshape(c)
112- ROOT out = (f32[5000,6000], f32[300,20,5000]) tuple(e,r)
113- }
114- EOF
108+ # ENTRY computation {
109+ # p = f32[5000,6000]{1,0} parameter(0)
110+ # e = f32[5000,6000]{1,0} sqrt(p)
111+ # c = f32[6000,5000] transpose(p), dimensions={1,0}
112+ # r = f32[300,20,5000] reshape(c)
113+ # ROOT out = (f32[5000,6000], f32[300,20,5000]) tuple(e,r)
114+ # }
115+ # EOF
115116
116- - name : Wait For Connection
117- uses : google-ml-infra/actions/ci_connection@main
118- with :
119- halt-dispatch-input : ${{ inputs.halt-for-connection }}
117+ # - name: Wait For Connection
118+ # uses: google-ml-infra/actions/ci_connection@main
119+ # with:
120+ # halt-dispatch-input: ${{ inputs.halt-for-connection }}
120121
121122 # - name: Run an HLO file
122123 # working-directory: xla
@@ -128,19 +129,19 @@ jobs:
128129 # with:
129130 # halt-dispatch-input: ${{ inputs.halt-for-connection }}
130131
131- - name : Download parse_xla_logs.py
132- working-directory : xla
133- run : wget https://raw.githubusercontent.com/juliagmt-google/xla/main/.github/workflows/parse_xla_logs.py
132+ # - name: Download parse_xla_logs.py
133+ # working-directory: xla
134+ # run: wget https://raw.githubusercontent.com/juliagmt-google/xla/main/.github/workflows/parse_xla_logs.py
134135
135- - name : Parse XLA logs
136- working-directory : xla
137- run : python parse_xla_logs.py results/gpu_hlo_backend.log
136+ # - name: Parse XLA logs
137+ # working-directory: xla
138+ # run: python parse_xla_logs.py results/gpu_hlo_backend.log
138139
139- - name : Upload Results
140- uses : actions/upload-artifact@v4
141- with :
142- name : gpu-xla-benchmarks
143- path : xla/results
140+ # - name: Upload Results
141+ # uses: actions/upload-artifact@v4
142+ # with:
143+ # name: gpu-xla-benchmarks
144+ # path: xla/results
144145 # # jax-build-and-test:
145146 # # runs-on: linux-x86-g2-48-l4-4gpu # Use a GPU-enabled runner
146147 # # container:
0 commit comments