Skip to content

Commit a457e16

Browse files
Update benchmarks.yml
1 parent 278a117 commit a457e16

File tree

1 file changed

+30
-29
lines changed

1 file changed

+30
-29
lines changed

.github/workflows/benchmarks.yml

Lines changed: 30 additions & 29 deletions
Original file line numberDiff line numberDiff line change
@@ -67,6 +67,7 @@ jobs:
6767
6868
- name: Run run.sh for E2E benchmarks flax_2b (within venv)
6969
working-directory: xla/xla/backends/cpu/benchmarks/e2e/gemma2/flax_2b
70+
timeout-minutes: 30 # Set the timeout to 30 minutes
7071
run: |
7172
bash run.sh
7273
@@ -97,26 +98,26 @@ jobs:
9798
# with:
9899
# halt-dispatch-input: ${{ inputs.halt-for-connection }}
99100

100-
- name: Create gpu_hlo_backend.hlo
101-
working-directory: xla
102-
run: |
103-
cat << EOF > gpu_hlo_backend.hlo
104-
HloModule module
105-
// CHECK: is_scheduled=true
101+
# - name: Create gpu_hlo_backend.hlo
102+
# working-directory: xla
103+
# run: |
104+
# cat << EOF > gpu_hlo_backend.hlo
105+
# HloModule module
106+
# // CHECK: is_scheduled=true
106107

107-
ENTRY computation {
108-
p = f32[5000,6000]{1,0} parameter(0)
109-
e = f32[5000,6000]{1,0} sqrt(p)
110-
c = f32[6000,5000] transpose(p), dimensions={1,0}
111-
r = f32[300,20,5000] reshape(c)
112-
ROOT out = (f32[5000,6000], f32[300,20,5000]) tuple(e,r)
113-
}
114-
EOF
108+
# ENTRY computation {
109+
# p = f32[5000,6000]{1,0} parameter(0)
110+
# e = f32[5000,6000]{1,0} sqrt(p)
111+
# c = f32[6000,5000] transpose(p), dimensions={1,0}
112+
# r = f32[300,20,5000] reshape(c)
113+
# ROOT out = (f32[5000,6000], f32[300,20,5000]) tuple(e,r)
114+
# }
115+
# EOF
115116

116-
- name: Wait For Connection
117-
uses: google-ml-infra/actions/ci_connection@main
118-
with:
119-
halt-dispatch-input: ${{ inputs.halt-for-connection }}
117+
# - name: Wait For Connection
118+
# uses: google-ml-infra/actions/ci_connection@main
119+
# with:
120+
# halt-dispatch-input: ${{ inputs.halt-for-connection }}
120121

121122
# - name: Run an HLO file
122123
# working-directory: xla
@@ -128,19 +129,19 @@ jobs:
128129
# with:
129130
# halt-dispatch-input: ${{ inputs.halt-for-connection }}
130131

131-
- name: Download parse_xla_logs.py
132-
working-directory: xla
133-
run: wget https://raw.githubusercontent.com/juliagmt-google/xla/main/.github/workflows/parse_xla_logs.py
132+
# - name: Download parse_xla_logs.py
133+
# working-directory: xla
134+
# run: wget https://raw.githubusercontent.com/juliagmt-google/xla/main/.github/workflows/parse_xla_logs.py
134135

135-
- name: Parse XLA logs
136-
working-directory: xla
137-
run: python parse_xla_logs.py results/gpu_hlo_backend.log
136+
# - name: Parse XLA logs
137+
# working-directory: xla
138+
# run: python parse_xla_logs.py results/gpu_hlo_backend.log
138139

139-
- name: Upload Results
140-
uses: actions/upload-artifact@v4
141-
with:
142-
name: gpu-xla-benchmarks
143-
path: xla/results
140+
# - name: Upload Results
141+
# uses: actions/upload-artifact@v4
142+
# with:
143+
# name: gpu-xla-benchmarks
144+
# path: xla/results
144145
# # jax-build-and-test:
145146
# # runs-on: linux-x86-g2-48-l4-4gpu # Use a GPU-enabled runner
146147
# # container:

0 commit comments

Comments
 (0)