@@ -33,17 +33,11 @@ jobs:
3333 with :
3434 repository : juliagmt-google/xla
3535 path : xla
36-
37- - name : Create results directory
38- working-directory : xla
39- run : mkdir -p results
40- - name : Install Kaggle API client
41- run : pip install kaggle
4236
43- # - name: Get GPU spec
44- # working-directory: xla
45- # continue-on-error: true
46- # run: nvidia-smi
37+ - name : Get GPU spec
38+ working-directory : xla
39+ continue-on-error : true
40+ run : nvidia-smi
4741
4842 - name : Configure XLA
4943 working-directory : xla
@@ -65,63 +59,63 @@ jobs:
6559 with :
6660 halt-dispatch-input : ${{ inputs.halt-for-connection }}
6761
68- - name : Build hlo_runner_main
69- working-directory : xla
70- run : bazel build -c opt --config=cuda --dynamic_mode=off //xla/tools/multihost_hlo_runner:hlo_runner_main
62+ # - name: Build hlo_runner_main
63+ # working-directory: xla
64+ # run: bazel build -c opt --config=cuda --dynamic_mode=off //xla/tools/multihost_hlo_runner:hlo_runner_main
7165
7266 # - name: Wait For Connection
7367 # uses: google-ml-infra/actions/ci_connection@main
7468 # with:
7569 # halt-dispatch-input: ${{ inputs.halt-for-connection }}
7670
77- - name : Create gpu_hlo_backend.hlo
78- working-directory : xla
79- run : |
80- cat << EOF > gpu_hlo_backend.hlo
81- HloModule module
82- // CHECK: is_scheduled=true
71+ # - name: Create gpu_hlo_backend.hlo
72+ # working-directory: xla
73+ # run: |
74+ # cat << EOF > gpu_hlo_backend.hlo
75+ # HloModule module
76+ # // CHECK: is_scheduled=true
8377
84- ENTRY computation {
85- p = f32[5000,6000]{1,0} parameter(0)
86- e = f32[5000,6000]{1,0} sqrt(p)
87- c = f32[6000,5000] transpose(p), dimensions={1,0}
88- r = f32[300,20,5000] reshape(c)
89- ROOT out = (f32[5000,6000], f32[300,20,5000]) tuple(e,r)
90- }
91- EOF
78+ # ENTRY computation {
79+ # p = f32[5000,6000]{1,0} parameter(0)
80+ # e = f32[5000,6000]{1,0} sqrt(p)
81+ # c = f32[6000,5000] transpose(p), dimensions={1,0}
82+ # r = f32[300,20,5000] reshape(c)
83+ # ROOT out = (f32[5000,6000], f32[300,20,5000]) tuple(e,r)
84+ # }
85+ # EOF
9286
9387 # - name: Wait For Connection
9488 # uses: google-ml-infra/actions/ci_connection@main
9589 # with:
9690 # halt-dispatch-input: ${{ inputs.halt-for-connection }}
9791
98- - name : Run an HLO file
99- working-directory : xla
100- run : |
101- nvidia-smi --query-gpu=utilization.gpu --format=csv -l 1 > results/gpu_utilization_gpu_backend.log & ./bazel-bin/xla/tools/multihost_hlo_runner/hlo_runner_main --device_type=gpu --log_output=True --use_spmd_partitioning gpu_hlo_backend.hlo
92+ # - name: Run an HLO file
93+ # working-directory: xla
94+ # run: |
95+ # nvidia-smi --query-gpu=utilization.gpu --format=csv -l 1 > results/gpu_utilization_gpu_backend.log & ./bazel-bin/xla/tools/multihost_hlo_runner/hlo_runner_main --device_type=gpu --log_output=True --use_spmd_partitioning gpu_hlo_backend.hlo
10296
103- cd results
104- ls
105- cat gpu_hlo_backend.hlo.log
97+ # cd results
98+ # ls
99+ # cat gpu_hlo_backend.hlo.log
106100
107- - name : Wait For Connection
108- uses : google-ml-infra/actions/ci_connection@main
109- with :
110- halt-dispatch-input : ${{ inputs.halt-for-connection }}
101+ # - name: Wait For Connection
102+ # uses: google-ml-infra/actions/ci_connection@main
103+ # with:
104+ # halt-dispatch-input: ${{ inputs.halt-for-connection }}
111105
112- - name : Download parse_xla_logs.py
113- working-directory : xla
114- run : wget https://raw.githubusercontent.com/juliagmt-google/xla/main/.github/workflows/parse_xla_logs.py
106+ # - name: Download parse_xla_logs.py
107+ # working-directory: xla
108+ # run: wget https://raw.githubusercontent.com/juliagmt-google/xla/main/.github/workflows/parse_xla_logs.py
115109
116- - name : Parse XLA logs
117- working-directory : xla
118- run : python parse_xla_logs.py results/gpu_hlo_backend.hlo.log
110+ # - name: Parse XLA logs
111+ # working-directory: xla
112+ # run: python parse_xla_logs.py results/gpu_hlo_backend.hlo.log
119113
120- - name : Upload Results
121- uses : actions/upload-artifact@v4
122- with :
123- name : gpu-xla-benchmarks
124- path : xla/results
114+ # - name: Upload Results
115+ # uses: actions/upload-artifact@v4
116+ # with:
117+ # name: gpu-xla-benchmarks
118+ # path: xla/results
125119 # # jax-build-and-test:
126120 # # runs-on: linux-x86-g2-48-l4-4gpu # Use a GPU-enabled runner
127121 # # container:
0 commit comments