5959 with :
6060 halt-dispatch-input : ${{ inputs.halt-for-connection }}
6161
62-
63- - name : Build hlo_runner_main_gpu
64- working-directory : xla
65- run : bazel build -c opt --config=cuda --dynamic_mode=off //xla/tools/multihost_hlo_runner:hlo_runner_main_gpu
6662 - name : Build hlo_runner_main
6763 working-directory : xla
6864 run : bazel build -c opt --config=cuda --dynamic_mode=off //xla/tools/multihost_hlo_runner:hlo_runner_main
69-
70- - name : Build test_gpu_profiler
71- working-directory : xla
72- run : bazel build -c opt --config=cuda --dynamic_mode=off //xla/tools/multihost_hlo_runner:test_gpu_profiler
7365
7466 # - name: Wait For Connection
7567 # uses: google-ml-infra/actions/ci_connection@main
@@ -88,44 +80,23 @@ jobs:
8880 }
8981 EOF
9082
91- - name : Wait For Connection
92- uses : google-ml-infra/actions/ci_connection@main
93- with :
94- halt-dispatch-input : ${{ inputs.halt-for-connection }}
95-
96- # - name: Run specific HLO file
97- # working-directory: xla
98- # run: |
99- # nvidia-smi --query-gpu=utilization.gpu --format=csv -l 1 > gpu_utilization.log & bazel run -c opt --config=cuda --dynamic_mode=off //xla/tools/multihost_hlo_runner:hlo_runner_main_gpu -- --device_type=gpu --log_output=True --use_spmd_partitioning isolated_convolution.hlo &> results/isolated_convolution.log
100- # - name: Wait For Connection
101- # uses: google-ml-infra/actions/ci_connection@main
102- # with:
103- # halt-dispatch-input: ${{ inputs.halt-for-connection }}
104- - name : Run specific HLO file hlo_runner_main_gpu
105- working-directory : xla
106- run : |
107- nvidia-smi --query-gpu=utilization.gpu --format=csv -l 1 > results.gpu_utilization.log & bazel run -c opt --config=cuda --dynamic_mode=off //xla/tools/multihost_hlo_runner:hlo_runner_main_gpu -- --device_type=gpu --log_output=True --use_spmd_partitioning isolated_convolution.hlo &> results/hlo_runner_main_gpu_isolated_convolution.log
108-
10983 - name : Run specific HLO file hlo_runner_main
11084 working-directory : xla
11185 run : |
112- nvidia-smi --query-gpu=utilization.gpu --format=csv -l 1 > results/gpu_utilization_v2.log & bazel run -c opt --config=cuda --dynamic_mode=off // xla/tools/multihost_hlo_runner: hlo_runner_main -- -- device_type=gpu --log_output=True --use_spmd_partitioning isolated_convolution.hlo &> results/hlo_runner_main_isolated_convolution.log
86+ nvidia-smi --query-gpu=utilization.gpu --format=csv -l 1 > results/gpu_utilization_v2.log & ./ bazel-bin/ xla/tools/multihost_hlo_runner/ hlo_runner_main --device_type=gpu --use_spmd_partitioning isolated_convolution.hlo &> results/hlo_runner_main_isolated_convolution.log
11387
114- - name : Run test_gpu_profiler
115- working-directory : xla
116- run : |
117- nvidia-smi --query-gpu=utilization.gpu --format=csv -l 1 > results/gpu_utilization_profiler.log & ./bazel-bin/xla/tools/multihost_hlo_runner/test_gpu_profiler &> results/test_gpu_profiler.log
11888 - name : Wait For Connection
11989 uses : google-ml-infra/actions/ci_connection@main
12090 with :
12191 halt-dispatch-input : ${{ inputs.halt-for-connection }}
92+
12293 - name : Download parse_xla_logs.py
12394 working-directory : xla
12495 run : wget https://raw.githubusercontent.com/juliagmt-google/xla/main/.github/workflows/parse_xla_logs.py
12596
12697 - name : Parse XLA logs
12798 working-directory : xla
128- run : python parse_xla_logs.py results/isolated_convolution.hlo .log
99+ run : python parse_xla_logs.py results/hlo_runner_main_isolated_convolution .log
129100
130101 - name : Upload Results
131102 uses : actions/upload-artifact@v4
0 commit comments