Skip to content

Commit 341e2a1

Browse files
Update benchmarks.yml
1 parent 33ce83c commit 341e2a1

File tree

1 file changed

+24
-23
lines changed

1 file changed

+24
-23
lines changed

.github/workflows/benchmarks.yml

Lines changed: 24 additions & 23 deletions
Original file line numberDiff line numberDiff line change
@@ -71,39 +71,40 @@ jobs:
7171
with:
7272
repository: openxla/xla # Replace with your fork if needed
7373
path: xla
74-
75-
- name: Configure XLA
76-
working-directory: xla
77-
run: ./configure.py --backend CUDA --nccl
74+
75+
# - name: Build XLA with GPU support with RBE
76+
# working-directory: xla
77+
# continue-on-error: true
78+
# run: bazel build --build_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only --test_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only,requires-gpu-sm75-only,requires-gpu-sm60,requires-gpu-sm70,-requires-gpu-sm80,-requires-gpu-sm80-only,-requires-gpu-sm90,-requires-gpu-sm90-only,-requires-gpu-sm100,-requires-gpu-sm100-only,-requires-gpu-amd --config=warnings --config=rbe_linux_cuda_nvcc --run_under=//tools/ci_build/gpu_build:parallel_gpu_execute --repo_env=TF_CUDA_COMPUTE_CAPABILITIES=7.5 --@cuda_driver//:enable_forward_compatibility=true --test_output=errors --verbose_failures --keep_going --nobuild_tests_only --profile=profile.json.gz --flaky_test_attempts=3 --jobs=150 --bes_upload_mode=fully_async --nobuild -- //xla/... //build_tools/... @tsl//tsl/...
79+
80+
# - name: Run XLA tests with GPU support with RBE
81+
# working-directory: xla
82+
# continue-on-error: true
83+
# run: bazel test --build_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only --test_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only,requires-gpu-sm75-only,requires-gpu-sm60,requires-gpu-sm70,-requires-gpu-sm80,-requires-gpu-sm80-only,-requires-gpu-sm90,-requires-gpu-sm90-only,-requires-gpu-sm100,-requires-gpu-sm100-only,-requires-gpu-amd --config=warnings --config=rbe_linux_cuda_nvcc --run_under=//tools/ci_build/gpu_build:parallel_gpu_execute --repo_env=TF_CUDA_COMPUTE_CAPABILITIES=7.5 --@cuda_driver//:enable_forward_compatibility=true --test_output=errors --verbose_failures --keep_going --nobuild_tests_only --profile=profile.json.gz --flaky_test_attempts=3 --jobs=150 --bes_upload_mode=fully_async -- //xla/... //build_tools/... @tsl//tsl/...
7884

79-
- name: Set TF_CPP_MAX_VLOG_LEVEL
80-
working-directory: xla
81-
run: echo "TF_CPP_MAX_VLOG_LEVEL=1" >> $GITHUB_ENV
85+
# - name: Run Profile Analysis
86+
# working-directory: xla
87+
# continue-on-error: true
88+
# run: bazel analyze-profile profile.json.gz
8289

90+
- name: Get GPU spec
91+
working-directory: xla
92+
continue-on-error: true
93+
run: nvidia-smi
8394
- name: Wait For Connection
8495
uses: google-ml-infra/actions/ci_connection@main
8596
with:
8697
halt-dispatch-input: ${{ inputs.halt-for-connection }}
87-
88-
- name: Build XLA with GPU support with RBE
89-
working-directory: xla
90-
continue-on-error: true
91-
run: bazel build --build_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only --test_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only,requires-gpu-sm75-only,requires-gpu-sm60,requires-gpu-sm70,-requires-gpu-sm80,-requires-gpu-sm80-only,-requires-gpu-sm90,-requires-gpu-sm90-only,-requires-gpu-sm100,-requires-gpu-sm100-only,-requires-gpu-amd --config=warnings --config=rbe_linux_cuda_nvcc --run_under=//tools/ci_build/gpu_build:parallel_gpu_execute --repo_env=TF_CUDA_COMPUTE_CAPABILITIES=7.5 --@cuda_driver//:enable_forward_compatibility=true --test_output=errors --verbose_failures --keep_going --nobuild_tests_only --profile=profile.json.gz --flaky_test_attempts=3 --jobs=150 --bes_upload_mode=fully_async --nobuild -- //xla/... //build_tools/... @tsl//tsl/...
92-
93-
- name: Run XLA tests with GPU support with RBE
98+
99+
- name: Configure XLA
94100
working-directory: xla
95-
continue-on-error: true
96-
run: bazel test --build_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only --test_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only,requires-gpu-sm75-only,requires-gpu-sm60,requires-gpu-sm70,-requires-gpu-sm80,-requires-gpu-sm80-only,-requires-gpu-sm90,-requires-gpu-sm90-only,-requires-gpu-sm100,-requires-gpu-sm100-only,-requires-gpu-amd --config=warnings --config=rbe_linux_cuda_nvcc --run_under=//tools/ci_build/gpu_build:parallel_gpu_execute --repo_env=TF_CUDA_COMPUTE_CAPABILITIES=7.5 --@cuda_driver//:enable_forward_compatibility=true --test_output=errors --verbose_failures --keep_going --nobuild_tests_only --profile=profile.json.gz --flaky_test_attempts=3 --jobs=150 --bes_upload_mode=fully_async -- //xla/... //build_tools/... @tsl//tsl/...
101+
run: ./configure.py --backend CUDA --nccl
97102

98-
- name: Run Profile Analysis
103+
- name: Set TF_CPP_MAX_VLOG_LEVEL
99104
working-directory: xla
100-
continue-on-error: true
101-
run: bazel analyze-profile profile.json.gz
105+
run: echo "TF_CPP_MAX_VLOG_LEVEL=1" >> $GITHUB_ENV
102106

103-
- name: Get GPU spec
104-
working-directory: xla
105-
continue-on-error: true
106-
run: nvidia-smi
107+
107108

108109
- name: Build run_hlo_module with GPU
109110
working-directory: openxla

0 commit comments

Comments
 (0)