@@ -71,39 +71,40 @@ jobs:
7171 with :
7272 repository : openxla/xla # Replace with your fork if needed
7373 path : xla
74-
75- - name : Configure XLA
76- working-directory : xla
77- run : ./configure.py --backend CUDA --nccl
74+
75+ # - name: Build XLA with GPU support with RBE
76+ # working-directory: xla
77+ # continue-on-error: true
78+ # run: bazel build --build_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only --test_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only,requires-gpu-sm75-only,requires-gpu-sm60,requires-gpu-sm70,-requires-gpu-sm80,-requires-gpu-sm80-only,-requires-gpu-sm90,-requires-gpu-sm90-only,-requires-gpu-sm100,-requires-gpu-sm100-only,-requires-gpu-amd --config=warnings --config=rbe_linux_cuda_nvcc --run_under=//tools/ci_build/gpu_build:parallel_gpu_execute --repo_env=TF_CUDA_COMPUTE_CAPABILITIES=7.5 --@cuda_driver//:enable_forward_compatibility=true --test_output=errors --verbose_failures --keep_going --nobuild_tests_only --profile=profile.json.gz --flaky_test_attempts=3 --jobs=150 --bes_upload_mode=fully_async --nobuild -- //xla/... //build_tools/... @tsl//tsl/...
79+
80+ # - name: Run XLA tests with GPU support with RBE
81+ # working-directory: xla
82+ # continue-on-error: true
83+ # run: bazel test --build_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only --test_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only,requires-gpu-sm75-only,requires-gpu-sm60,requires-gpu-sm70,-requires-gpu-sm80,-requires-gpu-sm80-only,-requires-gpu-sm90,-requires-gpu-sm90-only,-requires-gpu-sm100,-requires-gpu-sm100-only,-requires-gpu-amd --config=warnings --config=rbe_linux_cuda_nvcc --run_under=//tools/ci_build/gpu_build:parallel_gpu_execute --repo_env=TF_CUDA_COMPUTE_CAPABILITIES=7.5 --@cuda_driver//:enable_forward_compatibility=true --test_output=errors --verbose_failures --keep_going --nobuild_tests_only --profile=profile.json.gz --flaky_test_attempts=3 --jobs=150 --bes_upload_mode=fully_async -- //xla/... //build_tools/... @tsl//tsl/...
7884
79- - name : Set TF_CPP_MAX_VLOG_LEVEL
80- working-directory : xla
81- run : echo "TF_CPP_MAX_VLOG_LEVEL=1" >> $GITHUB_ENV
85+ # - name: Run Profile Analysis
86+ # working-directory: xla
87+ # continue-on-error: true
88+ # run: bazel analyze-profile profile.json.gz
8289
90+ - name : Get GPU spec
91+ working-directory : xla
92+ continue-on-error : true
93+ run : nvidia-smi
8394 - name : Wait For Connection
8495 uses : google-ml-infra/actions/ci_connection@main
8596 with :
8697 halt-dispatch-input : ${{ inputs.halt-for-connection }}
87-
88- - name : Build XLA with GPU support with RBE
89- working-directory : xla
90- continue-on-error : true
91- run : bazel build --build_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only --test_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only,requires-gpu-sm75-only,requires-gpu-sm60,requires-gpu-sm70,-requires-gpu-sm80,-requires-gpu-sm80-only,-requires-gpu-sm90,-requires-gpu-sm90-only,-requires-gpu-sm100,-requires-gpu-sm100-only,-requires-gpu-amd --config=warnings --config=rbe_linux_cuda_nvcc --run_under=//tools/ci_build/gpu_build:parallel_gpu_execute --repo_env=TF_CUDA_COMPUTE_CAPABILITIES=7.5 --@cuda_driver//:enable_forward_compatibility=true --test_output=errors --verbose_failures --keep_going --nobuild_tests_only --profile=profile.json.gz --flaky_test_attempts=3 --jobs=150 --bes_upload_mode=fully_async --nobuild -- //xla/... //build_tools/... @tsl//tsl/...
92-
93- - name : Run XLA tests with GPU support with RBE
98+
99+ - name : Configure XLA
94100 working-directory : xla
95- continue-on-error : true
96- run : bazel test --build_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only --test_tag_filters=-no_oss,requires-gpu-nvidia,gpu,-rocm-only,requires-gpu-sm75-only,requires-gpu-sm60,requires-gpu-sm70,-requires-gpu-sm80,-requires-gpu-sm80-only,-requires-gpu-sm90,-requires-gpu-sm90-only,-requires-gpu-sm100,-requires-gpu-sm100-only,-requires-gpu-amd --config=warnings --config=rbe_linux_cuda_nvcc --run_under=//tools/ci_build/gpu_build:parallel_gpu_execute --repo_env=TF_CUDA_COMPUTE_CAPABILITIES=7.5 --@cuda_driver//:enable_forward_compatibility=true --test_output=errors --verbose_failures --keep_going --nobuild_tests_only --profile=profile.json.gz --flaky_test_attempts=3 --jobs=150 --bes_upload_mode=fully_async -- //xla/... //build_tools/... @tsl//tsl/...
101+ run : ./configure.py --backend CUDA --nccl
97102
98- - name : Run Profile Analysis
103+ - name : Set TF_CPP_MAX_VLOG_LEVEL
99104 working-directory : xla
100- continue-on-error : true
101- run : bazel analyze-profile profile.json.gz
105+ run : echo "TF_CPP_MAX_VLOG_LEVEL=1" >> $GITHUB_ENV
102106
103- - name : Get GPU spec
104- working-directory : xla
105- continue-on-error : true
106- run : nvidia-smi
107+
107108
108109 - name : Build run_hlo_module with GPU
109110 working-directory : openxla
0 commit comments