Skip to content

Commit a379380

Browse files
authored
drop cuda11 support (#959)
fix #957 - Updated default cuda version to 12.9.1 - TODO: `notebooks/databricks/init-pip-cuda-11.8.sh` --------- Signed-off-by: Yanxuan Liu <[email protected]>
1 parent 096f5d0 commit a379380

File tree

14 files changed

+31
-31
lines changed

14 files changed

+31
-31
lines changed

ci/Dockerfile

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@
1414
# limitations under the License.
1515
#
1616

17-
ARG CUDA_VERSION=11.8.0
17+
ARG CUDA_VERSION=12.0.1
1818
FROM nvidia/cuda:${CUDA_VERSION}-devel-ubuntu22.04
1919

2020
# ubuntu22
@@ -48,5 +48,5 @@ RUN wget --quiet https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86
4848

4949
# install cuML
5050
ARG CUML_VER=25.06
51-
RUN conda install -y -c rapidsai -c conda-forge -c nvidia cuml=$CUML_VER cuvs=$CUML_VER python=3.10 cuda-version=11.8 numpy~=1.0 \
51+
RUN conda install -y -c rapidsai -c conda-forge -c nvidia cuml=$CUML_VER cuvs=$CUML_VER python=3.10 cuda-version=12.0 numpy~=1.0 \
5252
&& conda clean --all -f -y

ci/Jenkinsfile.premerge

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ import ipp.blossom.*
3030

3131
def githubHelper // blossom github helper
3232
def TEMP_IMAGE_BUILD = true
33-
def IMAGE_PREMERGE = "${common.ARTIFACTORY_NAME}/sw-spark-docker/rapids:ml-ubuntu22-cuda11.8.0-py310"
33+
def IMAGE_PREMERGE = "${common.ARTIFACTORY_NAME}/sw-spark-docker/rapids:ml-ubuntu22-cuda12.0.1-py310"
3434
def cpuImage = pod.getCPUYAML("${common.ARTIFACTORY_NAME}/sw-spark-docker/spark:rapids-databricks") // tooling image
3535
def PREMERGE_DOCKERFILE = 'ci/Dockerfile'
3636
def PREMERGE_TAG

docker/Dockerfile

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@
3535
#
3636
###
3737

38-
ARG CUDA_VERSION=11.5.2
38+
ARG CUDA_VERSION=12.0.1
3939
FROM nvidia/cuda:${CUDA_VERSION}-devel-ubuntu20.04
4040

4141
# ubuntu22

docker/Dockerfile.pip

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@
1414
# limitations under the License.
1515
#
1616

17-
ARG CUDA_VERSION=11.8.0
17+
ARG CUDA_VERSION=12.0.1
1818
FROM nvidia/cuda:${CUDA_VERSION}-devel-ubuntu22.04
1919

2020
ARG PYSPARK_VERSION=3.3.1

docker/Dockerfile.python

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@
1414
# limitations under the License.
1515
#
1616

17-
ARG CUDA_VERSION=11.8.0
17+
ARG CUDA_VERSION=12.0.1
1818
FROM nvidia/cuda:${CUDA_VERSION}-devel-ubuntu22.04
1919

2020
ARG CUML_VERSION=25.06
@@ -47,7 +47,7 @@ RUN wget --quiet https://repo.anaconda.com/miniconda/Miniconda3-py38_4.10.3-Linu
4747

4848
# install cuML
4949

50-
RUN conda install -y -c rapidsai -c conda-forge -c nvidia python=3.10 cuda-version=11.8 cuml=$CUML_VERSION numpy~=1.0 \
50+
RUN conda install -y -c rapidsai -c conda-forge -c nvidia python=3.10 cuda-version=12.0 cuml=$CUML_VERSION numpy~=1.0 \
5151
&& conda clean --all -f -y
5252

5353
# install python dependencies

docs/site/compatibility.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ The following table shows the currently supported algorithms. The goal is to ex
3131

3232
| Spark Rapids ML | CUDA | Spark | Python |
3333
| :-------------- | :---- | :----- | :----- |
34-
| 1.0.0 | 11.4+ | 3.3+ | 3.10+ |
34+
| 1.0.0 | 12.0+ | 3.3+ | 3.10+ |
3535

3636

3737
## Single vs Double precision inputs

notebooks/databricks/README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -11,16 +11,16 @@ If you already have a Databricks account, you can run the example notebooks on a
1111
```bash
1212
export WS_SAVE_DIR="/path/to/directory/in/workspace"
1313
databricks workspace mkdirs ${WS_SAVE_DIR} --profile ${PROFILE}
14-
databricks workspace import --format AUTO --file init-pip-cuda-11.8.sh ${WS_SAVE_DIR}/init-pip-cuda-11.8.sh --profile ${PROFILE}
14+
databricks workspace import --format AUTO --file init-pip-cuda-12.0.sh ${WS_SAVE_DIR}/init-pip-cuda-12.0.sh --profile ${PROFILE}
1515
```
1616
**Note**: the init script does the following on each Spark node:
17-
- updates the CUDA runtime to 11.8 (required for Spark Rapids ML dependencies).
17+
- updates the CUDA runtime to 12.0 (required for Spark Rapids ML dependencies).
1818
- downloads and installs the [Spark-Rapids](https://github.com/NVIDIA/spark-rapids) plugin for accelerating data loading and Spark SQL.
1919
- installs various `cuXX` dependencies via pip.
2020
- if the cluster environment variable `SPARK_RAPIDS_ML_NO_IMPORT_ENABLED=1` is define (see below), the init script also modifies a Databricks notebook kernel startup script to enable no-import change UX for the cluster. See [no-import-change](../README.md#no-import-change).
2121
- Create a cluster using **Databricks 13.3 LTS ML GPU Runtime** using at least two single-gpu workers and add the following configurations to the **Advanced options**.
2222
- **Init Scripts**
23-
- add the workspace path to the uploaded init script `${WS_SAVE_DIR}/init-pip-cuda-11.8.sh` as set above (but substitute variables manually in the form).
23+
- add the workspace path to the uploaded init script `${WS_SAVE_DIR}/init-pip-cuda-12.0.sh` as set above (but substitute variables manually in the form).
2424
- **Spark**
2525
- **Spark config**
2626
```

notebooks/databricks/init-pip-cuda-11.8.sh renamed to notebooks/databricks/init-pip-cuda-12.0.sh

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -21,24 +21,24 @@ set -ex
2121
RAPIDS_VERSION=25.6.0
2222
SPARK_RAPIDS_VERSION=25.04.0
2323

24-
curl -L https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/${SPARK_RAPIDS_VERSION}/rapids-4-spark_2.12-${SPARK_RAPIDS_VERSION}-cuda11.jar -o /databricks/jars/rapids-4-spark_2.12-${SPARK_RAPIDS_VERSION}.jar
24+
curl -L https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/${SPARK_RAPIDS_VERSION}/rapids-4-spark_2.12-${SPARK_RAPIDS_VERSION}-cuda12.jar -o /databricks/jars/rapids-4-spark_2.12-${SPARK_RAPIDS_VERSION}.jar
2525

26-
# install cudatoolkit 11.8 via runfile approach
27-
wget https://developer.download.nvidia.com/compute/cuda/11.8.0/local_installers/cuda_11.8.0_520.61.05_linux.run
28-
sh cuda_11.8.0_520.61.05_linux.run --silent --toolkit
26+
# install cudatoolkit 12.0 via runfile approach
27+
wget https://developer.download.nvidia.com/compute/cuda/12.0.1/local_installers/cuda_12.0.1_525.85.12_linux.run
28+
sh cuda_12.0.1_525.85.12_linux.run --silent --toolkit
2929

3030
# reset symlink and update library loading paths
3131
rm /usr/local/cuda
32-
ln -s /usr/local/cuda-11.8 /usr/local/cuda
32+
ln -s /usr/local/cuda-12.0 /usr/local/cuda
3333

3434
# upgrade pip
3535
/databricks/python/bin/pip install --upgrade pip
3636

3737
# install cudf, cuml and their rapids dependencies
3838
# using ~= pulls in latest micro version patches
39-
/databricks/python/bin/pip install cudf-cu11~=${RAPIDS_VERSION} \
40-
cuml-cu11~=${RAPIDS_VERSION} \
41-
cuvs-cu11~=${RAPIDS_VERSION} \
39+
/databricks/python/bin/pip install cudf-cu12~=${RAPIDS_VERSION} \
40+
cuml-cu12~=${RAPIDS_VERSION} \
41+
cuvs-cu12~=${RAPIDS_VERSION} \
4242
--extra-index-url=https://pypi.nvidia.com
4343

4444
# install spark-rapids-ml

python/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -18,11 +18,11 @@ This PySpark-compatible API leverages the RAPIDS cuML python API to provide GPU-
1818

1919
For simplicity, the following instructions just use Spark local mode, assuming a server with at least one GPU.
2020

21-
First, install RAPIDS cuML per [these instructions](https://rapids.ai/start.html). Example for CUDA Toolkit 11.8:
21+
First, install RAPIDS cuML per [these instructions](https://rapids.ai/start.html). Example for CUDA Toolkit 12.0:
2222
```bash
2323
conda create -n rapids-25.06 \
2424
-c rapidsai -c conda-forge -c nvidia \
25-
cuml=25.06 cuvs=25.06 python=3.10 cuda-version=11.8 numpy~=1.0
25+
cuml=25.06 cuvs=25.06 python=3.10 cuda-version=12.0 numpy~=1.0
2626
```
2727

2828
**Note**: while testing, we recommend using conda or docker to simplify installation and isolate your environment while experimenting. Once you have a working environment, you can then try installing directly, if necessary.

python/benchmark/databricks/gpu_cluster_spec.sh

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
# Copyright (c) 2024, NVIDIA CORPORATION.
1+
# Copyright (c) 2024-2025, NVIDIA CORPORATION.
22
#
33
# Licensed under the Apache License, Version 2.0 (the "License");
44
# you may not use this file except in compliance with the License.
@@ -54,7 +54,7 @@ cat <<EOF
5454
"init_scripts": [
5555
{
5656
"workspace": {
57-
"destination": "${INIT_SCRIPT_DIR}/init-pip-cuda-11.8.sh"
57+
"destination": "${INIT_SCRIPT_DIR}/init-pip-cuda-12.0.sh"
5858
}
5959
}
6060
],

0 commit comments

Comments
 (0)