Skip to content

Commit c7becc2

Browse files
authored
Merge pull request #690 from NVIDIA/branch-24.06
release 24.06 [skip ci]
2 parents df01b39 + 20fc7b9 commit c7becc2

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

43 files changed

+1021
-249
lines changed

.github/workflows/auto-merge.yml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ name: auto-merge HEAD to BASE
1818
on:
1919
pull_request_target:
2020
branches:
21-
- branch-24.04
21+
- branch-24.06
2222
types: [closed]
2323

2424
jobs:
@@ -29,14 +29,14 @@ jobs:
2929
steps:
3030
- uses: actions/checkout@v4
3131
with:
32-
ref: branch-24.04 # force to fetch from latest upstream instead of PR ref
32+
ref: branch-24.06 # force to fetch from latest upstream instead of PR ref
3333

3434
- name: auto-merge job
3535
uses: ./.github/workflows/auto-merge
3636
env:
3737
OWNER: NVIDIA
3838
REPO_NAME: spark-rapids-ml
39-
HEAD: branch-24.04
40-
BASE: branch-24.06
39+
HEAD: branch-24.06
40+
BASE: branch-24.08
4141
AUTOMERGE_TOKEN: ${{ secrets.AUTOMERGE_TOKEN }} # use to merge PR
4242

.github/workflows/blossom-ci.yml

Lines changed: 17 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
# Copyright (c) 2023, NVIDIA CORPORATION.
1+
# Copyright (c) 2023-2024, NVIDIA CORPORATION.
22
#
33
# Licensed under the Apache License, Version 2.0 (the "License");
44
# you may not use this file except in compliance with the License.
@@ -34,17 +34,20 @@ jobs:
3434
args: ${{ env.args }}
3535

3636
# This job only runs for pull request comments
37-
if: contains( '\
38-
lijinf2,\
39-
eordentlich,\
40-
wbo4958,\
41-
leewyang,\
42-
rongou,\
43-
wjxiz1992,\
44-
GaryShen2008,\
45-
NvTimLiu,\
46-
YanxuanLiu,\
47-
', format('{0},', github.actor)) && github.event.comment.body == 'build'
37+
if: |
38+
github.event.comment.body == 'build' &&
39+
(
40+
github.actor == 'lijinf2' ||
41+
github.actor == 'eordentlich' ||
42+
github.actor == 'wbo4958' ||
43+
github.actor == 'leewyang' ||
44+
github.actor == 'rongou' ||
45+
github.actor == 'wjxiz1992' ||
46+
github.actor == 'GaryShen2008' ||
47+
github.actor == 'NvTimLiu' ||
48+
github.actor == 'YanxuanLiu' ||
49+
github.actor == 'pxLi'
50+
)
4851
steps:
4952
- name: Check if comment is issued by authorized person
5053
run: blossom-ci
@@ -59,15 +62,15 @@ jobs:
5962
runs-on: ubuntu-latest
6063
steps:
6164
- name: Checkout code
62-
uses: actions/checkout@v3
65+
uses: actions/checkout@v4
6366
with:
6467
repository: ${{ fromJson(needs.Authorization.outputs.args).repo }}
6568
ref: ${{ fromJson(needs.Authorization.outputs.args).ref }}
6669
lfs: 'true'
6770

6871
# repo specific steps
6972
- name: Setup java
70-
uses: actions/setup-java@v3
73+
uses: actions/setup-java@v4
7174
with:
7275
distribution: adopt
7376
java-version: 8

ci/Dockerfile

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,6 +37,6 @@ RUN wget --quiet https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86
3737
&& conda config --set solver libmamba
3838

3939
# install cuML
40-
ARG CUML_VER=24.04
40+
ARG CUML_VER=24.06
4141
RUN conda install -y -c rapidsai -c conda-forge -c nvidia cuml=$CUML_VER python=3.9 cuda-version=11.8 \
4242
&& conda clean --all -f -y

docker/Dockerfile.pip

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ ARG CUDA_VERSION=11.8.0
1818
FROM nvidia/cuda:${CUDA_VERSION}-devel-ubuntu22.04
1919

2020
ARG PYSPARK_VERSION=3.3.1
21-
ARG RAPIDS_VERSION=24.4.0
21+
ARG RAPIDS_VERSION=24.6.0
2222
ARG ARCH=amd64
2323
#ARG ARCH=arm64
2424
# Install packages to build spark-rapids-ml

docker/Dockerfile.python

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@
1717
ARG CUDA_VERSION=11.8.0
1818
FROM nvidia/cuda:${CUDA_VERSION}-devel-ubuntu20.04
1919

20-
ARG CUML_VERSION=24.04
20+
ARG CUML_VERSION=24.06
2121

2222
# Install packages to build spark-rapids-ml
2323
RUN apt update -y \

docs/site/compatibility.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,4 +34,4 @@ Note: Spark does not provide a k-Nearest Neighbors (k-NN) implementation, but it
3434
## Single vs Double precision inputs
3535
The underlying cuML implementations all accept single precision (e.g. Float or float32) input types and offer the best performance in this case. As a result, by default, Spark RAPIDs ML converts Spark DataFrames supplied to `fit` and `transform` methods having double precision data types (i.e. `VectorUDT`, `ArrayType(DoubleType())`, `DoubleType()` columns) to single precision before passing them down to the cuML layer. Most of the cuML algorithm implementations also support double precision inputs. The Estimator (for all algorithms) constructor parameter `float32_inputs` can be used to control this behavior. The default value is `True` which forces the conversion to single precision for all algorithms, but it can be set to `False` in which case double precision input data is passed to those cuML algorithms which support it.
3636

37-
Currently all algorithms *except* the following support double precision: LogisticRegression, k-NN, UMAP.
37+
Currently all algorithms *except* the following support double precision: k-NN, UMAP.

docs/source/conf.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@
99
project = 'spark-rapids-ml'
1010
copyright = '2024, NVIDIA'
1111
author = 'NVIDIA'
12-
release = '24.04.0'
12+
release = '24.06.0'
1313

1414
# -- General configuration ---------------------------------------------------
1515
# https://www.sphinx-doc.org/en/master/usage/configuration.html#general-configuration

jvm/README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ the _project root path_ with:
7474
cd jvm
7575
mvn clean package
7676
```
77-
Then `rapids-4-spark-ml_2.12-24.02.0-SNAPSHOT.jar` will be generated under `target` folder.
77+
Then `rapids-4-spark-ml_2.12-24.04.1-SNAPSHOT.jar` will be generated under `target` folder.
7878
7979
Users can also use the _release_ version spark-rapids plugin as the dependency if it's already been
8080
released in public maven repositories, see [rapids-4-spark maven repository](https://mvnrepository.com/artifact/com.nvidia/rapids-4-spark)
@@ -94,8 +94,8 @@ repository, usually in your `~/.m2/repository`.
9494
9595
Add the artifact jar to the Spark, for example:
9696
```bash
97-
ML_JAR="target/rapids-4-spark-ml_2.12-24.02.0-SNAPSHOT.jar"
98-
PLUGIN_JAR="~/.m2/repository/com/nvidia/rapids-4-spark_2.12/24.02.0/rapids-4-spark_2.12-24.02.0.jar"
97+
ML_JAR="target/rapids-4-spark-ml_2.12-24.04.1-SNAPSHOT.jar"
98+
PLUGIN_JAR="~/.m2/repository/com/nvidia/rapids-4-spark_2.12/24.04.1/rapids-4-spark_2.12-24.04.1.jar"
9999
100100
$SPARK_HOME/bin/spark-shell --master $SPARK_MASTER \
101101
--driver-memory 20G \

notebooks/aws-emr/init-bootstrap-action.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ sudo chmod a+rwx -R /sys/fs/cgroup/devices
88
sudo yum install -y gcc openssl-devel bzip2-devel libffi-devel tar gzip wget make mysql-devel
99
sudo bash -c "wget https://www.python.org/ftp/python/3.9.9/Python-3.9.9.tgz && tar xzf Python-3.9.9.tgz && cd Python-3.9.9 && ./configure --enable-optimizations && make altinstall"
1010

11-
RAPIDS_VERSION=24.4.0
11+
RAPIDS_VERSION=24.6.0
1212

1313
# install scikit-learn
1414
sudo /usr/local/bin/pip3.9 install scikit-learn

notebooks/databricks/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -51,7 +51,7 @@ If you already have a Databricks account, you can run the example notebooks on a
5151
spark.task.resource.gpu.amount 1
5252
spark.databricks.delta.preview.enabled true
5353
spark.python.worker.reuse true
54-
spark.executorEnv.PYTHONPATH /databricks/jars/rapids-4-spark_2.12-24.02.0.jar:/databricks/spark/python
54+
spark.executorEnv.PYTHONPATH /databricks/jars/rapids-4-spark_2.12-24.04.1.jar:/databricks/spark/python
5555
spark.sql.execution.arrow.maxRecordsPerBatch 100000
5656
spark.rapids.memory.gpu.minAllocFraction 0.0001
5757
spark.plugins com.nvidia.spark.SQLPlugin

0 commit comments

Comments
 (0)