Skip to content

Commit a16627a

Browse files
committed
Release v21.06.2
Signed-off-by: Peixin Li <pxli@nyu.edu>
2 parents 3e399ac + bd971b9 commit a16627a

File tree

42 files changed

+148
-116
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

42 files changed

+148
-116
lines changed

.github/workflows/auto-merge.yml

Lines changed: 0 additions & 41 deletions
This file was deleted.

CHANGELOG.md

Lines changed: 17 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,20 @@
11
# Change log
2-
Generated on 2021-08-04
2+
Generated on 2021-08-12
3+
4+
## Release 21.06.2
5+
6+
### Bugs Fixed
7+
|||
8+
|:---|:---|
9+
|[#3191](https://github.com/NVIDIA/spark-rapids/issues/3191)|[BUG] Databricks parquetFilters build failure in db 8.2 runtime|
10+
11+
### PRs
12+
|||
13+
|:---|:---|
14+
|[#3208](https://github.com/NVIDIA/spark-rapids/pull/3208)|Update rapids plugin version to 21.06.2 [skip ci]|
15+
|[#3207](https://github.com/NVIDIA/spark-rapids/pull/3207)|Disable auto-merge from 21.06 to 21.08 [skip ci]|
16+
|[#3205](https://github.com/NVIDIA/spark-rapids/pull/3205)|Branch 21.06 databricks update [skip ci]|
17+
|[#3198](https://github.com/NVIDIA/spark-rapids/pull/3198)|Databricks parquetFilters api change in db 8.2 runtime|
318

419
## Release 21.06.1
520

@@ -11,6 +26,7 @@ Generated on 2021-08-04
1126
### PRs
1227
|||
1328
|:---|:---|
29+
|[#3127](https://github.com/NVIDIA/spark-rapids/pull/3127)|Update CHANGELOG for the release v21.06.1 [skip ci]|
1430
|[#3123](https://github.com/NVIDIA/spark-rapids/pull/3123)|Update rapids plugin version to 21.06.1 [skip ci]|
1531
|[#3118](https://github.com/NVIDIA/spark-rapids/pull/3118)|Fix databricks 3.0.1 for ParquetFilters api change|
1632
|[#3119](https://github.com/NVIDIA/spark-rapids/pull/3119)|Branch 21.06 databricks update [skip ci]|

api_validation/pom.xml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -22,10 +22,10 @@
2222
<parent>
2323
<groupId>com.nvidia</groupId>
2424
<artifactId>rapids-4-spark-parent</artifactId>
25-
<version>21.06.1</version>
25+
<version>21.06.2</version>
2626
</parent>
2727
<artifactId>rapids-4-spark-api-validation</artifactId>
28-
<version>21.06.1</version>
28+
<version>21.06.2</version>
2929

3030
<profiles>
3131
<profile>
@@ -78,7 +78,7 @@
7878
<dependency>
7979
<groupId>com.nvidia</groupId>
8080
<artifactId>rapids-4-spark-shims-aggregator_${scala.binary.version}</artifactId>
81-
<version>21.06.1</version>
81+
<version>21.06.2</version>
8282
<scope>provided</scope>
8383
</dependency>
8484
</dependencies>

dist/pom.xml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -22,12 +22,12 @@
2222
<parent>
2323
<groupId>com.nvidia</groupId>
2424
<artifactId>rapids-4-spark-parent</artifactId>
25-
<version>21.06.1</version>
25+
<version>21.06.2</version>
2626
</parent>
2727
<artifactId>rapids-4-spark_2.12</artifactId>
2828
<name>RAPIDS Accelerator for Apache Spark Distribution</name>
2929
<description>Creates the distribution package of the RAPIDS plugin for Apache Spark</description>
30-
<version>21.06.1</version>
30+
<version>21.06.2</version>
3131

3232
<dependencies>
3333
<dependency>

docs/configs.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ The following is the list of options that `rapids-plugin-4-spark` supports.
1010
On startup use: `--conf [conf key]=[conf value]`. For example:
1111

1212
```
13-
${SPARK_HOME}/bin/spark --jars 'rapids-4-spark_2.12-21.06.1.jar,cudf-21.06.1-cuda11.jar' \
13+
${SPARK_HOME}/bin/spark --jars 'rapids-4-spark_2.12-21.06.2.jar,cudf-21.06.1-cuda11.jar' \
1414
--conf spark.plugins=com.nvidia.spark.SQLPlugin \
1515
--conf spark.rapids.sql.incompatibleOps.enabled=true
1616
```
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
{"cells":[{"cell_type":"code","source":["dbutils.fs.mkdirs(\"dbfs:/databricks/init_scripts/\")\n \ndbutils.fs.put(\"/databricks/init_scripts/init.sh\",\"\"\"\n#!/bin/bash\nsudo wget -O /databricks/jars/rapids-4-spark_2.12-21.06.1.jar https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/21.06.1/rapids-4-spark_2.12-21.06.1.jar\nsudo wget -O /databricks/jars/cudf-21.06.1-cuda11.jar https://repo1.maven.org/maven2/ai/rapids/cudf/21.06.1/cudf-21.06.1-cuda11.jar\n\nsudo wget -O /etc/apt/preferences.d/cuda-repository-pin-600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu1804/x86_64/cuda-ubuntu1804.pin\nsudo wget -O ~/cuda-repo-ubuntu1804-11-0-local_11.0.3-450.51.06-1_amd64.deb https://developer.download.nvidia.com/compute/cuda/11.0.3/local_installers/cuda-repo-ubuntu1804-11-0-local_11.0.3-450.51.06-1_amd64.deb\nsudo dpkg -i ~/cuda-repo-ubuntu1804-11-0-local_11.0.3-450.51.06-1_amd64.deb\nsudo apt-key add /var/cuda-repo-ubuntu1804-11-0-local/7fa2af80.pub\nsudo apt-get update\nsudo apt -y install cuda-toolkit-11-0\"\"\", True)"],"metadata":{},"outputs":[],"execution_count":1},{"cell_type":"code","source":["%sh\ncd ../../dbfs/databricks/init_scripts\npwd\nls -ltr\ncat init.sh"],"metadata":{},"outputs":[],"execution_count":2},{"cell_type":"code","source":[""],"metadata":{},"outputs":[],"execution_count":3}],"metadata":{"name":"generate-init-script","notebookId":2645746662301564},"nbformat":4,"nbformat_minor":0}
1+
{"cells":[{"cell_type":"code","source":["dbutils.fs.mkdirs(\"dbfs:/databricks/init_scripts/\")\n \ndbutils.fs.put(\"/databricks/init_scripts/init.sh\",\"\"\"\n#!/bin/bash\nsudo wget -O /databricks/jars/rapids-4-spark_2.12-21.06.2.jar https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/21.06.2/rapids-4-spark_2.12-21.06.2.jar\nsudo wget -O /databricks/jars/cudf-21.06.1-cuda11.jar https://repo1.maven.org/maven2/ai/rapids/cudf/21.06.1/cudf-21.06.1-cuda11.jar\n\nsudo wget -O /etc/apt/preferences.d/cuda-repository-pin-600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu1804/x86_64/cuda-ubuntu1804.pin\nsudo wget -O ~/cuda-repo-ubuntu1804-11-0-local_11.0.3-450.51.06-1_amd64.deb https://developer.download.nvidia.com/compute/cuda/11.0.3/local_installers/cuda-repo-ubuntu1804-11-0-local_11.0.3-450.51.06-1_amd64.deb\nsudo dpkg -i ~/cuda-repo-ubuntu1804-11-0-local_11.0.3-450.51.06-1_amd64.deb\nsudo apt-key add /var/cuda-repo-ubuntu1804-11-0-local/7fa2af80.pub\nsudo apt-get update\nsudo apt -y install cuda-toolkit-11-0\"\"\", True)"],"metadata":{},"outputs":[],"execution_count":1},{"cell_type":"code","source":["%sh\ncd ../../dbfs/databricks/init_scripts\npwd\nls -ltr\ncat init.sh"],"metadata":{},"outputs":[],"execution_count":2},{"cell_type":"code","source":[""],"metadata":{},"outputs":[],"execution_count":3}],"metadata":{"name":"generate-init-script","notebookId":2645746662301564},"nbformat":4,"nbformat_minor":0}
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
{"cells":[{"cell_type":"code","source":["dbutils.fs.mkdirs(\"dbfs:/databricks/init_scripts/\")\n \ndbutils.fs.put(\"/databricks/init_scripts/init.sh\",\"\"\"\n#!/bin/bash\nsudo wget -O /databricks/jars/rapids-4-spark_2.12-21.06.1.jar https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/21.06.1/rapids-4-spark_2.12-21.06.1.jar\nsudo wget -O /databricks/jars/cudf-21.06.1-cuda11.jar https://repo1.maven.org/maven2/ai/rapids/cudf/21.06.1/cudf-21.06.1-cuda11.jar\"\"\", True)"],"metadata":{},"outputs":[],"execution_count":1},{"cell_type":"code","source":["%sh\ncd ../../dbfs/databricks/init_scripts\npwd\nls -ltr\ncat init.sh"],"metadata":{},"outputs":[],"execution_count":2},{"cell_type":"code","source":[""],"metadata":{},"outputs":[],"execution_count":3}],"metadata":{"name":"generate-init-script","notebookId":2645746662301564},"nbformat":4,"nbformat_minor":0}
1+
{"cells":[{"cell_type":"code","source":["dbutils.fs.mkdirs(\"dbfs:/databricks/init_scripts/\")\n \ndbutils.fs.put(\"/databricks/init_scripts/init.sh\",\"\"\"\n#!/bin/bash\nsudo wget -O /databricks/jars/rapids-4-spark_2.12-21.06.2.jar https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/21.06.2/rapids-4-spark_2.12-21.06.2.jar\nsudo wget -O /databricks/jars/cudf-21.06.1-cuda11.jar https://repo1.maven.org/maven2/ai/rapids/cudf/21.06.1/cudf-21.06.1-cuda11.jar\"\"\", True)"],"metadata":{},"outputs":[],"execution_count":1},{"cell_type":"code","source":["%sh\ncd ../../dbfs/databricks/init_scripts\npwd\nls -ltr\ncat init.sh"],"metadata":{},"outputs":[],"execution_count":2},{"cell_type":"code","source":[""],"metadata":{},"outputs":[],"execution_count":3}],"metadata":{"name":"generate-init-script","notebookId":2645746662301564},"nbformat":4,"nbformat_minor":0}

docs/download.md

Lines changed: 43 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,6 +18,48 @@ cuDF jar, that is either preinstalled in the Spark classpath on all nodes or sub
1818
that uses the RAPIDS Accelerator For Apache Spark. See the [getting-started
1919
guide](https://nvidia.github.io/spark-rapids/Getting-Started/) for more details.
2020

21+
## Release v21.06.2
22+
This is a patch release to address an issue with the plugin in the Databricks 8.2 ML runtime.
23+
24+
Hardware Requirements:
25+
26+
GPU Architecture: NVIDIA V100, T4 or A10/A30/A100 GPUs
27+
28+
Software Requirements:
29+
30+
OS: Ubuntu 18.04, Ubuntu 20.04 or CentOS 7, CentOS 8
31+
32+
CUDA & Nvidia Drivers*: 11.0 or 11.2 & v450.80.02+
33+
34+
Apache Spark 3.0.1, 3.0.2, 3.1.1, 3.1.2, Cloudera CDP 7.1.7, Databricks 7.3 ML LTS or 8.2 ML Runtime, and GCP Dataproc 2.0
35+
36+
Apache Hadoop 2.10+ or 3.1.1+ (3.1.1 for nvidia-docker version 2)
37+
38+
Python 3.6+, Scala 2.12, Java 8
39+
40+
*Some hardware may have a minimum driver version greater than v450.80.02+. Check the GPU spec sheet
41+
for your hardware's minimum driver version.
42+
43+
### Download v21.06.2
44+
* Download the [RAPIDS
45+
Accelerator for Apache Spark 21.06.2 jar](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/21.06.2/rapids-4-spark_2.12-21.06.2.jar)
46+
* Download the [RAPIDS cuDF 21.06.1 jar](https://repo1.maven.org/maven2/ai/rapids/cudf/21.06.1/cudf-21.06.1-cuda11.jar)
47+
48+
This package is built against CUDA 11.2 and has [CUDA forward
49+
compatibility](https://docs.nvidia.com/deploy/cuda-compatibility/index.html) enabled. It is tested
50+
on Tesla datacenter GPUs with CUDA 11.0 and 11.2. For those using other types of GPUs which
51+
do not have CUDA forward compatibility (for example, GeForce), CUDA 11.2 is required. Users will
52+
need to ensure the minimum driver (450.80.02) and CUDA toolkit are installed on each Spark node.
53+
54+
### Release Notes
55+
This release patches the plugin to address a backwards incompatible change to Parquet filters made
56+
by Databricks in the Databricks 8.2 ML runtime. More information is in [issue
57+
3191](https://github.com/NVIDIA/spark-rapids/issues/3191) in the RAPIDS Spark repository. See the
58+
[Release v21.06.0](#release-v21060) release notes for more detail about new features in 21.06.
59+
60+
For a detailed list of changes, please refer to the
61+
[CHANGELOG](https://github.com/NVIDIA/spark-rapids/blob/main/CHANGELOG.md).
62+
2163
## Release v21.06.1
2264
This is a patch release to address an issue with the plugin in the Databricks 7.3 ML LTS runtime.
2365

@@ -77,7 +119,7 @@ Software Requirements:
77119

78120
CUDA & Nvidia Drivers*: 11.0 or 11.2 & v450.80.02+
79121

80-
Apache Spark 3.0.1, 3.0.2, 3.1.1, 3.1.2, Cloudera CDP 7.1.7, Databricks 8.2 ML Runtime, and GCP Dataproc 2.0
122+
Apache Spark 3.0.1, 3.0.2, 3.1.1, 3.1.2, Cloudera CDP 7.1.7, and GCP Dataproc 2.0
81123

82124
Apache Hadoop 2.10+ or 3.1.1+ (3.1.1 for nvidia-docker version 2)
83125

docs/get-started/Dockerfile.cuda

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -51,7 +51,7 @@ COPY spark-3.0.2-bin-hadoop3.2/kubernetes/tests /opt/spark/tests
5151
COPY spark-3.0.2-bin-hadoop3.2/data /opt/spark/data
5252

5353
COPY cudf-21.06.1-cuda11.jar /opt/sparkRapidsPlugin
54-
COPY rapids-4-spark_2.12-21.06.1.jar /opt/sparkRapidsPlugin
54+
COPY rapids-4-spark_2.12-21.06.2.jar /opt/sparkRapidsPlugin
5555
COPY getGpusResources.sh /opt/sparkRapidsPlugin
5656

5757
RUN mkdir /opt/spark/python

docs/get-started/getting-started-databricks.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -48,13 +48,13 @@ cluster.
4848
version:
4949
- [Databricks 7.3 LTS
5050
ML](https://docs.databricks.com/release-notes/runtime/7.3ml.html#system-environment) runs CUDA 10.1
51-
Update 2. Users wishing to try 21.06.1 on Databricks 7.3 LTS ML will need to install the CUDA
51+
Update 2. Users wishing to try 21.06.1 or later on Databricks 7.3 LTS ML will need to install the CUDA
5252
11.0 toolkit on the cluster. This can be done with the [generate-init-script-cuda11.ipynb
5353
](../demo/Databricks/generate-init-script-cuda11.ipynb) init script, which installs both the RAPIDS
5454
Spark plugin and the CUDA 11 toolkit.
5555
- [Databricks 8.2
5656
ML](https://docs.databricks.com/release-notes/runtime/8.2ml.html#system-environment) has CUDA 11
57-
installed. In this case use
57+
installed. Users will need to use 21.06.2 or later on Databricks 8.2 ML. In this case use
5858
[generate-init-script.ipynb](../demo/Databricks/generate-init-script.ipynb) which will install
5959
the RAPIDS Spark plugin.
6060
2. Once you are in the notebook, click the “Run All” button.
@@ -110,7 +110,7 @@ Spark plugin and the CUDA 11 toolkit.
110110
```bash
111111
spark.rapids.sql.python.gpu.enabled true
112112
spark.python.daemon.module rapids.daemon_databricks
113-
spark.executorEnv.PYTHONPATH /databricks/jars/rapids-4-spark_2.12-21.06.1.jar:/databricks/spark/python
113+
spark.executorEnv.PYTHONPATH /databricks/jars/rapids-4-spark_2.12-21.06.2.jar:/databricks/spark/python
114114
```
115115

116116
7. Once you’ve added the Spark config, click “Confirm and Restart”.

0 commit comments

Comments
 (0)