You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/site/FAQ.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -24,5 +24,5 @@ Be aware that some Spark Rapids ML algorithms (such as NearestNeighbors) may con
24
24
### What are some possible causes of low-level CUDA and/or native code errors?
25
25
26
26
- NaNs or nulls in the input data. These are currently passed directly into the cuML layer and may trigger such errors.
27
-
- NCCL communication library does not allow communication between processes on the same GPU. Check your Spark GPU configs to ensure 1 task per GPU during fit() calls.
27
+
- NCCL communication library does not allow communication between processes on the same GPU. [Stage level scheduling](https://nvidia.github.io/spark-rapids-ml/performance.html#stage-level-scheduling) can avoid this but it is not supported in all cases. Check requirements and adjust your Spark GPU configs to ensure 1 task per GPU during fit() calls if needed.
0 commit comments