diff --git a/ci/Jenkinsfile-ppc64le-cuda10 b/ci/Jenkinsfile-ppc64le-cuda10 index 02404c38f..23ab09e5d 100644 --- a/ci/Jenkinsfile-ppc64le-cuda10 +++ b/ci/Jenkinsfile-ppc64le-cuda10 @@ -24,7 +24,7 @@ def linuxwheel = "ppc64le-centos7-cuda10.whl" def testtype = "dotest-single-gpu" def testtype_multi_gpu = "dotest-multi-gpu" def labelbuild = "ibm-power-gpu" -def labeltest_multi_gpu = "ibm-power-gpu && !mr-0xp2" +def labeltest_multi_gpu = "mr-0xp2" def labeltest = "ibm-power-gpu && !mr-0xp2" def labelruntime = "ibm-power-gpu" def doingbenchmark = "0" diff --git a/ci/base/Jenkinsfile-ppc64le-cuda10.base b/ci/base/Jenkinsfile-ppc64le-cuda10.base index 13d108687..99b5f6407 100644 --- a/ci/base/Jenkinsfile-ppc64le-cuda10.base +++ b/ci/base/Jenkinsfile-ppc64le-cuda10.base @@ -18,7 +18,7 @@ def linuxwheel = "ppc64le-centos7-cuda10.whl" def testtype = "dotest-single-gpu" def testtype_multi_gpu = "dotest-multi-gpu" def labelbuild = "ibm-power-gpu" -def labeltest_multi_gpu = "ibm-power-gpu && !mr-0xp2" +def labeltest_multi_gpu = "mr-0xp2" def labeltest = "ibm-power-gpu && !mr-0xp2" def labelruntime = "ibm-power-gpu" def doingbenchmark = "0" diff --git a/tests/python/open_data/gbm/test_xgboost.py b/tests/python/open_data/gbm/test_xgboost.py index 22874a8c6..4cc1d1ad8 100644 --- a/tests/python/open_data/gbm/test_xgboost.py +++ b/tests/python/open_data/gbm/test_xgboost.py @@ -113,7 +113,7 @@ def test_xgboost_covtype_multi_gpu(): with LocalCUDACluster(n_workers=n_gpus, threads_per_worker=1) as cluster: with Client(cluster) as client: # Convert input data from numpy to XGBoost format - partition_size = 100000 + partition_size = 10000 # remove when https://github.com/dmlc/xgboost/issues/4987 is fixed dask_X_train = da.from_array(X_train, partition_size)