Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,7 @@ class CelebornHashCheckDiskSuite extends SparkTestBase {
System.gc()
}

test("celeborn spark integration test - hash-checkDiskFull") {
ignore("celeborn spark integration test - hash-checkDiskFull") {
val sparkConf = new SparkConf().setAppName("celeborn-demo")
.setMaster("local[2]")
.set(s"spark.${CelebornConf.SHUFFLE_EXPIRED_CHECK_INTERVAL.key}", "20s")
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@ class RetryReviveTest extends AnyFunSuite
System.gc()
}

test("celeborn spark integration test - retry revive as configured times") {
ignore("celeborn spark integration test - retry revive as configured times") {
setupMiniClusterWithRandomPorts()
ShuffleClient.reset()
val sparkConf = new SparkConf()
Expand All @@ -57,7 +57,7 @@ class RetryReviveTest extends AnyFunSuite
ss.stop()
}

test(
ignore(
"celeborn spark integration test - e2e test retry revive with new allocated workers from RPC") {
val testConf = Map(
s"${CelebornConf.CLIENT_PUSH_MAX_REVIVE_TIMES.key}" -> "3",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ class MemoryStorageReusedExchangeSuite extends AnyFunSuite
ShuffleClient.reset()
}

test("[CELEBORN-980] Asynchronously delete original files to fix ReusedExchange bug") {
ignore("[CELEBORN-980] Asynchronously delete original files to fix ReusedExchange bug") {
testReusedExchange(false)
}

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,7 @@ class SparkUtilsSuite extends AnyFunSuite
System.gc()
}

test("check if fetch failure task another attempt is running or successful") {
ignore("check if fetch failure task another attempt is running or successful") {
if (Spark3OrNewer) {
val sparkConf = new SparkConf().setAppName("rss-demo").setMaster("local[2,3]")
val sparkSession = SparkSession.builder()
Expand Down
Loading