Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -17,12 +17,11 @@

package org.apache.spark.sql.catalyst.expressions

import org.apache.spark.SparkFunSuite
import org.apache.spark.sql.catalyst.dsl.expressions._
import org.apache.spark.sql.catalyst.plans.PlanTest
import org.apache.spark.sql.types.BooleanType

class ExtractPredicatesWithinOutputSetSuite extends SparkFunSuite with PlanTest {
class ExtractPredicatesWithinOutputSetSuite extends PlanTest {
private val a = AttributeReference("A", BooleanType)(exprId = ExprId(1))
private val b = AttributeReference("B", BooleanType)(exprId = ExprId(2))
private val c = AttributeReference("C", BooleanType)(exprId = ExprId(3))
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,6 @@ package org.apache.spark.sql.catalyst.plans

import java.util.TimeZone

import org.apache.spark.SparkFunSuite
import org.apache.spark.sql.catalyst.analysis._
import org.apache.spark.sql.catalyst.dsl.expressions._
import org.apache.spark.sql.catalyst.dsl.plans._
Expand All @@ -28,7 +27,7 @@ import org.apache.spark.sql.catalyst.plans.logical._
import org.apache.spark.sql.internal.SQLConf
import org.apache.spark.sql.types.{DataType, DoubleType, IntegerType, LongType, StringType}

class ConstraintPropagationSuite extends SparkFunSuite with PlanTest {
class ConstraintPropagationSuite extends PlanTest {

private def resolveColumn(tr: LocalRelation, columnName: String): Expression =
resolveColumn(tr.analyze, columnName)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -17,11 +17,11 @@

package org.apache.spark.sql

import org.apache.spark.{MapOutputTrackerMaster, SparkFunSuite}
import org.apache.spark.MapOutputTrackerMaster
import org.apache.spark.sql.classic.SparkSession
import org.apache.spark.sql.internal.SQLConf

class MapStatusEndToEndSuite extends SparkFunSuite with QueryTest {
class MapStatusEndToEndSuite extends QueryTest {
override def spark: SparkSession = SparkSession.builder()
.master("local")
.config(SQLConf.LEAF_NODE_DEFAULT_PARALLELISM.key, value = 5)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ import java.util.{Locale, UUID}
import scala.collection.mutable.ListBuffer
import scala.concurrent.Future

import org.apache.spark.{MapOutputStatistics, SparkFunSuite, TaskContext}
import org.apache.spark.{MapOutputStatistics, TaskContext}
import org.apache.spark.broadcast.Broadcast
import org.apache.spark.internal.Logging
import org.apache.spark.rdd.RDD
Expand Down Expand Up @@ -58,8 +58,7 @@ import org.apache.spark.unsafe.types.UTF8String
/**
* Test cases for the [[SparkSessionExtensions]].
*/
class SparkSessionExtensionSuite extends SparkFunSuite with SQLHelper with AdaptiveSparkPlanHelper
with PlanTest {
class SparkSessionExtensionSuite extends PlanTest with SQLHelper with AdaptiveSparkPlanHelper {
private def create(
builder: SparkSessionExtensionsProvider): Seq[SparkSessionExtensionsProvider] = Seq(builder)

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@ import org.apache.hadoop.mapreduce.{JobContext, OutputCommitter, TaskAttemptCont
import org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
import org.apache.parquet.hadoop.{ParquetOutputCommitter, ParquetOutputFormat}

import org.apache.spark.{LocalSparkContext, SparkFunSuite}
import org.apache.spark.LocalSparkContext
import org.apache.spark.sql.QueryTest
import org.apache.spark.sql.classic.SparkSession
import org.apache.spark.sql.internal.SQLConf
Expand All @@ -33,8 +33,7 @@ import org.apache.spark.tags.ExtendedSQLTest
* Test logic related to choice of output committers.
*/
@ExtendedSQLTest
class ParquetCommitterSuite extends SparkFunSuite with QueryTest
with LocalSparkContext {
class ParquetCommitterSuite extends QueryTest with LocalSparkContext {

private val PARQUET_COMMITTER = classOf[ParquetOutputCommitter].getCanonicalName

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ import java.util.UUID

import org.scalatest.Assertions._

import org.apache.spark.{SparkFunSuite, SparkNoSuchElementException, TaskContext}
import org.apache.spark.{SparkNoSuchElementException, TaskContext}
import org.apache.spark.rdd.RDD
import org.apache.spark.sql.QueryTest
import org.apache.spark.sql.catalyst.InternalRow
Expand All @@ -35,7 +35,7 @@ import org.apache.spark.sql.functions.col
import org.apache.spark.tags.ExtendedSQLTest

@ExtendedSQLTest
class ExecutorSideSQLConfSuite extends SparkFunSuite with QueryTest {
class ExecutorSideSQLConfSuite extends QueryTest {
import testImplicits._

protected var spark: SparkSession = null
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,6 @@ import java.util.Collections
import org.apache.hadoop.hive.metastore.api.FieldSchema
import org.apache.hadoop.hive.serde.serdeConstants

import org.apache.spark.SparkFunSuite
import org.apache.spark.sql.catalyst.dsl.expressions._
import org.apache.spark.sql.catalyst.expressions._
import org.apache.spark.sql.catalyst.plans.PlanTest
Expand All @@ -35,7 +34,7 @@ import org.apache.spark.unsafe.types.UTF8String
* A set of tests for the filter conversion logic used when pushing partition pruning into the
* metastore
*/
class FiltersSuite extends SparkFunSuite with PlanTest {
class FiltersSuite extends PlanTest {
private val shim = new Shim_v2_0

private val testTable = new org.apache.hadoop.hive.ql.metadata.Table("default", "test")
Expand Down