Trait/Object

com.holdenkarau.spark.testing

DataFrameSuiteBase

Related Docs: object DataFrameSuiteBase | package testing

Permalink

trait DataFrameSuiteBase extends TestSuite with SharedSparkContext with DataFrameSuiteBaseLike

:: Experimental :: Base class for testing Spark DataFrames.

Self Type
DataFrameSuiteBase with Suite
Linear Supertypes
DataFrameSuiteBaseLike, Serializable, Serializable, SharedSparkContext, SparkContextProvider, BeforeAndAfterAll, SuiteMixin, TestSuite, TestSuiteLike, AnyRef, Any
Known Subclasses
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. DataFrameSuiteBase
  2. DataFrameSuiteBaseLike
  3. Serializable
  4. Serializable
  5. SharedSparkContext
  6. SparkContextProvider
  7. BeforeAndAfterAll
  8. SuiteMixin
  9. TestSuite
  10. TestSuiteLike
  11. AnyRef
  12. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Abstract Value Members

  1. abstract def expectedTestCount(filter: Filter): Int

    Permalink
    Definition Classes
    SuiteMixin
  2. abstract def nestedSuites: IndexedSeq[Suite]

    Permalink
    Definition Classes
    SuiteMixin
  3. abstract def rerunner: Option[String]

    Permalink
    Definition Classes
    SuiteMixin
  4. abstract def runNestedSuites(args: Args): Status

    Permalink
    Attributes
    protected
    Definition Classes
    SuiteMixin
  5. abstract def runTest(testName: String, args: Args): Status

    Permalink
    Attributes
    protected
    Definition Classes
    SuiteMixin
  6. abstract def runTests(testName: Option[String], args: Args): Status

    Permalink
    Attributes
    protected
    Definition Classes
    SuiteMixin
  7. abstract def suiteId: String

    Permalink
    Definition Classes
    SuiteMixin
  8. abstract def suiteName: String

    Permalink
    Definition Classes
    SuiteMixin
  9. abstract def tags: Map[String, Set[String]]

    Permalink
    Definition Classes
    SuiteMixin
  10. abstract def testDataFor(testName: String, theConfigMap: ConfigMap): TestData

    Permalink
    Definition Classes
    SuiteMixin
  11. abstract def testNames: Set[String]

    Permalink
    Definition Classes
    SuiteMixin
  12. abstract val styleName: String

    Permalink
    Definition Classes
    SuiteMixin
    Annotations
    @deprecated
    Deprecated

    (Since version 3.1.0) The styleName lifecycle method has been deprecated and will be removed in a future version of ScalaTest with no replacement.

Concrete Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  4. def afterAll(): Unit

    Permalink
    Definition Classes
    DataFrameSuiteBaseSharedSparkContext → BeforeAndAfterAll
  5. def appID: String

    Permalink
    Definition Classes
    SparkContextProvider
  6. def appName: String

    Permalink
    Definition Classes
    SparkContextProvider
  7. def approxEquals(r1: Row, r2: Row, tol: Double, tolTimestamp: Duration): Boolean

    Permalink
    Definition Classes
    DataFrameSuiteBaseLike
  8. def approxEquals(r1: Row, r2: Row, tolTimestamp: Duration): Boolean

    Permalink
    Definition Classes
    DataFrameSuiteBaseLike
  9. def approxEquals(r1: Row, r2: Row, tol: Double): Boolean

    Permalink
    Definition Classes
    DataFrameSuiteBaseLike
  10. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  11. def assert[U](message: String, expected: U, actual: U)(implicit CT: ClassTag[U]): Unit

    Permalink
    Definition Classes
    TestSuiteTestSuiteLike
  12. def assert[U](expected: U, actual: U)(implicit CT: ClassTag[U]): Unit

    Permalink
    Definition Classes
    TestSuiteTestSuiteLike
  13. def assertDataFrameApproximateEquals(expected: DataFrame, result: DataFrame, tol: Double, tolTimestamp: Duration, customShow: (DataFrame) ⇒ Unit = _.show()): Unit

    Permalink

    Compares if two DataFrames are equal, checks that the schemas are the same.

    Compares if two DataFrames are equal, checks that the schemas are the same. When comparing inexact fields uses tol & tolTimestamp.

    tol

    max acceptable numeric tolerance, should be less than 1.

    tolTimestamp

    max acceptable timestamp tolerance.

    customShow

    unit function to customize the show method when dataframes are not equal. IE: df.show(false) or df.toJSON.show(false).

    Definition Classes
    DataFrameSuiteBaseLike
  14. def assertDataFrameDataEquals(expected: DataFrame, result: DataFrame): Unit

    Permalink

    Compares if two DataFrames are equal without caring about order of rows, by finding elements in one DataFrame that is not in the other.

    Compares if two DataFrames are equal without caring about order of rows, by finding elements in one DataFrame that is not in the other. The resulting DataFrame should be empty inferring the two DataFrames have the same elements. Does not compare the schema.

    Definition Classes
    DataFrameSuiteBaseLike
  15. def assertDataFrameEquals(expected: DataFrame, result: DataFrame, customShow: (DataFrame) ⇒ Unit = _.show()): Unit

    Permalink

    Compares if two DataFrames are equal, checks the schema and then if that matches checks if the rows are equal.

    Compares if two DataFrames are equal, checks the schema and then if that matches checks if the rows are equal.

    customShow

    unit function to customize the show method when dataframes are not equal. IE: df.show(false) or df.toJSON.show(false).

    Definition Classes
    DataFrameSuiteBaseLike
  16. def assertDataFrameNoOrderEquals(expected: DataFrame, result: DataFrame): Unit

    Permalink

    Compares if two DataFrames are equal without caring about order of rows, by finding elements in one DataFrame that is not in the other.

    Compares if two DataFrames are equal without caring about order of rows, by finding elements in one DataFrame that is not in the other. The resulting DataFrame should be empty inferring the two DataFrames have the same elements. Also verifies that the schema is identical.

    Definition Classes
    DataFrameSuiteBaseLike
  17. def assertEmpty[U](arr: Array[U])(implicit CT: ClassTag[U]): Unit

    Permalink
    Definition Classes
    TestSuiteTestSuiteLike
  18. def assertSchemasEqual(expected: StructType, result: StructType): Unit

    Permalink

    Compare if two schemas are equal, ignoring autoGeneratedAlias magic

    Compare if two schemas are equal, ignoring autoGeneratedAlias magic

    Definition Classes
    DataFrameSuiteBaseLike
  19. def assertSmallDataFrameDataEquals(expected: DataFrame, result: DataFrame): Unit

    Permalink

    Compares if two DataFrames are equal without caring about order of rows, by finding elements in one DataFrame that is not in the other.

    Compares if two DataFrames are equal without caring about order of rows, by finding elements in one DataFrame that is not in the other. Similar to the function assertDataFrameDataEquals but for small DataFrame that can be collected in memory for the comparison.

    Definition Classes
    DataFrameSuiteBaseLike
  20. def assertTrue(expected: Boolean): Unit

    Permalink
    Definition Classes
    TestSuiteTestSuiteLike
  21. def beforeAll(): Unit

    Permalink
    Definition Classes
    DataFrameSuiteBaseSharedSparkContext → BeforeAndAfterAll
  22. def builder(): Builder

    Permalink

    Constructs a configuration for hive or iceberg, where the metastore is located in a temp directory.

    Constructs a configuration for hive or iceberg, where the metastore is located in a temp directory.

    Definition Classes
    DataFrameSuiteBaseLike
  23. def clone(): AnyRef

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @IntrinsicCandidate() @throws( ... )
  24. def conf: SparkConf

    Permalink
    Definition Classes
    SparkContextProvider
  25. def enableHiveSupport: Boolean

    Permalink
    Attributes
    protected
    Definition Classes
    DataFrameSuiteBaseLike
  26. def enableIcebergSupport: Boolean

    Permalink
    Attributes
    protected
    Definition Classes
    DataFrameSuiteBaseLike
  27. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  28. def equals(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  29. def fail(message: String): Unit

    Permalink
    Definition Classes
    TestSuiteTestSuiteLike
  30. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
    Annotations
    @IntrinsicCandidate()
  31. def hashCode(): Int

    Permalink
    Definition Classes
    AnyRef → Any
    Annotations
    @IntrinsicCandidate()
  32. val icebergWarehouse: String

    Permalink
    Definition Classes
    DataFrameSuiteBaseLike
  33. implicit def impSqlContext: SQLContext

    Permalink
    Attributes
    protected
    Definition Classes
    DataFrameSuiteBaseLike
  34. val invokeBeforeAllAndAfterAllEvenIfNoTestsAreExpected: Boolean

    Permalink
    Definition Classes
    BeforeAndAfterAll
  35. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  36. lazy val localMetastorePath: String

    Permalink
    Definition Classes
    DataFrameSuiteBaseLike
  37. lazy val localWarehousePath: String

    Permalink
    Definition Classes
    DataFrameSuiteBaseLike
  38. val maxUnequalRowsToShow: Int

    Permalink
    Definition Classes
    DataFrameSuiteBaseLike
  39. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  40. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @IntrinsicCandidate()
  41. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @IntrinsicCandidate()
  42. implicit def reuseContextIfPossible: Boolean

    Permalink
    Attributes
    protected
    Definition Classes
    SharedSparkContext
  43. def run(testName: Option[String], args: Args): Status

    Permalink
    Definition Classes
    BeforeAndAfterAll → SuiteMixin
  44. def sc: SparkContext

    Permalink
  45. def setup(sc: SparkContext): Unit

    Permalink

    Setup work to be called when creating a new SparkContext.

    Setup work to be called when creating a new SparkContext. Default implementation currently sets a checkpoint directory.

    This _should_ be called by the context provider automatically.

    Definition Classes
    SparkContextProvider
  46. lazy val spark: SparkSession

    Permalink
    Definition Classes
    DataFrameSuiteBaseLike
  47. def sqlBeforeAllTestCases(): Unit

    Permalink
    Definition Classes
    DataFrameSuiteBaseLike
  48. lazy val sqlContext: SQLContext

    Permalink
    Definition Classes
    DataFrameSuiteBaseLike
  49. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  50. lazy val tempDir: File

    Permalink
    Definition Classes
    DataFrameSuiteBaseLike
  51. def toString(): String

    Permalink
    Definition Classes
    AnyRef → Any
  52. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  53. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  54. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  55. def withSQLConf(pairs: (String, String)*)(f: ⇒ Unit): Unit

    Permalink

    Sets all SQL configurations specified in pairs, calls f, and then restores all SQL configurations.

    Sets all SQL configurations specified in pairs, calls f, and then restores all SQL configurations. Taken from Spark SQLHelper.

    Attributes
    protected

Deprecated Value Members

  1. def assertDataFrameApproximateEquals(expected: DataFrame, result: DataFrame, tol: Double): Unit

    Permalink

    Compares if two DataFrames are equal, checks that the schemas are the same.

    Compares if two DataFrames are equal, checks that the schemas are the same. When comparing inexact fields uses tol.

    tol

    max acceptable tolerance for numeric (between(0, 1)) & timestamp (millis).

    Definition Classes
    DataFrameSuiteBaseLike
    Annotations
    @deprecated
    Deprecated

    (Since version 1.5.0) Use assertDataFrameApproximateEquals with timestamp tolerance

  2. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @Deprecated @deprecated @throws( classOf[java.lang.Throwable] )
    Deprecated

    (Since version ) see corresponding Javadoc for more information.

Inherited from DataFrameSuiteBaseLike

Inherited from Serializable

Inherited from Serializable

Inherited from SharedSparkContext

Inherited from SparkContextProvider

Inherited from BeforeAndAfterAll

Inherited from SuiteMixin

Inherited from TestSuite

Inherited from TestSuiteLike

Inherited from AnyRef

Inherited from Any

Ungrouped