Trait

com.holdenkarau.spark.testing

DataFrameSuiteBaseLike

Related Doc: package testing

Permalink

trait DataFrameSuiteBaseLike extends SparkContextProvider with TestSuiteLike with Serializable

Linear Supertypes
Serializable, Serializable, TestSuiteLike, SparkContextProvider, AnyRef, Any
Known Subclasses
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. DataFrameSuiteBaseLike
  2. Serializable
  3. Serializable
  4. TestSuiteLike
  5. SparkContextProvider
  6. AnyRef
  7. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Abstract Value Members

  1. abstract def assert[U](message: String, expected: U, actual: U)(implicit CT: ClassTag[U]): Unit

    Permalink
    Definition Classes
    TestSuiteLike
  2. abstract def assert[U](expected: U, actual: U)(implicit CT: ClassTag[U]): Unit

    Permalink
    Definition Classes
    TestSuiteLike
  3. abstract def assertEmpty[U](arr: Array[U])(implicit CT: ClassTag[U]): Unit

    Permalink
    Definition Classes
    TestSuiteLike
  4. abstract def assertTrue(expected: Boolean): Unit

    Permalink
    Definition Classes
    TestSuiteLike
  5. abstract def fail(message: String): Unit

    Permalink
    Definition Classes
    TestSuiteLike
  6. abstract def sc: SparkContext

    Permalink
    Definition Classes
    SparkContextProvider

Concrete Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  4. def appID: String

    Permalink
    Definition Classes
    SparkContextProvider
  5. def approxEquals(r1: Row, r2: Row, tol: Double): Boolean

    Permalink
  6. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  7. def assertDataFrameApproximateEquals(expected: DataFrame, result: DataFrame, tol: Double): Unit

    Permalink

    Compares if two DataFrames are equal, checks that the schemas are the same.

    Compares if two DataFrames are equal, checks that the schemas are the same. When comparing inexact fields uses tol.

    tol

    max acceptable tolerance, should be less than 1.

  8. def assertDataFrameDataEquals(expected: DataFrame, result: DataFrame): Unit

    Permalink

    Compares if two DataFrames are equal without caring about order of rows, by finding elements in one DataFrame that is not in the other.

    Compares if two DataFrames are equal without caring about order of rows, by finding elements in one DataFrame that is not in the other. The resulting DataFrame should be empty inferring the two DataFrames have the same elements. Does not compare the schema.

  9. def assertDataFrameEquals(expected: DataFrame, result: DataFrame): Unit

    Permalink

    Compares if two DataFrames are equal, checks the schema and then if that matches checks if the rows are equal.

  10. def assertDataFrameNoOrderEquals(expected: DataFrame, result: DataFrame): Unit

    Permalink

    Compares if two DataFrames are equal without caring about order of rows, by finding elements in one DataFrame that is not in the other.

    Compares if two DataFrames are equal without caring about order of rows, by finding elements in one DataFrame that is not in the other. The resulting DataFrame should be empty inferring the two DataFrames have the same elements. Also verifies that the schema is identical.

  11. def assertSmallDataFrameDataEquals(expected: DataFrame, result: DataFrame): Unit

    Permalink

    Compares if two DataFrames are equal without caring about order of rows, by finding elements in one DataFrame that is not in the other.

    Compares if two DataFrames are equal without caring about order of rows, by finding elements in one DataFrame that is not in the other. Similar to the function assertDataFrameDataEquals but for small DataFrame that can be collected in memory for the comparison.

  12. def builder(): Builder

    Permalink

    Constructs a configuration for hive or iceberg, where the metastore is located in a temp directory.

  13. def clone(): AnyRef

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @HotSpotIntrinsicCandidate() @throws( ... )
  14. def conf: SparkConf

    Permalink
    Definition Classes
    SparkContextProvider
  15. def enableHiveSupport: Boolean

    Permalink
    Attributes
    protected
  16. def enableIcebergSupport: Boolean

    Permalink
    Attributes
    protected
  17. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  18. def equals(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  19. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
    Annotations
    @HotSpotIntrinsicCandidate()
  20. def hashCode(): Int

    Permalink
    Definition Classes
    AnyRef → Any
    Annotations
    @HotSpotIntrinsicCandidate()
  21. val icebergWarehouse: String

    Permalink
  22. implicit def impSqlContext: SQLContext

    Permalink
    Attributes
    protected
  23. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  24. lazy val localMetastorePath: String

    Permalink
  25. lazy val localWarehousePath: String

    Permalink
  26. val maxUnequalRowsToShow: Int

    Permalink
  27. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  28. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @HotSpotIntrinsicCandidate()
  29. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @HotSpotIntrinsicCandidate()
  30. def setup(sc: SparkContext): Unit

    Permalink

    Setup work to be called when creating a new SparkContext.

    Setup work to be called when creating a new SparkContext. Default implementation currently sets a checkpoint directory.

    This _should_ be called by the context provider automatically.

    Definition Classes
    SparkContextProvider
  31. lazy val spark: SparkSession

    Permalink
  32. def sqlBeforeAllTestCases(): Unit

    Permalink
  33. lazy val sqlContext: SQLContext

    Permalink
  34. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  35. lazy val tempDir: File

    Permalink
  36. def toString(): String

    Permalink
    Definition Classes
    AnyRef → Any
  37. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  38. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  39. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )

Deprecated Value Members

  1. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @Deprecated @deprecated @throws( classOf[java.lang.Throwable] )
    Deprecated

    (Since version ) see corresponding Javadoc for more information.

Inherited from Serializable

Inherited from Serializable

Inherited from TestSuiteLike

Inherited from SparkContextProvider

Inherited from AnyRef

Inherited from Any

Ungrouped