object
SparkTestingUtils
Type Members
-
Value Members
-
final
def
!=(arg0: Any): Boolean
-
final
def
##(): Int
-
final
def
==(arg0: Any): Boolean
-
-
final
def
asInstanceOf[T0]: T0
-
def
assertDFEquals(expectedUnsorted: DataFrame, resultUnsorted: DataFrame, maxUnequalRowsToShow: Int, tol: Double = 0.0): Either[Error, String]
-
def
clone(): AnyRef
-
def
convertStringValueToDatatype(dataType: DataType, value: String): Any
-
def
createDF(spark: SparkSession, columns: Seq[String], values: Seq[Seq[Any]], defaults: Map[String, Any], typeMap: Map[String, DataType], port: String): DataFrame
-
def
createDfFromResourceFiles(spark: SparkSession, schemaDefinitionPath: String, dataPath: String, port: String): DataFrame
-
def
createDfFromSchemaFile(spark: SparkSession, schemaDefinitionPath: String, port: String, columns: Seq[String], values: Seq[Seq[Any]]): DataFrame
-
def
defaultForDatatype: PartialFunction[DataType, Any]
-
def
defaultsForSchema(schema: StructType): Map[String, Any]
-
final
def
eq(arg0: AnyRef): Boolean
-
def
equals(arg0: Any): Boolean
-
final
def
getClass(): Class[_]
-
def
hashCode(): Int
-
final
def
isInstanceOf[T0]: Boolean
-
def
mkString(row: Row, sep: String): String
-
final
def
ne(arg0: AnyRef): Boolean
-
final
def
notify(): Unit
-
final
def
notifyAll(): Unit
-
-
-
def
readData(spark: SparkSession, schema: StructType, path: String): DataFrame
-
def
readResource(path: String): String
-
def
rowEquals(r1: Row, r2: Row, tol: Double): Boolean
-
def
schemaToFieldMap(schema: StructType): Map[String, DataType]
-
final
def
synchronized[T0](arg0: ⇒ T0): T0
-
def
toString(): String
-
final
def
wait(arg0: Long, arg1: Int): Unit
-
final
def
wait(arg0: Long): Unit
-
final
def
wait(): Unit
-
def
writeData(pathPrefix: String, fileName: String, df: DataFrame, separator: String, maxRows: Int = 100): Unit
Deprecated Value Members
-
def
finalize(): Unit
Inherited from AnyRef
Inherited from Any