org.apache.spark.sql.catalyst

util

package util

Linear Supertypes
AnyRef, Any
Ordering
  1. Alphabetic
  2. By inheritance
Inherited
  1. util
  2. AnyRef
  3. Any
  1. Hide All
  2. Show all
Learn more about member selection
Visibility
  1. Public
  2. All

Type Members

  1. class AbstractScalaRowIterator[T] extends Iterator[T]

    Shim to allow us to implement scala.Iterator in Java.

  2. class ArrayBasedMapData extends MapData

  3. abstract class ArrayData extends SpecializedGetters with Serializable

  4. case class BadRecordException(record: () ⇒ UTF8String, partialResult: () ⇒ Option[InternalRow], cause: Throwable) extends Exception with Product with Serializable

    Exception thrown when the underlying parser meet a bad record and can't parse it.

  5. class CaseInsensitiveMap[T] extends Map[String, T] with Serializable

    Builds a map in which keys are case insensitive.

  6. class GenericArrayData extends ArrayData

  7. abstract class MapData extends Serializable

    This is an internal data representation for map type in Spark SQL.

  8. sealed trait ParseMode extends AnyRef

  9. class QuantileSummaries extends Serializable

    Helper class to compute approximate quantile summary.

  10. class StringKeyHashMap[T] extends AnyRef

Value Members

  1. object ArrayBasedMapData extends Serializable

  2. object ArrayData extends Serializable

  3. object CaseInsensitiveMap extends Serializable

  4. object CompressionCodecs

  5. object DateTimeUtils

    Helper functions for converting between internal and external date and time representations.

  6. object DropMalformedMode extends ParseMode with Product with Serializable

    This mode ignores the whole corrupted records.

  7. object FailFastMode extends ParseMode with Product with Serializable

    This mode throws an exception when it meets corrupted records.

  8. object NumberConverter

  9. object ParseMode extends Logging

  10. object PermissiveMode extends ParseMode with Product with Serializable

    This mode permissively parses the records.

  11. object QuantileSummaries extends Serializable

  12. object StringKeyHashMap

    Build a map with String type of key, and it also supports either key case sensitive or insensitive.

  13. object StringUtils

  14. object TypeUtils

    Helper functions to check for valid data types.

  15. def benchmark[A](f: ⇒ A): A

  16. def fileToString(file: File, encoding: String = "UTF-8"): String

  17. def quietly[A](f: ⇒ A): A

    Silences output to stderr or stdout for the duration of f

  18. def quoteIdentifier(name: String): String

  19. def resourceToBytes(resource: String, classLoader: ClassLoader = Utils.getSparkClassLoader): Array[Byte]

  20. def resourceToString(resource: String, encoding: String = "UTF-8", classLoader: ClassLoader = Utils.getSparkClassLoader): String

  21. def sideBySide(left: Seq[String], right: Seq[String]): Seq[String]

  22. def sideBySide(left: String, right: String): Seq[String]

  23. def stackTraceToString(t: Throwable): String

  24. def stringOrNull(a: AnyRef): String

  25. def stringToFile(file: File, str: String): File

  26. def toPrettySQL(e: Expression): String

  27. def usePrettyExpression(e: Expression): Expression

Inherited from AnyRef

Inherited from Any

Ungrouped