Object

org.apache.spark.sql.collection

Utils

Related Doc: package collection

Permalink

object Utils

Linear Supertypes
AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. Utils
  2. AnyRef
  3. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  4. def ERROR_NO_QCS(module: String): String

    Permalink
  5. final val SKIP_ANALYSIS_PREFIX: String("SAMPLE_")

    Permalink
  6. final val WEIGHTAGE_COLUMN_NAME: String("SNAPPY_SAMPLER_WEIGHTAGE")

    Permalink
  7. final val Z95Percent: Double

    Permalink
  8. final val Z95Squared: Double

    Permalink
  9. def analysisException(msg: String, cause: Option[Throwable] = None): AnalysisException

    Permalink
  10. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  11. def charMetadata(size: Int, md: Metadata): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    size

    the size parameter of the CHAR() column type

    md

    optional Metadata object to be merged into the result

    returns

    the result Metadata object to use for StructField

  12. def charMetadata(size: Int): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    size

    the size parameter of the CHAR() column type

    returns

    the result Metadata object to use for StructField

  13. def charMetadata(): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    returns

    the result Metadata object to use for StructField

  14. def classForName(className: String): Class[_]

    Permalink
  15. def clearDefaultSerializerAndCodec(): Unit

    Permalink
  16. def clone(): AnyRef

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  17. def closeJsonGenerator(gen: AnyRef): Unit

    Permalink
  18. def columnIndex(col: String, cols: Array[String], module: String): Int

    Permalink
  19. def createCatalystConverter(dataType: DataType): (Any) ⇒ Any

    Permalink
  20. def createScalaConverter(dataType: DataType): (Any) ⇒ Any

    Permalink
  21. def createStatsBuffer(statsData: Array[Byte], allocator: BufferAllocator): ByteBuffer

    Permalink
  22. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  23. def equals(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  24. def executorsListener(sc: SparkContext): Option[ExecutorsListener]

    Permalink
  25. def fieldIndex(relationOutput: Seq[Attribute], columnName: String, caseSensitive: Boolean): Int

    Permalink
  26. def fieldName(f: StructField): String

    Permalink
  27. def fillArray[T](a: Array[_ >: T], v: T, start: Int, endP1: Int): Unit

    Permalink
  28. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  29. def genTaskContextFunction(ctx: CodegenContext): String

    Permalink
  30. def generateJson(gen: AnyRef, row: InternalRow, columnIndex: Int, columnType: DataType): Unit

    Permalink
  31. def getActiveSession: Option[SparkSession]

    Permalink
  32. def getAllExecutorsMemoryStatus(sc: SparkContext): Map[BlockManagerId, (Long, Long)]

    Permalink
  33. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
  34. def getClientHostPort(netServer: String): String

    Permalink
  35. def getDriverClassName(url: String): String

    Permalink
  36. def getFields(o: Any): Map[String, Any]

    Permalink
  37. def getFixedPartitionRDD[T](sc: SparkContext, f: (TaskContext, Partition) ⇒ Iterator[T], partitioner: Partitioner, numPartitions: Int)(implicit arg0: ClassTag[T]): RDD[T]

    Permalink
  38. def getGenericRowValues(row: GenericRow): Array[Any]

    Permalink
  39. def getHostExecutorId(blockId: BlockManagerId): String

    Permalink
  40. def getInternalSparkConf(sc: SparkContext): SparkConf

    Permalink
  41. def getInternalType(dataType: DataType): Class[_]

    Permalink
  42. def getJsonGenerator(dataType: DataType, columnName: String, writer: Writer): AnyRef

    Permalink
  43. def getNumColumns(partitioning: Partitioning): Int

    Permalink
  44. def getPartitionData(blockId: BlockId, bm: BlockManager): ByteBuffer

    Permalink
  45. def getSQLDataType(dataType: DataType): DataType

    Permalink
  46. def getSchemaAndPlanFromBase(schemaOpt: Option[StructType], baseTableOpt: Option[String], catalog: SnappyStoreHiveCatalog, asSelect: Boolean, table: String, tableType: String): (StructType, Option[LogicalPlan])

    Permalink

    Get the result schema given an optional explicit schema and base table.

    Get the result schema given an optional explicit schema and base table. In case both are specified, then check compatibility between the two.

  47. def hasLowerCase(k: String): Boolean

    Permalink
  48. def hashCode(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  49. def immutableMap[A, B](m: Map[A, B]): Map[A, B]

    Permalink
  50. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  51. final def isLoner(sc: SparkContext): Boolean

    Permalink
  52. def mapExecutors[T](sc: SparkContext, f: (TaskContext, ExecutorLocalPartition) ⇒ Iterator[T], maxTries: Int)(implicit arg0: ClassTag[T]): Array[T]

    Permalink
  53. def mapExecutors[T](sc: SparkContext, f: () ⇒ Iterator[T], maxTries: Int = 30, blockManagerIds: Seq[BlockManagerId] = Seq.empty)(implicit arg0: ClassTag[T]): Array[T]

    Permalink
  54. def matchOption(optName: String, options: Map[String, Any]): Option[(String, Any)]

    Permalink
  55. def metricMethods: ((String) ⇒ String, (String) ⇒ String)

    Permalink
  56. def millisToDays(millisUtc: Long, tz: TimeZone): Int

    Permalink
  57. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  58. def newChunkedByteBuffer(chunks: Array[ByteBuffer]): ChunkedByteBuffer

    Permalink
  59. def newClusterSparkConf(props: Map[AnyRef, AnyRef]): SparkConf

    Permalink
  60. def newClusterSparkConf(): SparkConf

    Permalink
  61. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
  62. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
  63. def parseColumn(cv: Any, cols: Array[String], module: String, option: String): Int

    Permalink
  64. def parseColumnsAsClob(s: String): (Boolean, Set[String])

    Permalink
  65. def parseDouble(v: Any, module: String, option: String, min: Double, max: Double, exclusive: Boolean = true): Double

    Permalink
  66. def parseInteger(v: Any, module: String, option: String, min: Int = 1, max: Int = Int.MaxValue): Int

    Permalink
  67. def parseTimeInterval(optV: Any, module: String): Long

    Permalink

    Parse the given time interval value as long milliseconds.

    Parse the given time interval value as long milliseconds.

    See also

    timeIntervalSpec for the allowed string specification

  68. def parseTimestamp(ts: String, module: String, col: String): Long

    Permalink
  69. def qcsOf(qa: Array[String], cols: Array[String], module: String): (Array[Int], Array[String])

    Permalink
  70. def registerDriver(driver: String): Unit

    Permalink

    Register given driver class with Spark's loader.

  71. def registerDriverUrl(url: String): String

    Permalink

    Register driver for given JDBC URL and return the driver class name.

  72. def resolveQCS(options: Map[String, Any], fieldNames: Array[String], module: String): (Array[Int], Array[String])

    Permalink
  73. def resolveQCS(qcsV: Option[Any], fieldNames: Array[String], module: String): (Array[Int], Array[String])

    Permalink
  74. def schemaFields(schema: StructType): Map[String, StructField]

    Permalink
  75. def setDefaultConfProperty(conf: SparkConf, name: String, default: String): Unit

    Permalink
  76. def setDefaultSerializerAndCodec(conf: SparkConf): Unit

    Permalink
  77. def stringMetadata(md: Metadata = Metadata.empty): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is rendered as CLOB by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is rendered as CLOB by SnappyStore.

    md

    optional Metadata object to be merged into the result

    returns

    the result Metadata object to use for StructField

  78. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  79. def taskMemoryManager(context: TaskContext): TaskMemoryManager

    Permalink
  80. def toLowerCase(k: String): String

    Permalink
  81. def toOpenHashMap[K, V](map: Map[K, V]): ObjectObjectHashMap[K, V]

    Permalink
  82. def toString(): String

    Permalink
    Definition Classes
    AnyRef → Any
  83. def toUnsafeRow(buffer: ByteBuffer, numColumns: Int): UnsafeRow

    Permalink
  84. def toUpperCase(k: String): String

    Permalink
  85. def unAlias(e: Expression, childClass: Class[_] = null): Expression

    Permalink
    Annotations
    @tailrec()
  86. lazy val usingEnhancedSpark: Boolean

    Permalink
  87. def varcharMetadata(size: Int, md: Metadata): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    size

    the size parameter of the VARCHAR() column type

    md

    optional Metadata object to be merged into the result

    returns

    the result Metadata object to use for StructField

  88. def varcharMetadata(size: Int): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    size

    the size parameter of the VARCHAR() column type

    returns

    the result Metadata object to use for StructField

  89. def varcharMetadata(): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    returns

    the result Metadata object to use for StructField

  90. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  91. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  92. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  93. def withNewExecutionId[T](df: DataFrame, body: ⇒ T): T

    Permalink

    Wrap a DataFrame action to track all Spark jobs in the body so that we can connect them with an execution.

Inherited from AnyRef

Inherited from Any

Ungrouped