Object

org.apache.spark.sql.collection

Utils

Related Doc: package collection

Permalink

object Utils

Linear Supertypes
AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. Utils
  2. AnyRef
  3. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  4. final val EMPTY_STRING_ARRAY: Array[String]

    Permalink
  5. def ERROR_NO_QCS(module: String): String

    Permalink
  6. final val SKIP_ANALYSIS_PREFIX: String("SAMPLE_")

    Permalink
  7. final val WEIGHTAGE_COLUMN_NAME: String("SNAPPY_SAMPLER_WEIGHTAGE")

    Permalink
  8. final val Z95Percent: Double

    Permalink
  9. final val Z95Squared: Double

    Permalink
  10. def analysisException(msg: String, cause: Option[Throwable] = None): AnalysisException

    Permalink
  11. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  12. def charMetadata(size: Int, md: Metadata): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    size

    the size parameter of the CHAR() column type

    md

    optional Metadata object to be merged into the result

    returns

    the result Metadata object to use for StructField

  13. def charMetadata(size: Int): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    size

    the size parameter of the CHAR() column type

    returns

    the result Metadata object to use for StructField

  14. def charMetadata(): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    returns

    the result Metadata object to use for StructField

  15. def classForName(className: String): Class[_]

    Permalink
  16. def clearDefaultSerializerAndCodec(): Unit

    Permalink
  17. def clone(): AnyRef

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  18. def closeJsonGenerator(gen: AnyRef): Unit

    Permalink
  19. def columnIndex(col: String, cols: Array[String], module: String): Int

    Permalink
  20. def createCatalystConverter(dataType: DataType): (Any) ⇒ Any

    Permalink
  21. def createScalaConverter(dataType: DataType): (Any) ⇒ Any

    Permalink
  22. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  23. def equals(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  24. def executorsListener(sc: SparkContext): Option[ExecutorsListener]

    Permalink
  25. def fieldIndex(relationOutput: Seq[Attribute], columnName: String, caseSensitive: Boolean): Int

    Permalink
  26. def fillArray[T](a: Array[_ >: T], v: T, start: Int, endP1: Int): Unit

    Permalink
  27. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  28. def genTaskContextFunction(ctx: CodegenContext): String

    Permalink
  29. def generateJson(gen: AnyRef, row: InternalRow, columnIndex: Int, columnType: DataType): Unit

    Permalink
  30. def getActiveSession: Option[SparkSession]

    Permalink
  31. def getAllExecutorsMemoryStatus(sc: SparkContext): Map[BlockManagerId, (Long, Long)]

    Permalink
  32. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
  33. def getClientHostPort(netServer: String): String

    Permalink
  34. def getDriverClassName(url: String): String

    Permalink
  35. def getFields(o: Any): Map[String, Any]

    Permalink
  36. def getFixedPartitionRDD[T](sc: SparkContext, f: (TaskContext, Partition) ⇒ Iterator[T], partitioner: Partitioner, numPartitions: Int)(implicit arg0: ClassTag[T]): RDD[T]

    Permalink
  37. def getGenericRowValues(row: GenericRow): Array[Any]

    Permalink
  38. def getHostExecutorId(blockId: BlockManagerId): String

    Permalink
  39. def getInternalSparkConf(sc: SparkContext): SparkConf

    Permalink
  40. def getInternalType(dataType: DataType): Class[_]

    Permalink
  41. def getJsonGenerator(dataType: DataType, columnName: String, writer: Writer): AnyRef

    Permalink
  42. def getNumColumns(partitioning: Partitioning): Int

    Permalink
  43. def getPartitionData(blockId: BlockId, bm: BlockManager): ByteBuffer

    Permalink
  44. def getPartitions(region: Any, bucketId: Int): Array[Partition]

    Permalink
  45. def getPrunedPartition(partitionColumns: Seq[String], filters: Array[Expression], schema: StructType, numBuckets: Int, partitionColumnCount: Int): Int

    Permalink
  46. def getSQLDataType(dataType: DataType): DataType

    Permalink
  47. def getSchemaAndPlanFromBase(schemaOpt: Option[StructType], baseTableOpt: Option[String], catalog: SnappySessionCatalog, asSelect: Boolean, table: String, tableType: String): (StructType, Option[LogicalPlan])

    Permalink

    Get the result schema given an optional explicit schema and base table.

    Get the result schema given an optional explicit schema and base table. In case both are specified, then check compatibility between the two.

  48. def hasLowerCase(k: String): Boolean

    Permalink
  49. def hashCode(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  50. def immutableMap[A, B](m: Map[A, B]): Map[A, B]

    Permalink
  51. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  52. final def isLoner(sc: SparkContext): Boolean

    Permalink
  53. def mapExecutors[T](sc: SparkContext, f: (TaskContext, ExecutorLocalPartition) ⇒ Iterator[T], maxTries: Int)(implicit arg0: ClassTag[T]): Array[T]

    Permalink
  54. def mapExecutors[T](sc: SparkContext, f: () ⇒ Iterator[T], maxTries: Int = 30, blockManagerIds: Seq[BlockManagerId] = Nil)(implicit arg0: ClassTag[T]): Array[T]

    Permalink
  55. def matchOption(optName: String, options: Map[String, Any]): Option[(String, Any)]

    Permalink
  56. def metricMethods: ((String) ⇒ String, (String) ⇒ String)

    Permalink
  57. def millisToDays(millisUtc: Long, tz: TimeZone): Int

    Permalink
  58. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  59. def newChunkedByteBuffer(chunks: Array[ByteBuffer]): ChunkedByteBuffer

    Permalink
  60. def newClusterSparkConf(props: Map[AnyRef, AnyRef]): SparkConf

    Permalink
  61. def newClusterSparkConf(): SparkConf

    Permalink
  62. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
  63. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
  64. def parseColumn(cv: Any, cols: Array[String], module: String, option: String): Int

    Permalink
  65. def parseColumnsAsClob(s: String, session: SnappySession): (Boolean, Set[String])

    Permalink
  66. def parseDouble(v: Any, module: String, option: String, min: Double, max: Double, exclusive: Boolean = true): Double

    Permalink
  67. def parseInteger(v: Any, module: String, option: String, min: Int = 1, max: Int = Int.MaxValue): Int

    Permalink
  68. def parseTimeInterval(optV: Any, module: String): Long

    Permalink

    Parse the given time interval value as long milliseconds.

    Parse the given time interval value as long milliseconds.

    See also

    timeIntervalSpec for the allowed string specification

  69. def parseTimestamp(ts: String, module: String, col: String): Long

    Permalink
  70. def qcsOf(qa: Array[String], cols: Array[String], module: String): (Array[Int], Array[String])

    Permalink
  71. def registerDriver(driver: String): Unit

    Permalink

    Register given driver class with Spark's loader.

  72. def registerDriverUrl(url: String): String

    Permalink

    Register driver for given JDBC URL and return the driver class name.

  73. def resolveQCS(options: Map[String, Any], fieldNames: Array[String], module: String): (Array[Int], Array[String])

    Permalink
  74. def resolveQCS(qcsV: Option[Any], fieldNames: Array[String], module: String): (Array[Int], Array[String])

    Permalink
  75. def resultSetToSparkInternalRows(resultSet: ResultSet, schema: StructType, inputMetrics: InputMetrics = new InputMetrics): Iterator[InternalRow]

    Permalink
  76. def schemaFields(schema: StructType): Map[String, StructField]

    Permalink
  77. def setDefaultConfProperty(conf: SparkConf, name: String, default: String): Unit

    Permalink
  78. def setDefaultSerializerAndCodec(conf: SparkConf): Unit

    Permalink
  79. def sqlInternal(snappy: SnappySession, sqlText: String): CachedDataFrame

    Permalink
  80. def stringMetadata(md: Metadata = Metadata.empty): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is rendered as CLOB by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is rendered as CLOB by SnappyStore.

    md

    optional Metadata object to be merged into the result

    returns

    the result Metadata object to use for StructField

  81. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  82. def toLowerCase(k: String): String

    Permalink
  83. def toOpenHashMap[K, V](map: Map[K, V]): UnifiedMap[K, V]

    Permalink
  84. def toString(): String

    Permalink
    Definition Classes
    AnyRef → Any
  85. def toUpperCase(k: String): String

    Permalink
  86. def unAlias(e: Expression, childClass: Class[_] = null): Expression

    Permalink
    Annotations
    @tailrec()
  87. lazy val usingEnhancedSpark: Boolean

    Permalink
  88. def varcharMetadata(size: Int, md: Metadata): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    size

    the size parameter of the VARCHAR() column type

    md

    optional Metadata object to be merged into the result

    returns

    the result Metadata object to use for StructField

  89. def varcharMetadata(size: Int): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    size

    the size parameter of the VARCHAR() column type

    returns

    the result Metadata object to use for StructField

  90. def varcharMetadata(): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    returns

    the result Metadata object to use for StructField

  91. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  92. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  93. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  94. def withNewExecutionId[T](df: DataFrame, body: ⇒ T): T

    Permalink

    Wrap a DataFrame action to track all Spark jobs in the body so that we can connect them with an execution.

Inherited from AnyRef

Inherited from Any

Ungrouped