Class/Object

org.apache.spark.sql.internal

SQLConf

Related Docs: object SQLConf | package internal

Permalink

class SQLConf extends Serializable with Logging

A class that enables the setting and getting of mutable config parameters/hints.

In the presence of a SQLContext, these can be set and queried by passing SET commands into Spark SQL's query functions (i.e. sql()). Otherwise, users of this class can modify the hints by programmatically calling the setters and getters of this class.

SQLConf is thread-safe (internally synchronized, so safe to be used in multiple threads).

Linear Supertypes
Logging, Serializable, Serializable, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. SQLConf
  2. Logging
  3. Serializable
  4. Serializable
  5. AnyRef
  6. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new SQLConf()

    Permalink

Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  4. def adaptiveExecutionEnabled: Boolean

    Permalink
  5. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  6. def autoBroadcastJoinThreshold: Long

    Permalink
  7. def broadcastTimeout: Int

    Permalink
  8. def bucketingEnabled: Boolean

    Permalink
  9. def cartesianProductExecBufferInMemoryThreshold: Int

    Permalink
  10. def cartesianProductExecBufferSpillThreshold: Int

    Permalink
  11. def caseSensitiveAnalysis: Boolean

    Permalink
  12. def caseSensitiveInferenceMode: SQLConf.HiveCaseSensitiveInferenceMode.Value

    Permalink
  13. def cboEnabled: Boolean

    Permalink
  14. def checkpointLocation: Option[String]

    Permalink
  15. def clear(): Unit

    Permalink
  16. def clone(): SQLConf

    Permalink
    Definition Classes
    SQLConf → AnyRef
  17. def columnBatchSize: Int

    Permalink
  18. def columnNameOfCorruptRecord: String

    Permalink
  19. def constraintPropagationEnabled: Boolean

    Permalink
  20. def contains(key: String): Boolean

    Permalink

    Return whether a given key is set in this SQLConf.

  21. def convertCTAS: Boolean

    Permalink
  22. def copy(entries: (ConfigEntry[_], Any)*): SQLConf

    Permalink
  23. def crossJoinEnabled: Boolean

    Permalink
  24. def dataFramePivotMaxValues: Int

    Permalink
  25. def dataFrameRetainGroupColumns: Boolean

    Permalink
  26. def dataFrameSelfJoinAutoResolveAmbiguity: Boolean

    Permalink
  27. def defaultDataSourceName: String

    Permalink
  28. def defaultSizeInBytes: Long

    Permalink
  29. def enableRadixSort: Boolean

    Permalink
  30. def enableTwoLevelAggMap: Boolean

    Permalink
  31. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  32. def equals(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  33. def escapedStringLiterals: Boolean

    Permalink
  34. def exchangeReuseEnabled: Boolean

    Permalink
  35. def fallBackToHdfsForStatsEnabled: Boolean

    Permalink
  36. def fileCommitProtocolClass: String

    Permalink
  37. def fileSinkLogCleanupDelay: Long

    Permalink
  38. def fileSinkLogCompactInterval: Int

    Permalink
  39. def fileSinkLogDeletion: Boolean

    Permalink
  40. def fileSourceLogCleanupDelay: Long

    Permalink
  41. def fileSourceLogCompactInterval: Int

    Permalink
  42. def fileSourceLogDeletion: Boolean

    Permalink
  43. def filesMaxPartitionBytes: Long

    Permalink
  44. def filesOpenCostInBytes: Long

    Permalink
  45. def filesourcePartitionFileCacheSize: Long

    Permalink
  46. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  47. def gatherFastStats: Boolean

    Permalink
  48. def getAllConfs: Map[String, String]

    Permalink

    Return all the configuration properties that have been set (i.e.

    Return all the configuration properties that have been set (i.e. not the default). This creates a new copy of the config properties in the form of a Map.

  49. def getAllDefinedConfs: Seq[(String, String, String)]

    Permalink

    Return all the configuration definitions that have been defined in SQLConf.

    Return all the configuration definitions that have been defined in SQLConf. Each definition contains key, defaultValue and doc.

  50. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
  51. def getConf[T](entry: OptionalConfigEntry[T]): Option[T]

    Permalink

    Return the value of an optional Spark SQL configuration property for the given key.

    Return the value of an optional Spark SQL configuration property for the given key. If the key is not set yet, returns None.

  52. def getConf[T](entry: ConfigEntry[T]): T

    Permalink

    Return the value of Spark SQL configuration property for the given key.

    Return the value of Spark SQL configuration property for the given key. If the key is not set yet, return defaultValue in ConfigEntry.

  53. def getConf[T](entry: ConfigEntry[T], defaultValue: T): T

    Permalink

    Return the value of Spark SQL configuration property for the given key.

    Return the value of Spark SQL configuration property for the given key. If the key is not set yet, return defaultValue. This is useful when defaultValue in ConfigEntry is not the desired one.

  54. def getConfString(key: String, defaultValue: String): String

    Permalink

    Return the string value of Spark SQL configuration property for the given key.

    Return the string value of Spark SQL configuration property for the given key. If the key is not set yet, return defaultValue.

  55. def getConfString(key: String): String

    Permalink

    Return the value of Spark SQL configuration property for the given key.

    Return the value of Spark SQL configuration property for the given key.

    Annotations
    @throws( "if key is not set" )
  56. def groupByAliases: Boolean

    Permalink
  57. def groupByOrdinal: Boolean

    Permalink
  58. def hashCode(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  59. def hiveThriftServerSingleSession: Boolean

    Permalink
  60. def ignoreCorruptFiles: Boolean

    Permalink
  61. def inMemoryPartitionPruning: Boolean

    Permalink
  62. def initializeLogIfNecessary(isInterpreter: Boolean): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  63. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  64. def isParquetBinaryAsString: Boolean

    Permalink
  65. def isParquetINT64AsTimestampMillis: Boolean

    Permalink
  66. def isParquetINT96AsTimestamp: Boolean

    Permalink
  67. def isParquetSchemaMergingEnabled: Boolean

    Permalink
  68. def isParquetSchemaRespectSummaries: Boolean

    Permalink
  69. def isTraceEnabled(): Boolean

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  70. def isUnsupportedOperationCheckEnabled: Boolean

    Permalink
  71. def joinReorderCardWeight: Double

    Permalink
  72. def joinReorderDPStarFilter: Boolean

    Permalink
  73. def joinReorderDPThreshold: Int

    Permalink
  74. def joinReorderEnabled: Boolean

    Permalink
  75. def limitScaleUpFactor: Int

    Permalink
  76. def log: Logger

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  77. def logDebug(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  78. def logDebug(msg: ⇒ String): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  79. def logError(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  80. def logError(msg: ⇒ String): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  81. def logInfo(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  82. def logInfo(msg: ⇒ String): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  83. def logName: String

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  84. def logTrace(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  85. def logTrace(msg: ⇒ String): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  86. def logWarning(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  87. def logWarning(msg: ⇒ String): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  88. def manageFilesourcePartitions: Boolean

    Permalink
  89. def maxCaseBranchesForCodegen: Int

    Permalink
  90. def maxNestedViewDepth: Int

    Permalink
  91. def maxRecordsPerFile: Long

    Permalink
  92. def metastorePartitionPruning: Boolean

    Permalink
  93. def minBatchesToRetain: Int

    Permalink
  94. def minNumPostShufflePartitions: Int

    Permalink
  95. def ndvMaxError: Double

    Permalink
  96. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  97. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
  98. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
  99. def numShufflePartitions: Int

    Permalink
  100. def objectAggSortBasedFallbackThreshold: Int

    Permalink
  101. def optimizerInSetConversionThreshold: Int

    Permalink
  102. def optimizerMaxIterations: Int

    Permalink

    ************************ Spark SQL Params/Hints *******************

  103. def optimizerMetadataOnly: Boolean

    Permalink
  104. def orcFilterPushDown: Boolean

    Permalink
  105. def orderByOrdinal: Boolean

    Permalink
  106. def parallelPartitionDiscoveryParallelism: Int

    Permalink
  107. def parallelPartitionDiscoveryThreshold: Int

    Permalink
  108. def parquetCacheMetadata: Boolean

    Permalink
  109. def parquetCompressionCodec: String

    Permalink
  110. def parquetFilterPushDown: Boolean

    Permalink
  111. def parquetOutputCommitterClass: String

    Permalink
  112. def parquetVectorizedReaderEnabled: Boolean

    Permalink
  113. def partitionColumnTypeInferenceEnabled: Boolean

    Permalink
  114. def preferSortMergeJoin: Boolean

    Permalink
  115. def redactOptions(options: Map[String, String]): Map[String, String]

    Permalink

    Redacts the given option map according to the description of SQL_OPTIONS_REDACTION_PATTERN.

  116. def resolver: Resolver

    Permalink

    Returns the Resolver for the current configuration, which can be used to determine if two identifiers are equal.

  117. def runSQLonFile: Boolean

    Permalink
  118. def sessionLocalTimeZone: String

    Permalink
  119. def setConf[T](entry: ConfigEntry[T], value: T): Unit

    Permalink

    Set the given Spark SQL configuration property.

  120. def setConf(props: Properties): Unit

    Permalink

    Set Spark SQL configuration properties.

  121. def setConfString(key: String, value: String): Unit

    Permalink

    Set the given Spark SQL configuration property using a string value.

  122. val settings: Map[String, String]

    Permalink

    Only low degree of contention is expected for conf, thus NOT using ConcurrentHashMap.

    Only low degree of contention is expected for conf, thus NOT using ConcurrentHashMap.

    Attributes
    protected[org.apache.spark]
  123. def sortMergeJoinExecBufferInMemoryThreshold: Int

    Permalink
  124. def sortMergeJoinExecBufferSpillThreshold: Int

    Permalink
  125. def starSchemaDetection: Boolean

    Permalink
  126. def starSchemaFTRatio: Double

    Permalink
  127. def stateStoreMinDeltasForSnapshot: Int

    Permalink
  128. def streamingFileCommitProtocolClass: String

    Permalink
  129. def streamingMetricsEnabled: Boolean

    Permalink
  130. def streamingNoDataProgressEventInterval: Long

    Permalink
  131. def streamingPollingDelay: Long

    Permalink
  132. def streamingProgressRetention: Int

    Permalink
  133. def streamingSchemaInference: Boolean

    Permalink
  134. def subexpressionEliminationEnabled: Boolean

    Permalink
  135. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  136. def tableRelationCacheSize: Int

    Permalink
  137. def targetPostShuffleInputSize: Long

    Permalink
  138. def toString(): String

    Permalink
    Definition Classes
    AnyRef → Any
  139. def unsetConf(entry: ConfigEntry[_]): Unit

    Permalink
  140. def unsetConf(key: String): Unit

    Permalink
  141. def useCompression: Boolean

    Permalink
  142. def useObjectHashAggregation: Boolean

    Permalink
  143. def variableSubstituteDepth: Int

    Permalink
  144. def variableSubstituteEnabled: Boolean

    Permalink
  145. def verifyPartitionPath: Boolean

    Permalink
  146. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  147. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  148. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  149. def warehousePath: String

    Permalink
  150. def wholeStageEnabled: Boolean

    Permalink
  151. def wholeStageFallback: Boolean

    Permalink
  152. def wholeStageMaxNumFields: Int

    Permalink
  153. def windowExecBufferInMemoryThreshold: Int

    Permalink
  154. def windowExecBufferSpillThreshold: Int

    Permalink
  155. def writeLegacyParquetFormat: Boolean

    Permalink

Inherited from Logging

Inherited from Serializable

Inherited from Serializable

Inherited from AnyRef

Inherited from Any

Ungrouped