class SparkSession extends sql.SparkSession with Logging with ColumnConversions
The entry point to programming Spark with the Dataset and DataFrame API.
In environments that this has been created upfront (e.g. REPL, notebooks), use the builder to get an existing session:
SparkSession.builder().getOrCreate()
The builder can also be used to create a new session:
SparkSession.builder .master("local") .appName("Word Count") .config("spark.some.config.option", "some-value") .getOrCreate()
- Self Type
- SparkSession
- Annotations
- @Stable()
- Alphabetic
- By Inheritance
- SparkSession
- ColumnConversions
- Logging
- SparkSession
- Closeable
- AutoCloseable
- Serializable
- AnyRef
- Any
- Hide All
- Show All
- Public
- Protected
Type Members
- implicit class LogStringContext extends AnyRef
- Definition Classes
- Logging
Value Members
- final def !=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
- final def ##: Int
- Definition Classes
- AnyRef → Any
- final def ==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
- def addArtifact(source: String, target: String): Unit
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @Experimental()
- def addArtifact(bytes: Array[Byte], target: String): Unit
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @Experimental()
- def addArtifact(uri: URI): Unit
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @Experimental()
- def addArtifact(path: String): Unit
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @Experimental()
- def addArtifacts(uri: URI*): Unit
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @Experimental() @varargs()
- def addTag(tag: String): Unit
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- final def asInstanceOf[T0]: T0
- Definition Classes
- Any
- def baseRelationToDataFrame(baseRelation: BaseRelation): DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- lazy val catalog: Catalog
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @transient()
- def clearTags(): Unit
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def clone(): AnyRef
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.CloneNotSupportedException]) @IntrinsicCandidate() @native()
- def close(): Unit
Stop the underlying
SparkContext
.Stop the underlying
SparkContext
.- Definition Classes
- SparkSession → Closeable → AutoCloseable
- Since
2.1.0
- lazy val conf: RuntimeConfig
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @transient()
- val converter: ColumnNodeToExpressionConverter
- Attributes
- protected[sql]
- Definition Classes
- SparkSession → ColumnConversions
- def createDataFrame(data: List[_], beanClass: Class[_]): DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def createDataFrame(rdd: JavaRDD[_], beanClass: Class[_]): DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def createDataFrame(rdd: RDD[_], beanClass: Class[_]): DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def createDataFrame(rows: List[Row], schema: StructType): DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @DeveloperApi()
- def createDataFrame(rowRDD: JavaRDD[Row], schema: StructType): DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @DeveloperApi()
- def createDataFrame(rowRDD: RDD[Row], schema: StructType): DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @DeveloperApi()
- def createDataFrame[A <: Product](data: Seq[A])(implicit arg0: scala.reflect.api.JavaUniverse.TypeTag[A]): DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def createDataFrame[A <: Product](rdd: RDD[A])(implicit arg0: scala.reflect.api.JavaUniverse.TypeTag[A]): DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def createDataset[T](data: List[T])(implicit arg0: Encoder[T]): Dataset[T]
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def createDataset[T](data: RDD[T])(implicit arg0: Encoder[T]): Dataset[T]
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def createDataset[T](data: Seq[T])(implicit arg0: Encoder[T]): Dataset[T]
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- lazy val emptyDataFrame: DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @transient()
- def emptyDataset[T](implicit arg0: Encoder[T]): Dataset[T]
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- final def eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
- def equals(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef → Any
- def executeCommand(runner: String, command: String, options: Map[String, String]): DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @Unstable()
- def experimental: ExperimentalMethods
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @Experimental() @Unstable()
- def expression(column: Column): Expression
Convert a Column into an Expression.
Convert a Column into an Expression.
- Definition Classes
- ColumnConversions
- Annotations
- @DeveloperApi()
- final def getClass(): Class[_ <: AnyRef]
- Definition Classes
- AnyRef → Any
- Annotations
- @IntrinsicCandidate() @native()
- def getTags(): Set[String]
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @IntrinsicCandidate() @native()
- def initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
- Attributes
- protected
- Definition Classes
- Logging
- def initializeLogIfNecessary(isInterpreter: Boolean): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def interruptAll(): Seq[String]
Request to interrupt all currently running SQL operations of this session.
Request to interrupt all currently running SQL operations of this session.
- returns
Sequence of SQL execution IDs requested to be interrupted.
- Definition Classes
- SparkSession → SparkSession
- Since
4.0.0
- Note
Only DataFrame/SQL operations started by this session can be interrupted.
,This method will wait up to 60 seconds for the interruption request to be issued.
- def interruptOperation(operationId: String): Seq[String]
Request to interrupt a SQL operation of this session, given its SQL execution ID.
Request to interrupt a SQL operation of this session, given its SQL execution ID.
- returns
The execution ID requested to be interrupted, as a single-element sequence, or an empty sequence if the operation is not started by this session.
- Definition Classes
- SparkSession → SparkSession
- Since
4.0.0
- Note
Only DataFrame/SQL operations started by this session can be interrupted.
,This method will wait up to 60 seconds for the interruption request to be issued.
- def interruptTag(tag: String): Seq[String]
Request to interrupt all currently running SQL operations of this session with the given job tag.
Request to interrupt all currently running SQL operations of this session with the given job tag.
- returns
Sequence of SQL execution IDs requested to be interrupted.
- Definition Classes
- SparkSession → SparkSession
- Since
4.0.0
- Note
Only DataFrame/SQL operations started by this session can be interrupted.
,This method will wait up to 60 seconds for the interruption request to be issued.
- final def isInstanceOf[T0]: Boolean
- Definition Classes
- Any
- def isTraceEnabled(): Boolean
- Attributes
- protected
- Definition Classes
- Logging
- def listenerManager: ExecutionListenerManager
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def log: Logger
- Attributes
- protected
- Definition Classes
- Logging
- def logDebug(msg: => String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logDebug(entry: LogEntry, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logDebug(entry: LogEntry): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logDebug(msg: => String): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logError(msg: => String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logError(entry: LogEntry, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logError(entry: LogEntry): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logError(msg: => String): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logInfo(msg: => String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logInfo(entry: LogEntry, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logInfo(entry: LogEntry): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logInfo(msg: => String): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logName: String
- Attributes
- protected
- Definition Classes
- Logging
- def logTrace(msg: => String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logTrace(entry: LogEntry, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logTrace(entry: LogEntry): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logTrace(msg: => String): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logWarning(msg: => String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logWarning(entry: LogEntry, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logWarning(entry: LogEntry): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logWarning(msg: => String): Unit
- Attributes
- protected
- Definition Classes
- Logging
- final def ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
- def newSession(): SparkSession
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- final def notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @IntrinsicCandidate() @native()
- final def notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @IntrinsicCandidate() @native()
- def parseDataType(dataTypeString: String): DataType
Parses the data type in our internal string representation.
Parses the data type in our internal string representation. The data type string should have the same format as the one generated by
toString
in scala. It is only used by PySpark.- Attributes
- protected[sql]
- def range(start: Long, end: Long, step: Long, numPartitions: Int): Dataset[Long]
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def range(start: Long, end: Long, step: Long): Dataset[Long]
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def range(start: Long, end: Long): Dataset[Long]
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def range(end: Long): Dataset[Long]
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def read: DataFrameReader
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def readStream: DataStreamReader
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def removeTag(tag: String): Unit
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- lazy val sessionState: SessionState
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @Unstable() @transient()
- lazy val sharedState: SharedState
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @Unstable() @transient()
- val sparkContext: SparkContext
- Definition Classes
- SparkSession → SparkSession
- def sql(sqlText: String): DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def sql(sqlText: String, args: Map[String, Any]): DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def sql(sqlText: String, args: Map[String, Any]): DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def sql(sqlText: String, args: Array[_]): DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- val sqlContext: SQLContext
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def stop(): Unit
- Definition Classes
- SparkSession
- def streams: StreamingQueryManager
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- Annotations
- @Unstable()
- final def synchronized[T0](arg0: => T0): T0
- Definition Classes
- AnyRef
- def table(tableName: String): DataFrame
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def time[T](f: => T): T
- Definition Classes
- SparkSession
- implicit def toRichColumn(column: Column): RichColumn
Wrap a Column with a RichColumn to provide the
expr
andnamed
methods.Wrap a Column with a RichColumn to provide the
expr
andnamed
methods.- Definition Classes
- ColumnConversions
- Annotations
- @DeveloperApi()
- def toString(): String
- Definition Classes
- AnyRef → Any
- def tvf: TableValuedFunction
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def udf: UDFRegistration
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- def version: String
<invalid inheritdoc annotation>
<invalid inheritdoc annotation>
- Definition Classes
- SparkSession → SparkSession
- final def wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.InterruptedException])
- final def wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.InterruptedException]) @native()
- final def wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.InterruptedException])
- def withActive[T](block: => T): T
- Definition Classes
- SparkSession
- Annotations
- @DeveloperApi()
- def withLogContext(context: Map[String, String])(body: => Unit): Unit
- Attributes
- protected
- Definition Classes
- Logging
- object implicits extends SQLImplicits
- Definition Classes
- SparkSession → SparkSession
Deprecated Value Members
- def finalize(): Unit
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.Throwable]) @Deprecated
- Deprecated
(Since version 9)