Package

org.apache.spark.sql.execution

datasources

Permalink

package datasources

Visibility
  1. Public
  2. All

Type Members

  1. class CaseInsensitiveMap extends Map[String, String] with Serializable

    Permalink

    Builds a map in which keys are case insensitive

  2. case class CreateTableUsing(tableIdent: TableIdentifier, userSpecifiedSchema: Option[StructType], provider: String, temporary: Boolean, options: Map[String, String], allowExisting: Boolean, managedIfNoPath: Boolean) extends LogicalPlan with Command with Product with Serializable

    Permalink

    Used to represent the operation of create table using a data source.

    Used to represent the operation of create table using a data source.

    allowExisting

    If it is true, we will do nothing when the table already exists. If it is false, an exception will be thrown

  3. case class CreateTableUsingAsSelect(tableIdent: TableIdentifier, provider: String, temporary: Boolean, partitionColumns: Array[String], mode: SaveMode, options: Map[String, String], child: LogicalPlan) extends catalyst.plans.logical.UnaryNode with Product with Serializable

    Permalink

    A node used to support CTAS statements and saveAsTable for the data source API.

    A node used to support CTAS statements and saveAsTable for the data source API. This node is a UnaryNode instead of a Command because we want the analyzer can analyze the logical plan that will be used to populate the table. So, PreWriteCheck can detect cases that are not allowed.

  4. case class CreateTempTableUsing(tableIdent: TableIdentifier, userSpecifiedSchema: Option[StructType], provider: String, options: Map[String, String]) extends LogicalPlan with RunnableCommand with Product with Serializable

    Permalink
  5. case class CreateTempTableUsingAsSelect(tableIdent: TableIdentifier, provider: String, partitionColumns: Array[String], mode: SaveMode, options: Map[String, String], query: LogicalPlan) extends LogicalPlan with RunnableCommand with Product with Serializable

    Permalink
  6. class DDLException extends RuntimeException

    Permalink

    The exception thrown from the DDL parser.

  7. class DDLParser extends AbstractSparkSQLParser with DataTypeParser with Logging

    Permalink

    A parser for foreign DDL commands.

  8. class DefaultSource extends RelationProvider with DataSourceRegister

    Permalink
  9. case class DescribeCommand(table: LogicalPlan, isExtended: Boolean) extends LogicalPlan with Command with Product with Serializable

    Permalink

    Returned for the "DESCRIBE [EXTENDED] [dbName.]tableName" command.

    Returned for the "DESCRIBE [EXTENDED] [dbName.]tableName" command.

    table

    The table to be described.

    isExtended

    True if "DESCRIBE EXTENDED" is used. Otherwise, false. It is effective only when the table is a Hive table.

  10. case class RefreshTable(tableIdent: TableIdentifier) extends LogicalPlan with RunnableCommand with Product with Serializable

    Permalink
  11. case class ResolvedDataSource(provider: Class[_], relation: BaseRelation) extends Product with Serializable

    Permalink

Value Members

  1. object ResolvedDataSource extends Logging with Serializable

    Permalink
  2. package jdbc

    Permalink
  3. package json

    Permalink

Ungrouped