Class

org.apache.spark.sql.execution

SparkPlanner

Related Doc: package execution

Permalink

class SparkPlanner extends SparkStrategies

Linear Supertypes
SparkStrategies, QueryPlanner[SparkPlan], AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. SparkPlanner
  2. SparkStrategies
  3. QueryPlanner
  4. AnyRef
  5. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new SparkPlanner(sparkContext: SparkContext, conf: SQLConf, extraStrategies: Seq[Strategy])

    Permalink

Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  4. object Aggregation extends Strategy

    Permalink

    Used to plan the aggregate operator for expressions based on the AggregateFunction2 interface.

    Used to plan the aggregate operator for expressions based on the AggregateFunction2 interface.

    Definition Classes
    SparkStrategies
  5. object BasicOperators extends Strategy

    Permalink
    Definition Classes
    SparkStrategies
  6. object DDLStrategy extends Strategy

    Permalink
    Definition Classes
    SparkStrategies
  7. object InMemoryScans extends Strategy

    Permalink
    Definition Classes
    SparkStrategies
  8. object JoinSelection extends Strategy with PredicateHelper

    Permalink

    Select the proper physical plan for join based on joining keys and size of logical plan.

    Select the proper physical plan for join based on joining keys and size of logical plan.

    At first, uses the ExtractEquiJoinKeys pattern to find joins where at least some of the predicates can be evaluated by matching join keys. If found, Join implementations are chosen with the following precedence:

    - Broadcast: if one side of the join has an estimated physical size that is smaller than the user-configurable SQLConf.AUTO_BROADCASTJOIN_THRESHOLD threshold or if that side has an explicit broadcast hint (e.g. the user applied the org.apache.spark.sql.functions.broadcast() function to a DataFrame), then that side of the join will be broadcasted and the other side will be streamed, with no shuffling performed. If both sides of the join are eligible to be broadcasted then the - Shuffle hash join: if the average size of a single partition is small enough to build a hash table. - Sort merge: if the matching join keys are sortable.

    If there is no joining keys, Join implementations are chosen with the following precedence: - BroadcastNestedLoopJoin: if one side of the join could be broadcasted - CartesianProduct: for Inner join - BroadcastNestedLoopJoin

    Definition Classes
    SparkStrategies
  9. object SpecialLimits extends Strategy

    Permalink

    Plans special cases of limit operators.

    Plans special cases of limit operators.

    Definition Classes
    SparkStrategies
  10. object StatefulAggregationStrategy extends Strategy

    Permalink

    Used to plan aggregation queries that are computed incrementally as part of a StreamingQuery.

    Used to plan aggregation queries that are computed incrementally as part of a StreamingQuery. Currently this rule is injected into the planner on-demand, only when planning in a org.apache.spark.sql.execution.streaming.StreamExecution

    Definition Classes
    SparkStrategies
  11. object StreamingRelationStrategy extends Strategy

    Permalink

    This strategy is just for explaining Dataset/DataFrame created by spark.readStream.

    This strategy is just for explaining Dataset/DataFrame created by spark.readStream. It won't affect the execution, because StreamingRelation will be replaced with StreamingExecutionRelation in StreamingQueryManager and StreamingExecutionRelation will be replaced with the real relation using the Source in StreamExecution.

    Definition Classes
    SparkStrategies
  12. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  13. def clone(): AnyRef

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  14. val conf: SQLConf

    Permalink
  15. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  16. def equals(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  17. val extraStrategies: Seq[Strategy]

    Permalink
  18. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  19. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
  20. def hashCode(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  21. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  22. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  23. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
  24. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
  25. def numPartitions: Int

    Permalink
  26. def plan(plan: LogicalPlan): Iterator[SparkPlan]

    Permalink
    Definition Classes
    SparkPlanner → QueryPlanner
  27. def pruneFilterProject(projectList: Seq[NamedExpression], filterPredicates: Seq[Expression], prunePushedDownFilters: (Seq[Expression]) ⇒ Seq[Expression], scanBuilder: (Seq[Attribute]) ⇒ SparkPlan): SparkPlan

    Permalink

    Used to build table scan operators where complex projection and filtering are done using separate physical operators.

    Used to build table scan operators where complex projection and filtering are done using separate physical operators. This function returns the given scan operator with Project and Filter nodes added only when needed. For example, a Project operator is only used when the final desired output requires complex expressions to be evaluated or when columns can be further eliminated out after filtering has been done.

    The prunePushedDownFilters parameter is used to remove those filters that can be optimized away by the filter pushdown optimization.

    The required attributes for both filtering and expression evaluation are passed to the provided scanBuilder function so that it can avoid unnecessary column materialization.

  28. lazy val singleRowRdd: RDD[InternalRow]

    Permalink
    Attributes
    protected
    Definition Classes
    SparkStrategies
  29. val sparkContext: SparkContext

    Permalink
  30. def strategies: Seq[Strategy]

    Permalink
    Definition Classes
    SparkPlanner → QueryPlanner
  31. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  32. def toString(): String

    Permalink
    Definition Classes
    AnyRef → Any
  33. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  34. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  35. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )

Inherited from SparkStrategies

Inherited from QueryPlanner[SparkPlan]

Inherited from AnyRef

Inherited from Any

Ungrouped