class CacheManager extends Logging with AdaptiveSparkPlanHelper
Provides support in a SQLContext for caching query results and automatically using these cached
results when subsequent queries are executed. Data is cached using byte buffers stored in an
InMemoryRelation. This relation is automatically substituted query plans that return the
sameResult
as the originally cached query.
Internal to Spark SQL.
- Alphabetic
- By Inheritance
- CacheManager
- AdaptiveSparkPlanHelper
- Logging
- AnyRef
- Any
- Hide All
- Show All
- Public
- All
Instance Constructors
- new CacheManager()
Value Members
-
final
def
!=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
##(): Int
- Definition Classes
- AnyRef → Any
-
final
def
==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
asInstanceOf[T0]: T0
- Definition Classes
- Any
-
def
cacheQuery(query: Dataset[_], tableName: Option[String] = None, storageLevel: StorageLevel = MEMORY_AND_DISK): Unit
Caches the data produced by the logical representation of the given Dataset.
Caches the data produced by the logical representation of the given Dataset. Unlike
RDD.cache()
, the default storage level is set to beMEMORY_AND_DISK
because recomputing the in-memory columnar representation of the underlying table is expensive. -
def
clearCache(): Unit
Clears all cached tables.
-
def
clone(): AnyRef
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()
-
def
collect[B](p: SparkPlan)(pf: PartialFunction[SparkPlan, B]): Seq[B]
Returns a Seq containing the result of applying a partial function to all elements in this tree on which the function is defined.
Returns a Seq containing the result of applying a partial function to all elements in this tree on which the function is defined.
- Definition Classes
- AdaptiveSparkPlanHelper
-
def
collectFirst[B](p: SparkPlan)(pf: PartialFunction[SparkPlan, B]): Option[B]
Finds and returns the first SparkPlan of the tree for which the given partial function is defined (pre-order), and applies the partial function to it.
Finds and returns the first SparkPlan of the tree for which the given partial function is defined (pre-order), and applies the partial function to it.
- Definition Classes
- AdaptiveSparkPlanHelper
-
def
collectLeaves(p: SparkPlan): Seq[SparkPlan]
Returns a Seq containing the leaves in this tree.
Returns a Seq containing the leaves in this tree.
- Definition Classes
- AdaptiveSparkPlanHelper
-
def
collectWithSubqueries[B](p: SparkPlan)(f: PartialFunction[SparkPlan, B]): Seq[B]
Returns a sequence containing the result of applying a partial function to all elements in this plan, also considering all the plans in its (nested) subqueries
Returns a sequence containing the result of applying a partial function to all elements in this plan, also considering all the plans in its (nested) subqueries
- Definition Classes
- AdaptiveSparkPlanHelper
-
final
def
eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
def
equals(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
def
finalize(): Unit
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( classOf[java.lang.Throwable] )
-
def
find(p: SparkPlan)(f: (SparkPlan) ⇒ Boolean): Option[SparkPlan]
Find the first SparkPlan that satisfies the condition specified by
f
.Find the first SparkPlan that satisfies the condition specified by
f
. The condition is recursively applied to this node and all of its children (pre-order).- Definition Classes
- AdaptiveSparkPlanHelper
-
def
flatMap[A](p: SparkPlan)(f: (SparkPlan) ⇒ TraversableOnce[A]): Seq[A]
Returns a Seq by applying a function to all nodes in this tree and using the elements of the resulting collections.
Returns a Seq by applying a function to all nodes in this tree and using the elements of the resulting collections.
- Definition Classes
- AdaptiveSparkPlanHelper
-
def
foreach(p: SparkPlan)(f: (SparkPlan) ⇒ Unit): Unit
Runs the given function on this node and then recursively on children.
Runs the given function on this node and then recursively on children.
- f
the function to be applied to each node in the tree.
- Definition Classes
- AdaptiveSparkPlanHelper
-
def
foreachUp(p: SparkPlan)(f: (SparkPlan) ⇒ Unit): Unit
Runs the given function recursively on children then on this node.
Runs the given function recursively on children then on this node.
- f
the function to be applied to each node in the tree.
- Definition Classes
- AdaptiveSparkPlanHelper
-
final
def
getClass(): Class[_]
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
def
getOrCloneSessionWithAqeOff[T](session: SparkSession): SparkSession
Returns a cloned SparkSession with adaptive execution disabled, or the original SparkSession if its adaptive execution is already disabled.
Returns a cloned SparkSession with adaptive execution disabled, or the original SparkSession if its adaptive execution is already disabled.
- Definition Classes
- AdaptiveSparkPlanHelper
-
def
hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
def
initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
- Attributes
- protected
- Definition Classes
- Logging
-
def
initializeLogIfNecessary(isInterpreter: Boolean): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
isEmpty: Boolean
Checks if the cache is empty.
-
final
def
isInstanceOf[T0]: Boolean
- Definition Classes
- Any
-
def
isTraceEnabled(): Boolean
- Attributes
- protected
- Definition Classes
- Logging
-
def
log: Logger
- Attributes
- protected
- Definition Classes
- Logging
-
def
logDebug(msg: ⇒ String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logDebug(msg: ⇒ String): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logError(msg: ⇒ String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logError(msg: ⇒ String): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logInfo(msg: ⇒ String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logInfo(msg: ⇒ String): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logName: String
- Attributes
- protected
- Definition Classes
- Logging
-
def
logTrace(msg: ⇒ String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logTrace(msg: ⇒ String): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logWarning(msg: ⇒ String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logWarning(msg: ⇒ String): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
lookupCachedData(plan: LogicalPlan): Option[CachedData]
Optionally returns cached data for the given LogicalPlan.
-
def
lookupCachedData(query: Dataset[_]): Option[CachedData]
Optionally returns cached data for the given Dataset
-
def
mapPlans[A](p: SparkPlan)(f: (SparkPlan) ⇒ A): Seq[A]
Returns a Seq containing the result of applying the given function to each node in this tree in a preorder traversal.
Returns a Seq containing the result of applying the given function to each node in this tree in a preorder traversal.
- f
the function to be applied.
- Definition Classes
- AdaptiveSparkPlanHelper
-
final
def
ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
final
def
notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
final
def
notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
def
recacheByPath(spark: SparkSession, resourcePath: Path, fs: FileSystem): Unit
Tries to re-cache all the cache entries that contain
resourcePath
in one or moreHadoopFsRelation
node(s) as part of its logical plan. -
def
recacheByPath(spark: SparkSession, resourcePath: String): Unit
Tries to re-cache all the cache entries that contain
resourcePath
in one or moreHadoopFsRelation
node(s) as part of its logical plan. -
def
recacheByPlan(spark: SparkSession, plan: LogicalPlan): Unit
Tries to re-cache all the cache entries that refer to the given plan.
-
def
stripAQEPlan(p: SparkPlan): SparkPlan
Strip the executePlan of AdaptiveSparkPlanExec leaf node.
Strip the executePlan of AdaptiveSparkPlanExec leaf node.
- Definition Classes
- AdaptiveSparkPlanHelper
-
def
subqueriesAll(p: SparkPlan): Seq[SparkPlan]
Returns a sequence containing the subqueries in this plan, also including the (nested) subquries in its children
Returns a sequence containing the subqueries in this plan, also including the (nested) subquries in its children
- Definition Classes
- AdaptiveSparkPlanHelper
-
final
def
synchronized[T0](arg0: ⇒ T0): T0
- Definition Classes
- AnyRef
-
def
toString(): String
- Definition Classes
- AnyRef → Any
-
def
uncacheQuery(spark: SparkSession, plan: LogicalPlan, cascade: Boolean, blocking: Boolean = false): Unit
Un-cache the given plan or all the cache entries that refer to the given plan.
Un-cache the given plan or all the cache entries that refer to the given plan.
- spark
The Spark session.
- plan
The plan to be un-cached.
- cascade
If true, un-cache all the cache entries that refer to the given plan; otherwise un-cache the given plan only.
- blocking
Whether to block until all blocks are deleted.
-
def
uncacheQuery(query: Dataset[_], cascade: Boolean): Unit
Un-cache the given plan or all the cache entries that refer to the given plan.
-
def
useCachedData(plan: LogicalPlan): LogicalPlan
Replaces segments of the given logical plan with cached versions where possible.
-
final
def
wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()