class FileScanRDD extends RDD[InternalRow]
Value Members
- final def !=(arg0: Any): Boolean
- final def ##: Int
- def ++(other: RDD[InternalRow]): RDD[InternalRow]
- final def ==(arg0: Any): Boolean
- def aggregate[U](zeroValue: U)(seqOp: (U, InternalRow) => U, combOp: (U, U) => U)(implicit arg0: ClassTag[U]): U
- final def asInstanceOf[T0]: T0
- def barrier(): RDDBarrier[InternalRow]
- def cache(): FileScanRDD.this.type
- def cartesian[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(InternalRow, U)]
- def checkpoint(): Unit
- def cleanShuffleDependencies(blocking: Boolean): Unit
- def clearDependencies(): Unit
- def clone(): AnyRef
- def coalesce(numPartitions: Int, shuffle: Boolean, partitionCoalescer: Option[PartitionCoalescer])(implicit ord: Ordering[InternalRow]): RDD[InternalRow]
- def collect[U](f: PartialFunction[InternalRow, U])(implicit arg0: ClassTag[U]): RDD[U]
- def collect(): Array[InternalRow]
- def compute(split: Partition, context: TaskContext): Iterator[InternalRow]
- def context: SparkContext
- def count(): Long
- def countApprox(timeout: Long, confidence: Double): PartialResult[BoundedDouble]
- def countApproxDistinct(relativeSD: Double): Long
- def countApproxDistinct(p: Int, sp: Int): Long
- def countByValue()(implicit ord: Ordering[InternalRow]): Map[InternalRow, Long]
- def countByValueApprox(timeout: Long, confidence: Double)(implicit ord: Ordering[InternalRow]): PartialResult[Map[InternalRow, BoundedDouble]]
- final def dependencies: Seq[Dependency[_]]
- def distinct(): RDD[InternalRow]
- def distinct(numPartitions: Int)(implicit ord: Ordering[InternalRow]): RDD[InternalRow]
- final def eq(arg0: AnyRef): Boolean
- def equals(arg0: AnyRef): Boolean
- val filePartitions: Seq[FilePartition]
- def filter(f: (InternalRow) => Boolean): RDD[InternalRow]
- def finalize(): Unit
- def first(): InternalRow
- def firstParent[U](implicit arg0: ClassTag[U]): RDD[U]
- def flatMap[U](f: (InternalRow) => TraversableOnce[U])(implicit arg0: ClassTag[U]): RDD[U]
- def fold(zeroValue: InternalRow)(op: (InternalRow, InternalRow) => InternalRow): InternalRow
- def foreach(f: (InternalRow) => Unit): Unit
- def foreachPartition(f: (Iterator[InternalRow]) => Unit): Unit
- def getCheckpointFile: Option[String]
- final def getClass(): Class[_ <: AnyRef]
- def getDependencies: Seq[Dependency[_]]
- final def getNumPartitions: Int
- def getOutputDeterministicLevel: rdd.DeterministicLevel.Value
- def getPartitions: Array[Partition]
- def getPreferredLocations(split: Partition): Seq[String]
- def getResourceProfile(): ResourceProfile
- def getStorageLevel: StorageLevel
- def glom(): RDD[Array[InternalRow]]
- def groupBy[K](f: (InternalRow) => K, p: Partitioner)(implicit kt: ClassTag[K], ord: Ordering[K]): RDD[(K, Iterable[InternalRow])]
- def groupBy[K](f: (InternalRow) => K, numPartitions: Int)(implicit kt: ClassTag[K]): RDD[(K, Iterable[InternalRow])]
- def groupBy[K](f: (InternalRow) => K)(implicit kt: ClassTag[K]): RDD[(K, Iterable[InternalRow])]
- def hashCode(): Int
- val id: Int
- def initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
- def initializeLogIfNecessary(isInterpreter: Boolean): Unit
- def intersection(other: RDD[InternalRow], numPartitions: Int): RDD[InternalRow]
- def intersection(other: RDD[InternalRow], partitioner: Partitioner)(implicit ord: Ordering[InternalRow]): RDD[InternalRow]
- def intersection(other: RDD[InternalRow]): RDD[InternalRow]
- lazy val isBarrier_: Boolean
- def isCheckpointed: Boolean
- def isEmpty(): Boolean
- final def isInstanceOf[T0]: Boolean
- def isTraceEnabled(): Boolean
- final def iterator(split: Partition, context: TaskContext): Iterator[InternalRow]
- def keyBy[K](f: (InternalRow) => K): RDD[(K, InternalRow)]
- def localCheckpoint(): FileScanRDD.this.type
- def log: Logger
- def logDebug(msg: => String, throwable: Throwable): Unit
- def logDebug(msg: => String): Unit
- def logError(msg: => String, throwable: Throwable): Unit
- def logError(msg: => String): Unit
- def logInfo(msg: => String, throwable: Throwable): Unit
- def logInfo(msg: => String): Unit
- def logName: String
- def logTrace(msg: => String, throwable: Throwable): Unit
- def logTrace(msg: => String): Unit
- def logWarning(msg: => String, throwable: Throwable): Unit
- def logWarning(msg: => String): Unit
- def map[U](f: (InternalRow) => U)(implicit arg0: ClassTag[U]): RDD[U]
- def mapPartitions[U](f: (Iterator[InternalRow]) => Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]
- def mapPartitionsWithEvaluator[U](evaluatorFactory: PartitionEvaluatorFactory[InternalRow, U])(implicit arg0: ClassTag[U]): RDD[U]
- def mapPartitionsWithIndex[U](f: (Int, Iterator[InternalRow]) => Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]
- def max()(implicit ord: Ordering[InternalRow]): InternalRow
- val metadataColumns: Seq[AttributeReference]
- def min()(implicit ord: Ordering[InternalRow]): InternalRow
- var name: String
- final def ne(arg0: AnyRef): Boolean
- final def notify(): Unit
- final def notifyAll(): Unit
- def parent[U](j: Int)(implicit arg0: ClassTag[U]): RDD[U]
- val partitioner: Option[Partitioner]
- final def partitions: Array[Partition]
- def persist(): FileScanRDD.this.type
- def persist(newLevel: StorageLevel): FileScanRDD.this.type
- def pipe(command: Seq[String], env: Map[String, String], printPipeContext: ((String) => Unit) => Unit, printRDDElement: (InternalRow, (String) => Unit) => Unit, separateWorkingDir: Boolean, bufferSize: Int, encoding: String): RDD[String]
- def pipe(command: String, env: Map[String, String]): RDD[String]
- def pipe(command: String): RDD[String]
- final def preferredLocations(split: Partition): Seq[String]
- def randomSplit(weights: Array[Double], seed: Long): Array[RDD[InternalRow]]
- val readSchema: StructType
- def reduce(f: (InternalRow, InternalRow) => InternalRow): InternalRow
- def repartition(numPartitions: Int)(implicit ord: Ordering[InternalRow]): RDD[InternalRow]
- def sample(withReplacement: Boolean, fraction: Double, seed: Long): RDD[InternalRow]
- def saveAsObjectFile(path: String): Unit
- def saveAsTextFile(path: String, codec: Class[_ <: CompressionCodec]): Unit
- def saveAsTextFile(path: String): Unit
- def setName(_name: String): FileScanRDD.this.type
- def sortBy[K](f: (InternalRow) => K, ascending: Boolean, numPartitions: Int)(implicit ord: Ordering[K], ctag: ClassTag[K]): RDD[InternalRow]
- def sparkContext: SparkContext
- def subtract(other: RDD[InternalRow], p: Partitioner)(implicit ord: Ordering[InternalRow]): RDD[InternalRow]
- def subtract(other: RDD[InternalRow], numPartitions: Int): RDD[InternalRow]
- def subtract(other: RDD[InternalRow]): RDD[InternalRow]
- final def synchronized[T0](arg0: => T0): T0
- def take(num: Int): Array[InternalRow]
- def takeOrdered(num: Int)(implicit ord: Ordering[InternalRow]): Array[InternalRow]
- def takeSample(withReplacement: Boolean, num: Int, seed: Long): Array[InternalRow]
- def toDebugString: String
- def toJavaRDD(): JavaRDD[InternalRow]
- def toLocalIterator: Iterator[InternalRow]
- def toString(): String
- def top(num: Int)(implicit ord: Ordering[InternalRow]): Array[InternalRow]
- def treeAggregate[U](zeroValue: U, seqOp: (U, InternalRow) => U, combOp: (U, U) => U, depth: Int, finalAggregateOnExecutor: Boolean)(implicit arg0: ClassTag[U]): U
- def treeAggregate[U](zeroValue: U)(seqOp: (U, InternalRow) => U, combOp: (U, U) => U, depth: Int)(implicit arg0: ClassTag[U]): U
- def treeReduce(f: (InternalRow, InternalRow) => InternalRow, depth: Int): InternalRow
- def union(other: RDD[InternalRow]): RDD[InternalRow]
- def unpersist(blocking: Boolean): FileScanRDD.this.type
- final def wait(): Unit
- final def wait(arg0: Long, arg1: Int): Unit
- final def wait(arg0: Long): Unit
- def withResources(rp: ResourceProfile): FileScanRDD.this.type
- def zip[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(InternalRow, U)]
- def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D])(f: (Iterator[InternalRow], Iterator[B], Iterator[C], Iterator[D]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]
- def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D], preservesPartitioning: Boolean)(f: (Iterator[InternalRow], Iterator[B], Iterator[C], Iterator[D]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]
- def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C])(f: (Iterator[InternalRow], Iterator[B], Iterator[C]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]
- def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C], preservesPartitioning: Boolean)(f: (Iterator[InternalRow], Iterator[B], Iterator[C]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]
- def zipPartitions[B, V](rdd2: RDD[B])(f: (Iterator[InternalRow], Iterator[B]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]
- def zipPartitions[B, V](rdd2: RDD[B], preservesPartitioning: Boolean)(f: (Iterator[InternalRow], Iterator[B]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]
- def zipPartitionsWithEvaluator[U](rdd2: RDD[InternalRow], evaluatorFactory: PartitionEvaluatorFactory[InternalRow, U])(implicit arg0: ClassTag[U]): RDD[U]
- def zipWithIndex(): RDD[(InternalRow, Long)]
- def zipWithUniqueId(): RDD[(InternalRow, Long)]
Inherited from RDD[InternalRow]
Value Members
- def ++(other: RDD[InternalRow]): RDD[InternalRow]
- def aggregate[U](zeroValue: U)(seqOp: (U, InternalRow) => U, combOp: (U, U) => U)(implicit arg0: ClassTag[U]): U
- def barrier(): RDDBarrier[InternalRow]
- def cache(): FileScanRDD.this.type
- def cartesian[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(InternalRow, U)]
- def checkpoint(): Unit
- def cleanShuffleDependencies(blocking: Boolean): Unit
- def clearDependencies(): Unit
- def coalesce(numPartitions: Int, shuffle: Boolean, partitionCoalescer: Option[PartitionCoalescer])(implicit ord: Ordering[InternalRow]): RDD[InternalRow]
- def collect[U](f: PartialFunction[InternalRow, U])(implicit arg0: ClassTag[U]): RDD[U]
- def collect(): Array[InternalRow]
- def context: SparkContext
- def count(): Long
- def countApprox(timeout: Long, confidence: Double): PartialResult[BoundedDouble]
- def countApproxDistinct(relativeSD: Double): Long
- def countApproxDistinct(p: Int, sp: Int): Long
- def countByValue()(implicit ord: Ordering[InternalRow]): Map[InternalRow, Long]
- def countByValueApprox(timeout: Long, confidence: Double)(implicit ord: Ordering[InternalRow]): PartialResult[Map[InternalRow, BoundedDouble]]
- final def dependencies: Seq[Dependency[_]]
- def distinct(): RDD[InternalRow]
- def distinct(numPartitions: Int)(implicit ord: Ordering[InternalRow]): RDD[InternalRow]
- def filter(f: (InternalRow) => Boolean): RDD[InternalRow]
- def first(): InternalRow
- def firstParent[U](implicit arg0: ClassTag[U]): RDD[U]
- def flatMap[U](f: (InternalRow) => TraversableOnce[U])(implicit arg0: ClassTag[U]): RDD[U]
- def fold(zeroValue: InternalRow)(op: (InternalRow, InternalRow) => InternalRow): InternalRow
- def foreach(f: (InternalRow) => Unit): Unit
- def foreachPartition(f: (Iterator[InternalRow]) => Unit): Unit
- def getCheckpointFile: Option[String]
- def getDependencies: Seq[Dependency[_]]
- final def getNumPartitions: Int
- def getOutputDeterministicLevel: rdd.DeterministicLevel.Value
- def getResourceProfile(): ResourceProfile
- def getStorageLevel: StorageLevel
- def glom(): RDD[Array[InternalRow]]
- def groupBy[K](f: (InternalRow) => K, p: Partitioner)(implicit kt: ClassTag[K], ord: Ordering[K]): RDD[(K, Iterable[InternalRow])]
- def groupBy[K](f: (InternalRow) => K, numPartitions: Int)(implicit kt: ClassTag[K]): RDD[(K, Iterable[InternalRow])]
- def groupBy[K](f: (InternalRow) => K)(implicit kt: ClassTag[K]): RDD[(K, Iterable[InternalRow])]
- val id: Int
- def intersection(other: RDD[InternalRow], numPartitions: Int): RDD[InternalRow]
- def intersection(other: RDD[InternalRow], partitioner: Partitioner)(implicit ord: Ordering[InternalRow]): RDD[InternalRow]
- def intersection(other: RDD[InternalRow]): RDD[InternalRow]
- lazy val isBarrier_: Boolean
- def isCheckpointed: Boolean
- def isEmpty(): Boolean
- final def iterator(split: Partition, context: TaskContext): Iterator[InternalRow]
- def keyBy[K](f: (InternalRow) => K): RDD[(K, InternalRow)]
- def localCheckpoint(): FileScanRDD.this.type
- def map[U](f: (InternalRow) => U)(implicit arg0: ClassTag[U]): RDD[U]
- def mapPartitions[U](f: (Iterator[InternalRow]) => Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]
- def mapPartitionsWithEvaluator[U](evaluatorFactory: PartitionEvaluatorFactory[InternalRow, U])(implicit arg0: ClassTag[U]): RDD[U]
- def mapPartitionsWithIndex[U](f: (Int, Iterator[InternalRow]) => Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]
- def max()(implicit ord: Ordering[InternalRow]): InternalRow
- def min()(implicit ord: Ordering[InternalRow]): InternalRow
- var name: String
- def parent[U](j: Int)(implicit arg0: ClassTag[U]): RDD[U]
- val partitioner: Option[Partitioner]
- final def partitions: Array[Partition]
- def persist(): FileScanRDD.this.type
- def persist(newLevel: StorageLevel): FileScanRDD.this.type
- def pipe(command: Seq[String], env: Map[String, String], printPipeContext: ((String) => Unit) => Unit, printRDDElement: (InternalRow, (String) => Unit) => Unit, separateWorkingDir: Boolean, bufferSize: Int, encoding: String): RDD[String]
- def pipe(command: String, env: Map[String, String]): RDD[String]
- def pipe(command: String): RDD[String]
- final def preferredLocations(split: Partition): Seq[String]
- def randomSplit(weights: Array[Double], seed: Long): Array[RDD[InternalRow]]
- def reduce(f: (InternalRow, InternalRow) => InternalRow): InternalRow
- def repartition(numPartitions: Int)(implicit ord: Ordering[InternalRow]): RDD[InternalRow]
- def sample(withReplacement: Boolean, fraction: Double, seed: Long): RDD[InternalRow]
- def saveAsObjectFile(path: String): Unit
- def saveAsTextFile(path: String, codec: Class[_ <: CompressionCodec]): Unit
- def saveAsTextFile(path: String): Unit
- def setName(_name: String): FileScanRDD.this.type
- def sortBy[K](f: (InternalRow) => K, ascending: Boolean, numPartitions: Int)(implicit ord: Ordering[K], ctag: ClassTag[K]): RDD[InternalRow]
- def sparkContext: SparkContext
- def subtract(other: RDD[InternalRow], p: Partitioner)(implicit ord: Ordering[InternalRow]): RDD[InternalRow]
- def subtract(other: RDD[InternalRow], numPartitions: Int): RDD[InternalRow]
- def subtract(other: RDD[InternalRow]): RDD[InternalRow]
- def take(num: Int): Array[InternalRow]
- def takeOrdered(num: Int)(implicit ord: Ordering[InternalRow]): Array[InternalRow]
- def takeSample(withReplacement: Boolean, num: Int, seed: Long): Array[InternalRow]
- def toDebugString: String
- def toJavaRDD(): JavaRDD[InternalRow]
- def toLocalIterator: Iterator[InternalRow]
- def toString(): String
- def top(num: Int)(implicit ord: Ordering[InternalRow]): Array[InternalRow]
- def treeAggregate[U](zeroValue: U, seqOp: (U, InternalRow) => U, combOp: (U, U) => U, depth: Int, finalAggregateOnExecutor: Boolean)(implicit arg0: ClassTag[U]): U
- def treeAggregate[U](zeroValue: U)(seqOp: (U, InternalRow) => U, combOp: (U, U) => U, depth: Int)(implicit arg0: ClassTag[U]): U
- def treeReduce(f: (InternalRow, InternalRow) => InternalRow, depth: Int): InternalRow
- def union(other: RDD[InternalRow]): RDD[InternalRow]
- def unpersist(blocking: Boolean): FileScanRDD.this.type
- def withResources(rp: ResourceProfile): FileScanRDD.this.type
- def zip[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(InternalRow, U)]
- def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D])(f: (Iterator[InternalRow], Iterator[B], Iterator[C], Iterator[D]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]
- def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D], preservesPartitioning: Boolean)(f: (Iterator[InternalRow], Iterator[B], Iterator[C], Iterator[D]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]
- def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C])(f: (Iterator[InternalRow], Iterator[B], Iterator[C]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]
- def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C], preservesPartitioning: Boolean)(f: (Iterator[InternalRow], Iterator[B], Iterator[C]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]
- def zipPartitions[B, V](rdd2: RDD[B])(f: (Iterator[InternalRow], Iterator[B]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]
- def zipPartitions[B, V](rdd2: RDD[B], preservesPartitioning: Boolean)(f: (Iterator[InternalRow], Iterator[B]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]
- def zipPartitionsWithEvaluator[U](rdd2: RDD[InternalRow], evaluatorFactory: PartitionEvaluatorFactory[InternalRow, U])(implicit arg0: ClassTag[U]): RDD[U]
- def zipWithIndex(): RDD[(InternalRow, Long)]
- def zipWithUniqueId(): RDD[(InternalRow, Long)]
Inherited from Logging
Value Members
- def initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
- def initializeLogIfNecessary(isInterpreter: Boolean): Unit
- def isTraceEnabled(): Boolean
- def log: Logger
- def logDebug(msg: => String, throwable: Throwable): Unit
- def logDebug(msg: => String): Unit
- def logError(msg: => String, throwable: Throwable): Unit
- def logError(msg: => String): Unit
- def logInfo(msg: => String, throwable: Throwable): Unit
- def logInfo(msg: => String): Unit
- def logName: String
- def logTrace(msg: => String, throwable: Throwable): Unit
- def logTrace(msg: => String): Unit
- def logWarning(msg: => String, throwable: Throwable): Unit
- def logWarning(msg: => String): Unit
Inherited from AnyRef
Value Members
- final def !=(arg0: Any): Boolean
- final def ##: Int
- final def ==(arg0: Any): Boolean
- def clone(): AnyRef
- final def eq(arg0: AnyRef): Boolean
- def equals(arg0: AnyRef): Boolean
- def finalize(): Unit
- final def getClass(): Class[_ <: AnyRef]
- def hashCode(): Int
- final def ne(arg0: AnyRef): Boolean
- final def notify(): Unit
- final def notifyAll(): Unit
- final def synchronized[T0](arg0: => T0): T0
- final def wait(): Unit
- final def wait(arg0: Long, arg1: Int): Unit
- final def wait(arg0: Long): Unit
Inherited from Any
Value Members
- final def asInstanceOf[T0]: T0
- final def isInstanceOf[T0]: Boolean
Ungrouped
- final def !=(arg0: Any): Boolean
- final def ##: Int
- def ++(other: RDD[InternalRow]): RDD[InternalRow]
- final def ==(arg0: Any): Boolean
- def aggregate[U](zeroValue: U)(seqOp: (U, InternalRow) => U, combOp: (U, U) => U)(implicit arg0: ClassTag[U]): U
- final def asInstanceOf[T0]: T0
- def barrier(): RDDBarrier[InternalRow]
- def cache(): FileScanRDD.this.type
- def cartesian[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(InternalRow, U)]
- def checkpoint(): Unit
- def cleanShuffleDependencies(blocking: Boolean): Unit
- def clearDependencies(): Unit
- def clone(): AnyRef
- def coalesce(numPartitions: Int, shuffle: Boolean, partitionCoalescer: Option[PartitionCoalescer])(implicit ord: Ordering[InternalRow]): RDD[InternalRow]
- def collect[U](f: PartialFunction[InternalRow, U])(implicit arg0: ClassTag[U]): RDD[U]
- def collect(): Array[InternalRow]
- def compute(split: Partition, context: TaskContext): Iterator[InternalRow]
- def context: SparkContext
- def count(): Long
- def countApprox(timeout: Long, confidence: Double): PartialResult[BoundedDouble]
- def countApproxDistinct(relativeSD: Double): Long
- def countApproxDistinct(p: Int, sp: Int): Long
- def countByValue()(implicit ord: Ordering[InternalRow]): Map[InternalRow, Long]
- def countByValueApprox(timeout: Long, confidence: Double)(implicit ord: Ordering[InternalRow]): PartialResult[Map[InternalRow, BoundedDouble]]
- final def dependencies: Seq[Dependency[_]]
- def distinct(): RDD[InternalRow]
- def distinct(numPartitions: Int)(implicit ord: Ordering[InternalRow]): RDD[InternalRow]
- final def eq(arg0: AnyRef): Boolean
- def equals(arg0: AnyRef): Boolean
- val filePartitions: Seq[FilePartition]
- def filter(f: (InternalRow) => Boolean): RDD[InternalRow]
- def finalize(): Unit
- def first(): InternalRow
- def firstParent[U](implicit arg0: ClassTag[U]): RDD[U]
- def flatMap[U](f: (InternalRow) => TraversableOnce[U])(implicit arg0: ClassTag[U]): RDD[U]
- def fold(zeroValue: InternalRow)(op: (InternalRow, InternalRow) => InternalRow): InternalRow
- def foreach(f: (InternalRow) => Unit): Unit
- def foreachPartition(f: (Iterator[InternalRow]) => Unit): Unit
- def getCheckpointFile: Option[String]
- final def getClass(): Class[_ <: AnyRef]
- def getDependencies: Seq[Dependency[_]]
- final def getNumPartitions: Int
- def getOutputDeterministicLevel: rdd.DeterministicLevel.Value
- def getPartitions: Array[Partition]
- def getPreferredLocations(split: Partition): Seq[String]
- def getResourceProfile(): ResourceProfile
- def getStorageLevel: StorageLevel
- def glom(): RDD[Array[InternalRow]]
- def groupBy[K](f: (InternalRow) => K, p: Partitioner)(implicit kt: ClassTag[K], ord: Ordering[K]): RDD[(K, Iterable[InternalRow])]
- def groupBy[K](f: (InternalRow) => K, numPartitions: Int)(implicit kt: ClassTag[K]): RDD[(K, Iterable[InternalRow])]
- def groupBy[K](f: (InternalRow) => K)(implicit kt: ClassTag[K]): RDD[(K, Iterable[InternalRow])]
- def hashCode(): Int
- val id: Int
- def initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
- def initializeLogIfNecessary(isInterpreter: Boolean): Unit
- def intersection(other: RDD[InternalRow], numPartitions: Int): RDD[InternalRow]
- def intersection(other: RDD[InternalRow], partitioner: Partitioner)(implicit ord: Ordering[InternalRow]): RDD[InternalRow]
- def intersection(other: RDD[InternalRow]): RDD[InternalRow]
- lazy val isBarrier_: Boolean
- def isCheckpointed: Boolean
- def isEmpty(): Boolean
- final def isInstanceOf[T0]: Boolean
- def isTraceEnabled(): Boolean
- final def iterator(split: Partition, context: TaskContext): Iterator[InternalRow]
- def keyBy[K](f: (InternalRow) => K): RDD[(K, InternalRow)]
- def localCheckpoint(): FileScanRDD.this.type
- def log: Logger
- def logDebug(msg: => String, throwable: Throwable): Unit
- def logDebug(msg: => String): Unit
- def logError(msg: => String, throwable: Throwable): Unit
- def logError(msg: => String): Unit
- def logInfo(msg: => String, throwable: Throwable): Unit
- def logInfo(msg: => String): Unit
- def logName: String
- def logTrace(msg: => String, throwable: Throwable): Unit
- def logTrace(msg: => String): Unit
- def logWarning(msg: => String, throwable: Throwable): Unit
- def logWarning(msg: => String): Unit
- def map[U](f: (InternalRow) => U)(implicit arg0: ClassTag[U]): RDD[U]
- def mapPartitions[U](f: (Iterator[InternalRow]) => Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]
- def mapPartitionsWithEvaluator[U](evaluatorFactory: PartitionEvaluatorFactory[InternalRow, U])(implicit arg0: ClassTag[U]): RDD[U]
- def mapPartitionsWithIndex[U](f: (Int, Iterator[InternalRow]) => Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]
- def max()(implicit ord: Ordering[InternalRow]): InternalRow
- val metadataColumns: Seq[AttributeReference]
- def min()(implicit ord: Ordering[InternalRow]): InternalRow
- var name: String
- final def ne(arg0: AnyRef): Boolean
- final def notify(): Unit
- final def notifyAll(): Unit
- def parent[U](j: Int)(implicit arg0: ClassTag[U]): RDD[U]
- val partitioner: Option[Partitioner]
- final def partitions: Array[Partition]
- def persist(): FileScanRDD.this.type
- def persist(newLevel: StorageLevel): FileScanRDD.this.type
- def pipe(command: Seq[String], env: Map[String, String], printPipeContext: ((String) => Unit) => Unit, printRDDElement: (InternalRow, (String) => Unit) => Unit, separateWorkingDir: Boolean, bufferSize: Int, encoding: String): RDD[String]
- def pipe(command: String, env: Map[String, String]): RDD[String]
- def pipe(command: String): RDD[String]
- final def preferredLocations(split: Partition): Seq[String]
- def randomSplit(weights: Array[Double], seed: Long): Array[RDD[InternalRow]]
- val readSchema: StructType
- def reduce(f: (InternalRow, InternalRow) => InternalRow): InternalRow
- def repartition(numPartitions: Int)(implicit ord: Ordering[InternalRow]): RDD[InternalRow]
- def sample(withReplacement: Boolean, fraction: Double, seed: Long): RDD[InternalRow]
- def saveAsObjectFile(path: String): Unit
- def saveAsTextFile(path: String, codec: Class[_ <: CompressionCodec]): Unit
- def saveAsTextFile(path: String): Unit
- def setName(_name: String): FileScanRDD.this.type
- def sortBy[K](f: (InternalRow) => K, ascending: Boolean, numPartitions: Int)(implicit ord: Ordering[K], ctag: ClassTag[K]): RDD[InternalRow]
- def sparkContext: SparkContext
- def subtract(other: RDD[InternalRow], p: Partitioner)(implicit ord: Ordering[InternalRow]): RDD[InternalRow]
- def subtract(other: RDD[InternalRow], numPartitions: Int): RDD[InternalRow]
- def subtract(other: RDD[InternalRow]): RDD[InternalRow]
- final def synchronized[T0](arg0: => T0): T0
- def take(num: Int): Array[InternalRow]
- def takeOrdered(num: Int)(implicit ord: Ordering[InternalRow]): Array[InternalRow]
- def takeSample(withReplacement: Boolean, num: Int, seed: Long): Array[InternalRow]
- def toDebugString: String
- def toJavaRDD(): JavaRDD[InternalRow]
- def toLocalIterator: Iterator[InternalRow]
- def toString(): String
- def top(num: Int)(implicit ord: Ordering[InternalRow]): Array[InternalRow]
- def treeAggregate[U](zeroValue: U, seqOp: (U, InternalRow) => U, combOp: (U, U) => U, depth: Int, finalAggregateOnExecutor: Boolean)(implicit arg0: ClassTag[U]): U
- def treeAggregate[U](zeroValue: U)(seqOp: (U, InternalRow) => U, combOp: (U, U) => U, depth: Int)(implicit arg0: ClassTag[U]): U
- def treeReduce(f: (InternalRow, InternalRow) => InternalRow, depth: Int): InternalRow
- def union(other: RDD[InternalRow]): RDD[InternalRow]
- def unpersist(blocking: Boolean): FileScanRDD.this.type
- final def wait(): Unit
- final def wait(arg0: Long, arg1: Int): Unit
- final def wait(arg0: Long): Unit
- def withResources(rp: ResourceProfile): FileScanRDD.this.type
- def zip[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(InternalRow, U)]
- def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D])(f: (Iterator[InternalRow], Iterator[B], Iterator[C], Iterator[D]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]
- def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D], preservesPartitioning: Boolean)(f: (Iterator[InternalRow], Iterator[B], Iterator[C], Iterator[D]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]
- def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C])(f: (Iterator[InternalRow], Iterator[B], Iterator[C]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]
- def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C], preservesPartitioning: Boolean)(f: (Iterator[InternalRow], Iterator[B], Iterator[C]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]
- def zipPartitions[B, V](rdd2: RDD[B])(f: (Iterator[InternalRow], Iterator[B]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]
- def zipPartitions[B, V](rdd2: RDD[B], preservesPartitioning: Boolean)(f: (Iterator[InternalRow], Iterator[B]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]
- def zipPartitionsWithEvaluator[U](rdd2: RDD[InternalRow], evaluatorFactory: PartitionEvaluatorFactory[InternalRow, U])(implicit arg0: ClassTag[U]): RDD[U]
- def zipWithIndex(): RDD[(InternalRow, Long)]
- def zipWithUniqueId(): RDD[(InternalRow, Long)]