com.datastax.spark.connector.rdd

CassandraTableScanRDD

class CassandraTableScanRDD[R] extends CassandraRDD[R] with CassandraTableRowReaderProvider[R] with SplitSizeEstimator[R]

RDD representing a Table Scan of A Cassandra table.

This class is the main entry point for analyzing data in Cassandra database with Spark. Obtain objects of this class by calling com.datastax.spark.connector.SparkContextFunctions.cassandraTable.

Configuration properties should be passed in the SparkConf configuration of SparkContext. CassandraRDD needs to open connection to Cassandra, therefore it requires appropriate connection property values to be present in SparkConf. For the list of required and available properties, see CassandraConnector.

CassandraRDD divides the data set into smaller partitions, processed locally on every cluster node. A data partition consists of one or more contiguous token ranges. To reduce the number of roundtrips to Cassandra, every partition is fetched in batches.

The following properties control the number of partitions and the fetch size: - spark.cassandra.input.split.size_in_mb: approx amount of data to be fetched into a single Spark partition, default 64 MB - spark.cassandra.input.fetch.size_in_rows: number of CQL rows fetched per roundtrip, default 1000

A CassandraRDD object gets serialized and sent to every Spark Executor, which then calls the compute method to fetch the data on every node. The getPreferredLocations method tells Spark the preferred nodes to fetch a partition from, so that the data for the partition are at the same node the task was sent to. If Cassandra nodes are collocated with Spark nodes, the queries are always sent to the Cassandra process running on the same node as the Spark Executor process, hence data are not transferred between nodes. If a Cassandra node fails or gets overloaded during read, the queries are retried to a different node.

By default, reads are performed at ConsistencyLevel.LOCAL_ONE in order to leverage data-locality and minimize network traffic. This read consistency level is controlled by the spark.cassandra.input.consistency.level property.

Linear Supertypes
SplitSizeEstimator[R], CassandraTableRowReaderProvider[R], CassandraRDD[R], RDD[R], Logging, Serializable, Serializable, AnyRef, Any
Known Subclasses
Ordering
  1. Alphabetic
  2. By inheritance
Inherited
  1. CassandraTableScanRDD
  2. SplitSizeEstimator
  3. CassandraTableRowReaderProvider
  4. CassandraRDD
  5. RDD
  6. Logging
  7. Serializable
  8. Serializable
  9. AnyRef
  10. Any
Implicitly
  1. by rddToPairRDDFunctions
  2. by numericRDDToDoubleRDDFunctions
  3. by doubleRDDToDoubleRDDFunctions
  4. by rddToOrderedRDDFunctions
  5. by rddToSequenceFileRDDFunctions
  6. by rddToAsyncRDDActions
  7. by toCassandraTableScanRDDPairFunctions
  8. by toPairRDDFunctions
  9. by toCassandraTableScanFunctions
  10. by toRDDFunctions
  11. by any2stringadd
  12. by any2stringfmt
  13. by any2ArrowAssoc
  14. by any2Ensuring
  1. Hide All
  2. Show all
Learn more about member selection
Visibility
  1. Public
  2. All

Type Members

  1. type Self = CassandraTableScanRDD[R]

    This is slightly different than Scala this.

    This is slightly different than Scala this.type. this.type is the unique singleton type of an object which is not compatible with other instances of the same type, so returning anything other than this is not really possible without lying to the compiler by explicit casts. Here SelfType is used to return a copy of the object - a different instance of the same type

    Definition Classes
    CassandraTableScanRDDCassandraRDD

Value Members

  1. final def !=(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  2. final def !=(arg0: Any): Boolean

    Definition Classes
    Any
  3. final def ##(): Int

    Definition Classes
    AnyRef → Any
  4. def +(other: String): String

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to StringAdd performed by method any2stringadd in scala.Predef.
    Definition Classes
    StringAdd
  5. def ++(other: RDD[R]): RDD[R]

    Definition Classes
    RDD
  6. def ->[B](y: B): (CassandraTableScanRDD[R], B)

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to ArrowAssoc[CassandraTableScanRDD[R]] performed by method any2ArrowAssoc in scala.Predef.
    Definition Classes
    ArrowAssoc
    Annotations
    @inline()
  7. final def ==(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  8. final def ==(arg0: Any): Boolean

    Definition Classes
    Any
  9. def aggregate[U](zeroValue: U)(seqOp: (U, R) ⇒ U, combOp: (U, U) ⇒ U)(implicit arg0: ClassTag[U]): U

    Definition Classes
    RDD
  10. def aggregateByKey[U](zeroValue: U)(seqOp: (U, V) ⇒ U, combOp: (U, U) ⇒ U)(implicit arg0: ClassTag[U]): RDD[(K, U)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  11. def aggregateByKey[U](zeroValue: U, numPartitions: Int)(seqOp: (U, V) ⇒ U, combOp: (U, U) ⇒ U)(implicit arg0: ClassTag[U]): RDD[(K, U)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  12. def aggregateByKey[U](zeroValue: U, partitioner: Partitioner)(seqOp: (U, V) ⇒ U, combOp: (U, U) ⇒ U)(implicit arg0: ClassTag[U]): RDD[(K, U)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  13. def applyPartitioner[TokenValue, T <: Token[TokenValue]](partitioner: CassandraPartitioner[K, TokenValue, T]): CassandraTableScanRDD[(K, V)]

    Use a specific CassandraPartitioner to use with this PairRDD.

    Use a specific CassandraPartitioner to use with this PairRDD.

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to CassandraTableScanPairRDDFunctions[K, V] performed by method toCassandraTableScanRDDPairFunctions in com.datastax.spark.connector. This conversion will take place only if R is (K, V) (R =:= (K, V)).
    Definition Classes
    CassandraTableScanPairRDDFunctions
  14. def applyPartitionerFrom[X](thatRdd: CassandraTableScanRDD[(K, X)]): CassandraTableScanRDD[(K, V)]

    Use the CassandraPartitioner from another CassandraTableScanRDD which shares the same key type.

    Use the CassandraPartitioner from another CassandraTableScanRDD which shares the same key type. All Partition Keys columns must also be present in the keys of the target RDD.

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to CassandraTableScanPairRDDFunctions[K, V] performed by method toCassandraTableScanRDDPairFunctions in com.datastax.spark.connector. This conversion will take place only if R is (K, V) (R =:= (K, V)).
    Definition Classes
    CassandraTableScanPairRDDFunctions
  15. def as[B, A0, A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11](f: (A0, A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6], arg8: TypeConverter[A7], arg9: TypeConverter[A8], arg10: TypeConverter[A9], arg11: TypeConverter[A10], arg12: TypeConverter[A11]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  16. def as[B, A0, A1, A2, A3, A4, A5, A6, A7, A8, A9, A10](f: (A0, A1, A2, A3, A4, A5, A6, A7, A8, A9, A10) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6], arg8: TypeConverter[A7], arg9: TypeConverter[A8], arg10: TypeConverter[A9], arg11: TypeConverter[A10]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  17. def as[B, A0, A1, A2, A3, A4, A5, A6, A7, A8, A9](f: (A0, A1, A2, A3, A4, A5, A6, A7, A8, A9) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6], arg8: TypeConverter[A7], arg9: TypeConverter[A8], arg10: TypeConverter[A9]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  18. def as[B, A0, A1, A2, A3, A4, A5, A6, A7, A8](f: (A0, A1, A2, A3, A4, A5, A6, A7, A8) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6], arg8: TypeConverter[A7], arg9: TypeConverter[A8]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  19. def as[B, A0, A1, A2, A3, A4, A5, A6, A7](f: (A0, A1, A2, A3, A4, A5, A6, A7) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6], arg8: TypeConverter[A7]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  20. def as[B, A0, A1, A2, A3, A4, A5, A6](f: (A0, A1, A2, A3, A4, A5, A6) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  21. def as[B, A0, A1, A2, A3, A4, A5](f: (A0, A1, A2, A3, A4, A5) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  22. def as[B, A0, A1, A2, A3, A4](f: (A0, A1, A2, A3, A4) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  23. def as[B, A0, A1, A2, A3](f: (A0, A1, A2, A3) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  24. def as[B, A0, A1, A2](f: (A0, A1, A2) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  25. def as[B, A0, A1](f: (A0, A1) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  26. def as[B, A0](f: (A0) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0]): CassandraRDD[B]

    Maps each row into object of a different type using provided function taking column value(s) as argument(s).

    Maps each row into object of a different type using provided function taking column value(s) as argument(s). Can be used to convert each row to a tuple or a case class object:

    sc.cassandraTable("ks", "table")
      .select("column1")
      .as((s: String) => s)                 // yields CassandraRDD[String]
    
    sc.cassandraTable("ks", "table")
      .select("column1", "column2")
      .as((_: String, _: Long))             // yields CassandraRDD[(String, Long)]
    
    case class MyRow(key: String, value: Long)
    sc.cassandraTable("ks", "table")
      .select("column1", "column2")
      .as(MyRow)                            // yields CassandraRDD[MyRow]
    Definition Classes
    CassandraRDD
  27. final def asInstanceOf[T0]: T0

    Definition Classes
    Any
  28. def cache(): CassandraTableScanRDD.this.type

    Definition Classes
    RDD
  29. def cartesian[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(R, U)]

    Definition Classes
    RDD
  30. def cassandraCount(): Long

    Counts the number of items in this RDD by selecting count(*) on Cassandra table

    Counts the number of items in this RDD by selecting count(*) on Cassandra table

    Definition Classes
    CassandraTableScanRDDCassandraRDD
  31. lazy val cassandraPartitionerClassName: String

    Attributes
    protected
    Definition Classes
    CassandraTableRowReaderProvider
  32. def checkColumnsExistence(columns: Seq[ColumnRef]): Seq[ColumnRef]

    Attributes
    protected
    Definition Classes
    CassandraTableRowReaderProvider
  33. def checkpoint(): Unit

    Definition Classes
    RDD
  34. implicit val classTag: ClassTag[R]

  35. def clearDependencies(): Unit

    Attributes
    protected
    Definition Classes
    RDD
  36. def clone(): AnyRef

    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  37. def clusteringOrder(order: ClusteringOrder): Self

    Adds a CQL ORDER BY clause to the query.

    Adds a CQL ORDER BY clause to the query. It can be applied only in case there are clustering columns and primary key predicate is pushed down in where. It is useful when the default direction of ordering rows within a single Cassandra partition needs to be changed.

    Definition Classes
    CassandraRDD
  38. val clusteringOrder: Option[ClusteringOrder]

    Definition Classes
    CassandraTableScanRDDCassandraRDD
  39. def coalesce(numPartitions: Int, shuffle: Boolean = false, partitionCoalescer: Option[PartitionCoalescer])(implicit ord: Ordering[R] = null): RDD[R]

    This method overrides the default spark behavior and will not create a CoalesceRDD.

    This method overrides the default spark behavior and will not create a CoalesceRDD. Instead it will reduce the number of partitions by adjusting the partitioning of C* data on read. Using this method will override spark.cassandra.input.split.size. The method is useful with where() method call, when actual size of data is smaller then the table size. It has no effect if a partition key is used in where clause.

    numPartitions

    number of partitions

    shuffle

    whether to call shuffle after

    partitionCoalescer

    is ignored if no shuffle, or just passed to shuffled CoalesceRDD

    ord
    returns

    new CassandraTableScanRDD with predefined number of partitions

    Definition Classes
    CassandraTableScanRDD → RDD
  40. def cogroup[W1, W2, W3](other1: RDD[(K, W1)], other2: RDD[(K, W2)], other3: RDD[(K, W3)], numPartitions: Int): RDD[(K, (Iterable[V], Iterable[W1], Iterable[W2], Iterable[W3]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  41. def cogroup[W1, W2](other1: RDD[(K, W1)], other2: RDD[(K, W2)], numPartitions: Int): RDD[(K, (Iterable[V], Iterable[W1], Iterable[W2]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  42. def cogroup[W](other: RDD[(K, W)], numPartitions: Int): RDD[(K, (Iterable[V], Iterable[W]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  43. def cogroup[W1, W2](other1: RDD[(K, W1)], other2: RDD[(K, W2)]): RDD[(K, (Iterable[V], Iterable[W1], Iterable[W2]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  44. def cogroup[W](other: RDD[(K, W)]): RDD[(K, (Iterable[V], Iterable[W]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  45. def cogroup[W1, W2, W3](other1: RDD[(K, W1)], other2: RDD[(K, W2)], other3: RDD[(K, W3)]): RDD[(K, (Iterable[V], Iterable[W1], Iterable[W2], Iterable[W3]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  46. def cogroup[W1, W2](other1: RDD[(K, W1)], other2: RDD[(K, W2)], partitioner: Partitioner): RDD[(K, (Iterable[V], Iterable[W1], Iterable[W2]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  47. def cogroup[W](other: RDD[(K, W)], partitioner: Partitioner): RDD[(K, (Iterable[V], Iterable[W]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  48. def cogroup[W1, W2, W3](other1: RDD[(K, W1)], other2: RDD[(K, W2)], other3: RDD[(K, W3)], partitioner: Partitioner): RDD[(K, (Iterable[V], Iterable[W1], Iterable[W2], Iterable[W3]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  49. def collect[U](f: PartialFunction[R, U])(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
  50. def collect(): Array[R]

    Definition Classes
    RDD
  51. def collectAsMap(): Map[K, V]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  52. def collectAsync(): FutureAction[Seq[R]]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to AsyncRDDActions[R] performed by method rddToAsyncRDDActions in org.apache.spark.rdd.RDD. This conversion will take place only if R is accompanied by a ClassTag, which is a runtime representation of its type that survives erasure (R: ClassTag).
    Definition Classes
    AsyncRDDActions
  53. val columnNames: ColumnSelector

  54. def combineByKey[C](createCombiner: (V) ⇒ C, mergeValue: (C, V) ⇒ C, mergeCombiners: (C, C) ⇒ C): RDD[(K, C)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  55. def combineByKey[C](createCombiner: (V) ⇒ C, mergeValue: (C, V) ⇒ C, mergeCombiners: (C, C) ⇒ C, numPartitions: Int): RDD[(K, C)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  56. def combineByKey[C](createCombiner: (V) ⇒ C, mergeValue: (C, V) ⇒ C, mergeCombiners: (C, C) ⇒ C, partitioner: Partitioner, mapSideCombine: Boolean, serializer: Serializer): RDD[(K, C)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  57. def combineByKeyWithClassTag[C](createCombiner: (V) ⇒ C, mergeValue: (C, V) ⇒ C, mergeCombiners: (C, C) ⇒ C)(implicit ct: ClassTag[C]): RDD[(K, C)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
    Annotations
    @Experimental()
  58. def combineByKeyWithClassTag[C](createCombiner: (V) ⇒ C, mergeValue: (C, V) ⇒ C, mergeCombiners: (C, C) ⇒ C, numPartitions: Int)(implicit ct: ClassTag[C]): RDD[(K, C)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
    Annotations
    @Experimental()
  59. def combineByKeyWithClassTag[C](createCombiner: (V) ⇒ C, mergeValue: (C, V) ⇒ C, mergeCombiners: (C, C) ⇒ C, partitioner: Partitioner, mapSideCombine: Boolean, serializer: Serializer)(implicit ct: ClassTag[C]): RDD[(K, C)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
    Annotations
    @Experimental()
  60. def compute(split: Partition, context: TaskContext): Iterator[R]

    Definition Classes
    CassandraTableScanRDD → RDD
  61. val connector: CassandraConnector

  62. def consistencyLevel: ConsistencyLevel

    Attributes
    protected
    Definition Classes
    CassandraTableRowReaderProvider
  63. def context: SparkContext

    Definition Classes
    RDD
  64. def convertTo[B](implicit arg0: ClassTag[B], arg1: RowReaderFactory[B]): CassandraTableScanRDD[B]

    Attributes
    protected
    Definition Classes
    CassandraTableScanRDDCassandraRDD
  65. def copy(columnNames: ColumnSelector = columnNames, where: CqlWhereClause = where, limit: Option[CassandraLimit] = limit, clusteringOrder: Option[ClusteringOrder] = None, readConf: ReadConf = readConf, connector: CassandraConnector = connector): Self

    Allows to copy this RDD with changing some of the properties

    Allows to copy this RDD with changing some of the properties

    Attributes
    protected
    Definition Classes
    CassandraTableScanRDDCassandraRDD
  66. def count(): Long

    Definition Classes
    RDD
  67. def countApprox(timeout: Long, confidence: Double): PartialResult[BoundedDouble]

    Definition Classes
    RDD
  68. def countApproxDistinct(relativeSD: Double): Long

    Definition Classes
    RDD
  69. def countApproxDistinct(p: Int, sp: Int): Long

    Definition Classes
    RDD
  70. def countApproxDistinctByKey(relativeSD: Double): RDD[(K, Long)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  71. def countApproxDistinctByKey(relativeSD: Double, numPartitions: Int): RDD[(K, Long)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  72. def countApproxDistinctByKey(relativeSD: Double, partitioner: Partitioner): RDD[(K, Long)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  73. def countApproxDistinctByKey(p: Int, sp: Int, partitioner: Partitioner): RDD[(K, Long)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  74. def countAsync(): FutureAction[Long]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to AsyncRDDActions[R] performed by method rddToAsyncRDDActions in org.apache.spark.rdd.RDD. This conversion will take place only if R is accompanied by a ClassTag, which is a runtime representation of its type that survives erasure (R: ClassTag).
    Definition Classes
    AsyncRDDActions
  75. def countByKey(): Map[K, Long]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  76. def countByKeyApprox(timeout: Long, confidence: Double): PartialResult[Map[K, BoundedDouble]]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  77. def countByValue()(implicit ord: Ordering[R]): Map[R, Long]

    Definition Classes
    RDD
  78. def countByValueApprox(timeout: Long, confidence: Double)(implicit ord: Ordering[R]): PartialResult[Map[R, BoundedDouble]]

    Definition Classes
    RDD
  79. def deleteFromCassandra(keyspaceName: String, tableName: String, deleteColumns: ColumnSelector = SomeColumns(), keyColumns: ColumnSelector = PrimaryKeyColumns, writeConf: WriteConf = ...)(implicit connector: CassandraConnector = CassandraConnector(sparkContext), rwf: RowWriterFactory[R]): Unit

    Delete data from Cassandra table, using data from the RDD as primary keys.

    Delete data from Cassandra table, using data from the RDD as primary keys. Uses the specified column names.

    keyspaceName

    the name of the Keyspace to use

    tableName

    the name of the Table to use

    deleteColumns

    The list of column names to delete, empty ColumnSelector means full row.

    keyColumns

    Primary key columns selector, Optional. All RDD primary columns columns will be checked by default

    writeConf

    additional configuration object allowing to set consistency level, batch size, etc.

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to RDDFunctions[R] performed by method toRDDFunctions in com.datastax.spark.connector.
    Definition Classes
    RDDFunctionsWritableToCassandra
    See also

    com.datastax.spark.connector.writer.WritableToCassandra

  80. final def dependencies: Seq[Dependency[_]]

    Definition Classes
    RDD
  81. def distinct(): RDD[R]

    Definition Classes
    RDD
  82. def distinct(numPartitions: Int)(implicit ord: Ordering[R]): RDD[R]

    Definition Classes
    RDD
  83. def ensuring(cond: (CassandraTableScanRDD[R]) ⇒ Boolean, msg: ⇒ Any): CassandraTableScanRDD[R]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to Ensuring[CassandraTableScanRDD[R]] performed by method any2Ensuring in scala.Predef. This conversion will take place only if R is a superclass of Any and a subclass of (Nothing, Nothing) with Double (R >: Any <: (Nothing, Nothing) with Double).
    Definition Classes
    Ensuring
  84. def ensuring(cond: (CassandraTableScanRDD[R]) ⇒ Boolean): CassandraTableScanRDD[R]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to Ensuring[CassandraTableScanRDD[R]] performed by method any2Ensuring in scala.Predef. This conversion will take place only if R is a superclass of Any and a subclass of (Nothing, Nothing) with Double (R >: Any <: (Nothing, Nothing) with Double).
    Definition Classes
    Ensuring
  85. def ensuring(cond: Boolean, msg: ⇒ Any): CassandraTableScanRDD[R]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to Ensuring[CassandraTableScanRDD[R]] performed by method any2Ensuring in scala.Predef. This conversion will take place only if R is a superclass of Any and a subclass of (Nothing, Nothing) with Double (R >: Any <: (Nothing, Nothing) with Double).
    Definition Classes
    Ensuring
  86. def ensuring(cond: Boolean): CassandraTableScanRDD[R]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to Ensuring[CassandraTableScanRDD[R]] performed by method any2Ensuring in scala.Predef. This conversion will take place only if R is a superclass of Any and a subclass of (Nothing, Nothing) with Double (R >: Any <: (Nothing, Nothing) with Double).
    Definition Classes
    Ensuring
  87. final def eq(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  88. def equals(arg0: Any): Boolean

    Definition Classes
    AnyRef → Any
  89. def estimateSplitCount(splitSize: Long): Int

    Definition Classes
    SplitSizeEstimator
  90. def fetchSize: Int

    Attributes
    protected
    Definition Classes
    CassandraTableRowReaderProvider
  91. def filter(f: (R) ⇒ Boolean): RDD[R]

    Definition Classes
    RDD
  92. def filterByRange(lower: K, upper: K): RDD[(K, V)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to OrderedRDDFunctions[K, V, (K, V)] performed by method rddToOrderedRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type Ordering[K] is in scope
    2. an implicit value of type ClassTag[K] is in scope
    3. an implicit value of type ClassTag[V] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    OrderedRDDFunctions
  93. def finalize(): Unit

    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  94. def first(): R

    Definition Classes
    RDD
  95. def firstParent[U](implicit arg0: ClassTag[U]): RDD[U]

    Attributes
    protected[org.apache.spark]
    Definition Classes
    RDD
  96. def flatMap[U](f: (R) ⇒ TraversableOnce[U])(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
  97. def flatMapValues[U](f: (V) ⇒ TraversableOnce[U]): RDD[(K, U)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  98. def fold(zeroValue: R)(op: (R, R) ⇒ R): R

    Definition Classes
    RDD
  99. def foldByKey(zeroValue: V)(func: (V, V) ⇒ V): RDD[(K, V)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  100. def foldByKey(zeroValue: V, numPartitions: Int)(func: (V, V) ⇒ V): RDD[(K, V)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  101. def foldByKey(zeroValue: V, partitioner: Partitioner)(func: (V, V) ⇒ V): RDD[(K, V)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  102. def foreach(f: (R) ⇒ Unit): Unit

    Definition Classes
    RDD
  103. def foreachAsync(f: (R) ⇒ Unit): FutureAction[Unit]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to AsyncRDDActions[R] performed by method rddToAsyncRDDActions in org.apache.spark.rdd.RDD. This conversion will take place only if R is accompanied by a ClassTag, which is a runtime representation of its type that survives erasure (R: ClassTag).
    Definition Classes
    AsyncRDDActions
  104. def foreachPartition(f: (Iterator[R]) ⇒ Unit): Unit

    Definition Classes
    RDD
  105. def foreachPartitionAsync(f: (Iterator[R]) ⇒ Unit): FutureAction[Unit]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to AsyncRDDActions[R] performed by method rddToAsyncRDDActions in org.apache.spark.rdd.RDD. This conversion will take place only if R is accompanied by a ClassTag, which is a runtime representation of its type that survives erasure (R: ClassTag).
    Definition Classes
    AsyncRDDActions
  106. def formatted(fmtstr: String): String

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to StringFormat performed by method any2stringfmt in scala.Predef.
    Definition Classes
    StringFormat
    Annotations
    @inline()
  107. def fullOuterJoin[W](other: RDD[(K, W)], numPartitions: Int): RDD[(K, (Option[V], Option[W]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  108. def fullOuterJoin[W](other: RDD[(K, W)]): RDD[(K, (Option[V], Option[W]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  109. def fullOuterJoin[W](other: RDD[(K, W)], partitioner: Partitioner): RDD[(K, (Option[V], Option[W]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  110. def getCheckpointFile: Option[String]

    Definition Classes
    RDD
  111. final def getClass(): Class[_]

    Definition Classes
    AnyRef → Any
  112. def getDependencies: Seq[Dependency[_]]

    Attributes
    protected
    Definition Classes
    RDD
  113. final def getNumPartitions: Int

    Definition Classes
    RDD
    Annotations
    @Since( "1.6.0" )
  114. def getPartitions: Array[Partition]

    Definition Classes
    CassandraTableScanRDD → RDD
  115. def getPreferredLocations(split: Partition): Seq[String]

    Definition Classes
    CassandraTableScanRDD → RDD
  116. def getStorageLevel: StorageLevel

    Definition Classes
    RDD
  117. def glom(): RDD[Array[R]]

    Definition Classes
    RDD
  118. def groupBy[K](f: (R) ⇒ K, p: Partitioner)(implicit kt: ClassTag[K], ord: Ordering[K]): RDD[(K, Iterable[R])]

    Definition Classes
    RDD
  119. def groupBy[K](f: (R) ⇒ K, numPartitions: Int)(implicit kt: ClassTag[K]): RDD[(K, Iterable[R])]

    Definition Classes
    RDD
  120. def groupBy[K](f: (R) ⇒ K)(implicit kt: ClassTag[K]): RDD[(K, Iterable[R])]

    Definition Classes
    RDD
  121. def groupByKey(): RDD[(K, Iterable[V])]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  122. def groupByKey(numPartitions: Int): RDD[(K, Iterable[V])]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  123. def groupByKey(partitioner: Partitioner): RDD[(K, Iterable[V])]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  124. def groupWith[W1, W2, W3](other1: RDD[(K, W1)], other2: RDD[(K, W2)], other3: RDD[(K, W3)]): RDD[(K, (Iterable[V], Iterable[W1], Iterable[W2], Iterable[W3]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  125. def groupWith[W1, W2](other1: RDD[(K, W1)], other2: RDD[(K, W2)]): RDD[(K, (Iterable[V], Iterable[W1], Iterable[W2]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  126. def groupWith[W](other: RDD[(K, W)]): RDD[(K, (Iterable[V], Iterable[W]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  127. def hashCode(): Int

    Definition Classes
    AnyRef → Any
  128. val id: Int

    Definition Classes
    RDD
  129. def initializeLogIfNecessary(isInterpreter: Boolean): Unit

    Attributes
    protected
    Definition Classes
    Logging
  130. def intersection(other: RDD[R], numPartitions: Int): RDD[R]

    Definition Classes
    RDD
  131. def intersection(other: RDD[R], partitioner: Partitioner)(implicit ord: Ordering[R]): RDD[R]

    Definition Classes
    RDD
  132. def intersection(other: RDD[R]): RDD[R]

    Definition Classes
    RDD
  133. def isCheckpointed: Boolean

    Definition Classes
    RDD
  134. def isEmpty(): Boolean

    Definition Classes
    RDD
  135. final def isInstanceOf[T0]: Boolean

    Definition Classes
    Any
  136. def isTraceEnabled(): Boolean

    Attributes
    protected
    Definition Classes
    Logging
  137. final def iterator(split: Partition, context: TaskContext): Iterator[R]

    Definition Classes
    RDD
  138. def join[W](other: RDD[(K, W)], numPartitions: Int): RDD[(K, (V, W))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  139. def join[W](other: RDD[(K, W)]): RDD[(K, (V, W))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  140. def join[W](other: RDD[(K, W)], partitioner: Partitioner): RDD[(K, (V, W))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  141. def joinWithCassandraTable[R](keyspaceName: String, tableName: String, selectedColumns: ColumnSelector = AllColumns, joinColumns: ColumnSelector = PartitionKeyColumns)(implicit connector: CassandraConnector = CassandraConnector(sparkContext), newType: ClassTag[R], rrf: RowReaderFactory[R], ev: ValidRDDType[R], currentType: ClassTag[R], rwf: RowWriterFactory[R]): CassandraJoinRDD[R, R]

    Uses the data from RDD to join with a Cassandra table without retrieving the entire table.

    Uses the data from RDD to join with a Cassandra table without retrieving the entire table. Any RDD which can be used to saveToCassandra can be used to joinWithCassandra as well as any RDD which only specifies the partition Key of a Cassandra Table. This method executes single partition requests against the Cassandra Table and accepts the functional modifiers that a normal com.datastax.spark.connector.rdd.CassandraTableScanRDD takes.

    By default this method only uses the Partition Key for joining but any combination of columns which are acceptable to C* can be used in the join. Specify columns using joinColumns as a parameter or the on() method.

    Example With Prior Repartitioning:

    val source = sc.parallelize(keys).map(x => new KVRow(x))
    val repart = source.repartitionByCassandraReplica(keyspace, tableName, 10)
    val someCass = repart.joinWithCassandraTable(keyspace, tableName)

    Example Joining on Clustering Columns:

    val source = sc.parallelize(keys).map(x => (x, x * 100))
    val someCass = source.joinWithCassandraTable(keyspace, wideTable).on(SomeColumns("key", "group"))
    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to RDDFunctions[R] performed by method toRDDFunctions in com.datastax.spark.connector.
    Definition Classes
    RDDFunctions
  142. def keyAndApplyPartitionerFrom[K, X](thatRDD: CassandraTableScanRDD[(K, X)], columnSelector: ColumnSelector = PartitionKeyColumns)(implicit classTag: ClassTag[K], rrf: RowReaderFactory[K], rwf: RowWriterFactory[K]): CassandraTableScanRDD[(K, R)]

    Shortcut for rdd.keyBy[K].applyPartitionerFrom(thatRDD[K, V]) where K is the key type of the target RDD.

    Shortcut for rdd.keyBy[K].applyPartitionerFrom(thatRDD[K, V]) where K is the key type of the target RDD. This guarentees that the partitioner applied to this rdd will match the key type.

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to CassandraTableScanRDDFunctions[R] performed by method toCassandraTableScanFunctions in com.datastax.spark.connector.
    Definition Classes
    CassandraTableScanRDDFunctions
  143. def keyBy[K]()(implicit classtag: ClassTag[K], rrf: RowReaderFactory[K], rwf: RowWriterFactory[K]): CassandraTableScanRDD[(K, R)]

    Extracts a key of the given class from all the available columns.

    Extracts a key of the given class from all the available columns.

    See also

    keyBy(ColumnSelector)

  144. def keyBy[K](columns: ColumnRef*)(implicit classtag: ClassTag[K], rrf: RowReaderFactory[K], rwf: RowWriterFactory[K]): CassandraTableScanRDD[(K, R)]

    Extracts a key of the given class from the given columns.

    Extracts a key of the given class from the given columns.

    See also

    keyBy(ColumnSelector)

  145. def keyBy[K](columns: ColumnSelector)(implicit classtag: ClassTag[K], rrf: RowReaderFactory[K], rwf: RowWriterFactory[K]): CassandraTableScanRDD[(K, R)]

    Selects a subset of columns mapped to the key and returns an RDD of pairs.

    Selects a subset of columns mapped to the key and returns an RDD of pairs. Similar to the builtin Spark keyBy method, but this one uses implicit RowReaderFactory to construct the key objects. The selected columns must be available in the CassandraRDD.

    If the selected columns contain the complete partition key a CassandraPartitioner will also be created.

    columns

    column selector passed to the rrf to create the row reader, useful when the key is mapped to a tuple or a single value

  146. def keyBy[K](f: (R) ⇒ K): RDD[(K, R)]

    Definition Classes
    RDD
  147. def keyByCassandraReplica(keyspaceName: String, tableName: String, partitionKeyMapper: ColumnSelector = PartitionKeyColumns)(implicit connector: CassandraConnector = CassandraConnector(sparkContext), currentType: ClassTag[R], rwf: RowWriterFactory[R]): RDD[(Set[InetAddress], R)]

    Key every row in the RDD by with the IP Adresses of all of the Cassandra nodes which a contain a replica of the data specified by that row.

    Key every row in the RDD by with the IP Adresses of all of the Cassandra nodes which a contain a replica of the data specified by that row. The calling RDD must have rows that can be converted into the partition key of the given Cassandra Table.

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to RDDFunctions[R] performed by method toRDDFunctions in com.datastax.spark.connector.
    Definition Classes
    RDDFunctions
  148. def keys: RDD[K]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  149. val keyspaceName: String

  150. def leftJoinWithCassandraTable[R](keyspaceName: String, tableName: String, selectedColumns: ColumnSelector = AllColumns, joinColumns: ColumnSelector = PartitionKeyColumns)(implicit connector: CassandraConnector = CassandraConnector(sparkContext), newType: ClassTag[R], rrf: RowReaderFactory[R], ev: ValidRDDType[R], currentType: ClassTag[R], rwf: RowWriterFactory[R]): CassandraLeftJoinRDD[R, R]

    Uses the data from RDD to left join with a Cassandra table without retrieving the entire table.

    Uses the data from RDD to left join with a Cassandra table without retrieving the entire table. Any RDD which can be used to saveToCassandra can be used to leftJoinWithCassandra as well as any RDD which only specifies the partition Key of a Cassandra Table. This method executes single partition requests against the Cassandra Table and accepts the functional modifiers that a normal com.datastax.spark.connector.rdd.CassandraTableScanRDD takes.

    By default this method only uses the Partition Key for joining but any combination of columns which are acceptable to C* can be used in the join. Specify columns using joinColumns as a parameter or the on() method.

    Example With Prior Repartitioning:

    val source = sc.parallelize(keys).map(x => new KVRow(x))
    val repart = source.repartitionByCassandraReplica(keyspace, tableName, 10)
    val someCass = repart.leftJoinWithCassandraTable(keyspace, tableName)

    Example Joining on Clustering Columns:

    val source = sc.parallelize(keys).map(x => (x, x * 100))
    val someCass = source.leftJoinWithCassandraTable(keyspace, wideTable).on(SomeColumns("key", "group"))
    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to RDDFunctions[R] performed by method toRDDFunctions in com.datastax.spark.connector.
    Definition Classes
    RDDFunctions
  151. def leftOuterJoin[W](other: RDD[(K, W)], numPartitions: Int): RDD[(K, (V, Option[W]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  152. def leftOuterJoin[W](other: RDD[(K, W)]): RDD[(K, (V, Option[W]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  153. def leftOuterJoin[W](other: RDD[(K, W)], partitioner: Partitioner): RDD[(K, (V, Option[W]))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  154. def limit(rowLimit: Long): Self

    Adds the limit clause to CQL select statement.

    Adds the limit clause to CQL select statement. The limit will be applied for each created Spark partition. In other words, unless the data are fetched from a single Cassandra partition the number of results is unpredictable.

    The main purpose of passing limit clause is to fetch top n rows from a single Cassandra partition when the table is designed so that it uses clustering keys and a partition key predicate is passed to the where clause.

    Definition Classes
    CassandraRDD
  155. val limit: Option[CassandraLimit]

    Definition Classes
    CassandraTableScanRDDCassandraRDD
  156. def localCheckpoint(): CassandraTableScanRDD.this.type

    Definition Classes
    RDD
  157. def log: Logger

    Attributes
    protected
    Definition Classes
    Logging
  158. def logDebug(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  159. def logDebug(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  160. def logError(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  161. def logError(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  162. def logInfo(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  163. def logInfo(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  164. def logName: String

    Attributes
    protected
    Definition Classes
    Logging
  165. def logTrace(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  166. def logTrace(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  167. def logWarning(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  168. def logWarning(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  169. def lookup(key: K): Seq[V]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  170. def map[U](f: (R) ⇒ U)(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
  171. def mapPartitions[U](f: (Iterator[R]) ⇒ Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
  172. def mapPartitionsWithIndex[U](f: (Int, Iterator[R]) ⇒ Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
  173. def mapValues[U](f: (V) ⇒ U): RDD[(K, U)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  174. def max()(implicit ord: Ordering[R]): R

    Definition Classes
    RDD
  175. def min()(implicit ord: Ordering[R]): R

    Definition Classes
    RDD
  176. var name: String

    Definition Classes
    RDD
  177. def narrowColumnSelection(columns: Seq[ColumnRef]): Seq[ColumnRef]

    Filters currently selected set of columns with a new set of columns

    Filters currently selected set of columns with a new set of columns

    Definition Classes
    CassandraTableRowReaderProvider
  178. final def ne(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  179. final def notify(): Unit

    Definition Classes
    AnyRef
  180. final def notifyAll(): Unit

    Definition Classes
    AnyRef
  181. def parent[U](j: Int)(implicit arg0: ClassTag[U]): RDD[U]

    Attributes
    protected[org.apache.spark]
    Definition Classes
    RDD
  182. def partitionBy(partitioner: Partitioner): RDD[(K, V)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  183. lazy val partitionGenerator: CassandraPartitionGenerator[V, T]

  184. val partitioner: Option[Partitioner]

    Definition Classes
    CassandraTableScanRDD → RDD
  185. final def partitions: Array[Partition]

    Definition Classes
    RDD
  186. def perPartitionLimit(rowLimit: Long): Self

    Adds the PER PARTITION LIMIT clause to CQL select statement.

    Adds the PER PARTITION LIMIT clause to CQL select statement. The limit will be applied for every Cassandra Partition. Only Valid For Cassandra 3.6+

    Definition Classes
    CassandraRDD
  187. def persist(): CassandraTableScanRDD.this.type

    Definition Classes
    RDD
  188. def persist(newLevel: StorageLevel): CassandraTableScanRDD.this.type

    Definition Classes
    RDD
  189. def pipe(command: Seq[String], env: Map[String, String], printPipeContext: ((String) ⇒ Unit) ⇒ Unit, printRDDElement: (R, (String) ⇒ Unit) ⇒ Unit, separateWorkingDir: Boolean, bufferSize: Int, encoding: String): RDD[String]

    Definition Classes
    RDD
  190. def pipe(command: String, env: Map[String, String]): RDD[String]

    Definition Classes
    RDD
  191. def pipe(command: String): RDD[String]

    Definition Classes
    RDD
  192. final def preferredLocations(split: Partition): Seq[String]

    Definition Classes
    RDD
  193. def randomSplit(weights: Array[Double], seed: Long): Array[RDD[R]]

    Definition Classes
    RDD
  194. val readConf: ReadConf

  195. def reduce(f: (R, R) ⇒ R): R

    Definition Classes
    RDD
  196. def reduceByKey(func: (V, V) ⇒ V): RDD[(K, V)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  197. def reduceByKey(func: (V, V) ⇒ V, numPartitions: Int): RDD[(K, V)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  198. def reduceByKey(partitioner: Partitioner, func: (V, V) ⇒ V): RDD[(K, V)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  199. def reduceByKeyLocally(func: (V, V) ⇒ V): Map[K, V]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  200. def repartition(numPartitions: Int)(implicit ord: Ordering[R]): RDD[R]

    Definition Classes
    RDD
  201. def repartitionAndSortWithinPartitions(partitioner: Partitioner): RDD[(K, V)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to OrderedRDDFunctions[K, V, (K, V)] performed by method rddToOrderedRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type Ordering[K] is in scope
    2. an implicit value of type ClassTag[K] is in scope
    3. an implicit value of type ClassTag[V] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    OrderedRDDFunctions
  202. def repartitionByCassandraReplica(keyspaceName: String, tableName: String, partitionsPerHost: Int = 10, partitionKeyMapper: ColumnSelector = PartitionKeyColumns)(implicit connector: CassandraConnector = CassandraConnector(sparkContext), currentType: ClassTag[R], rwf: RowWriterFactory[R]): CassandraPartitionedRDD[R]

    Repartitions the data (via a shuffle) based upon the replication of the given keyspaceName and tableName.

    Repartitions the data (via a shuffle) based upon the replication of the given keyspaceName and tableName. Calling this method before using joinWithCassandraTable will ensure that requests will be coordinator local. partitionsPerHost Controls the number of Spark Partitions that will be created in this repartitioning event. The calling RDD must have rows that can be converted into the partition key of the given Cassandra Table.

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to RDDFunctions[R] performed by method toRDDFunctions in com.datastax.spark.connector.
    Definition Classes
    RDDFunctions
  203. def rightOuterJoin[W](other: RDD[(K, W)], numPartitions: Int): RDD[(K, (Option[V], W))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  204. def rightOuterJoin[W](other: RDD[(K, W)]): RDD[(K, (Option[V], W))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  205. def rightOuterJoin[W](other: RDD[(K, W)], partitioner: Partitioner): RDD[(K, (Option[V], W))]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  206. lazy val rowReader: RowReader[R]

  207. implicit val rowReaderFactory: RowReaderFactory[R]

    RowReaderFactory and ClassTag should be provided from implicit parameters in the constructor of the class implementing this trait

    RowReaderFactory and ClassTag should be provided from implicit parameters in the constructor of the class implementing this trait

    Definition Classes
    CassandraTableScanRDDCassandraTableRowReaderProvider
    See also

    CassandraTableScanRDD

  208. def sample(withReplacement: Boolean, fraction: Double, seed: Long): RDD[R]

    Definition Classes
    RDD
  209. def sampleByKey(withReplacement: Boolean, fractions: Map[K, Double], seed: Long): RDD[(K, V)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  210. def sampleByKeyExact(withReplacement: Boolean, fractions: Map[K, Double], seed: Long): RDD[(K, V)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  211. def saveAsCassandraTable(keyspaceName: String, tableName: String, columns: ColumnSelector = AllColumns, writeConf: WriteConf = ...)(implicit connector: CassandraConnector = CassandraConnector(sparkContext), rwf: RowWriterFactory[R], columnMapper: ColumnMapper[R]): Unit

    Saves the data from RDD to a new table with definition taken from the ColumnMapper for this class.

    Saves the data from RDD to a new table with definition taken from the ColumnMapper for this class.

    keyspaceName

    keyspace where to create a new table

    tableName

    name of the table to create; the table must not exist

    columns

    Selects the columns to save data to. Uses only the unique column names, and you must select at least all primary key columns. All other fields are discarded. Non-selected property/column names are left unchanged. This parameter does not affect table creation.

    writeConf

    additional configuration object allowing to set consistency level, batch size, etc.

    connector

    optional, implicit connector to Cassandra

    rwf

    factory for obtaining the row writer to be used to extract column values from items of the RDD

    columnMapper

    a column mapper determining the definition of the table

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to RDDFunctions[R] performed by method toRDDFunctions in com.datastax.spark.connector.
    Definition Classes
    RDDFunctions
  212. def saveAsCassandraTableEx(table: TableDef, columns: ColumnSelector = AllColumns, writeConf: WriteConf = ...)(implicit connector: CassandraConnector = CassandraConnector(sparkContext), rwf: RowWriterFactory[R]): Unit

    Saves the data from RDD to a new table defined by the given TableDef.

    Saves the data from RDD to a new table defined by the given TableDef.

    First it creates a new table with all columns from the TableDef and then it saves RDD content in the same way as saveToCassandra. The table must not exist prior to this call.

    table

    table definition used to create a new table

    columns

    Selects the columns to save data to. Uses only the unique column names, and you must select at least all primary key columns. All other fields are discarded. Non-selected property/column names are left unchanged. This parameter does not affect table creation.

    writeConf

    additional configuration object allowing to set consistency level, batch size, etc.

    connector

    optional, implicit connector to Cassandra

    rwf

    factory for obtaining the row writer to be used to extract column values from items of the RDD

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to RDDFunctions[R] performed by method toRDDFunctions in com.datastax.spark.connector.
    Definition Classes
    RDDFunctions
  213. def saveAsHadoopDataset(conf: JobConf): Unit

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  214. def saveAsHadoopFile(path: String, keyClass: Class[_], valueClass: Class[_], outputFormatClass: Class[_ <: OutputFormat[_, _]], conf: JobConf, codec: Option[Class[_ <: CompressionCodec]]): Unit

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  215. def saveAsHadoopFile(path: String, keyClass: Class[_], valueClass: Class[_], outputFormatClass: Class[_ <: OutputFormat[_, _]], codec: Class[_ <: CompressionCodec]): Unit

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  216. def saveAsHadoopFile[F <: OutputFormat[K, V]](path: String, codec: Class[_ <: CompressionCodec])(implicit fm: ClassTag[F]): Unit

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  217. def saveAsHadoopFile[F <: OutputFormat[K, V]](path: String)(implicit fm: ClassTag[F]): Unit

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  218. def saveAsNewAPIHadoopDataset(conf: Configuration): Unit

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  219. def saveAsNewAPIHadoopFile(path: String, keyClass: Class[_], valueClass: Class[_], outputFormatClass: Class[_ <: OutputFormat[_, _]], conf: Configuration): Unit

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  220. def saveAsNewAPIHadoopFile[F <: OutputFormat[K, V]](path: String)(implicit fm: ClassTag[F]): Unit

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  221. def saveAsObjectFile(path: String): Unit

    Definition Classes
    RDD
  222. def saveAsSequenceFile(path: String, codec: Option[Class[_ <: CompressionCodec]]): Unit

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to SequenceFileRDDFunctions[K, V] performed by method rddToSequenceFileRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type WritableFactory[K] is in scope
    4. an implicit value of type WritableFactory[V] is in scope
    5. R is (K, V) (R =:= (K, V))
    Definition Classes
    SequenceFileRDDFunctions
  223. def saveAsTextFile(path: String, codec: Class[_ <: CompressionCodec]): Unit

    Definition Classes
    RDD
  224. def saveAsTextFile(path: String): Unit

    Definition Classes
    RDD
  225. def saveToCassandra(keyspaceName: String, tableName: String, columns: ColumnSelector = AllColumns, writeConf: WriteConf = ...)(implicit connector: CassandraConnector = CassandraConnector(sparkContext), rwf: RowWriterFactory[R]): Unit

    Saves the data from RDD to a Cassandra table.

    Saves the data from RDD to a Cassandra table. Uses the specified column names.

    keyspaceName

    the name of the Keyspace to use

    tableName

    the name of the Table to use

    writeConf

    additional configuration object allowing to set consistency level, batch size, etc.

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to RDDFunctions[R] performed by method toRDDFunctions in com.datastax.spark.connector.
    Definition Classes
    RDDFunctionsWritableToCassandra
    See also

    com.datastax.spark.connector.writer.WritableToCassandra

  226. val sc: SparkContext

  227. def select(columns: ColumnRef*): Self

    Narrows down the selected set of columns.

    Narrows down the selected set of columns. Use this for better performance, when you don't need all the columns in the result RDD. When called multiple times, it selects the subset of the already selected columns, so after a column was removed by the previous select call, it is not possible to add it back.

    The selected columns are ColumnRef instances. This type allows to specify columns for straightforward retrieval and to read TTL or write time of regular columns as well. Implicit conversions included in com.datastax.spark.connector package make it possible to provide just column names (which is also backward compatible) and optional add .ttl or .writeTime suffix in order to create an appropriate ColumnRef instance.

    Definition Classes
    CassandraRDD
  228. def selectedColumnNames: Seq[String]

    Definition Classes
    CassandraRDD
  229. lazy val selectedColumnRefs: Seq[ColumnRef]

    Returns the columns to be selected from the table.

    Returns the columns to be selected from the table.

    Definition Classes
    CassandraTableRowReaderProvider
  230. def setName(_name: String): CassandraTableScanRDD.this.type

    Definition Classes
    RDD
  231. def sortBy[K](f: (R) ⇒ K, ascending: Boolean, numPartitions: Int)(implicit ord: Ordering[K], ctag: ClassTag[K]): RDD[R]

    Definition Classes
    RDD
  232. def sortByKey(ascending: Boolean, numPartitions: Int): RDD[(K, V)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to OrderedRDDFunctions[K, V, (K, V)] performed by method rddToOrderedRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type Ordering[K] is in scope
    2. an implicit value of type ClassTag[K] is in scope
    3. an implicit value of type ClassTag[V] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    OrderedRDDFunctions
  233. def spanBy[U](f: (R) ⇒ U): RDD[(U, Iterable[R])]

    Applies a function to each item, and groups consecutive items having the same value together.

    Applies a function to each item, and groups consecutive items having the same value together. Contrary to groupBy, items from the same group must be already next to each other in the original collection. Works locally on each partition, so items from different partitions will never be placed in the same group.

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to RDDFunctions[R] performed by method toRDDFunctions in com.datastax.spark.connector.
    Definition Classes
    RDDFunctions
  234. def spanByKey: RDD[(K, Seq[V])]

    Groups items with the same key, assuming the items with the same key are next to each other in the collection.

    Groups items with the same key, assuming the items with the same key are next to each other in the collection. It does not perform shuffle, therefore it is much faster than using much more universal Spark RDD groupByKey. For this method to be useful with Cassandra tables, the key must represent a prefix of the primary key, containing at least the partition key of the Cassandra table.

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to PairRDDFunctions[K, V] performed by method toPairRDDFunctions in com.datastax.spark.connector. This conversion will take place only if R is (K, V) (R =:= (K, V)).
    Definition Classes
    PairRDDFunctions
  235. def sparkContext: SparkContext

    Definition Classes
    RDD
  236. def splitCount: Option[Int]

    Attributes
    protected
    Definition Classes
    CassandraTableRowReaderProvider
  237. def splitSize: Long

    Attributes
    protected[com.datastax.spark.connector]
    Definition Classes
    CassandraTableRowReaderProvider
  238. def subtract(other: RDD[R], p: Partitioner)(implicit ord: Ordering[R]): RDD[R]

    Definition Classes
    RDD
  239. def subtract(other: RDD[R], numPartitions: Int): RDD[R]

    Definition Classes
    RDD
  240. def subtract(other: RDD[R]): RDD[R]

    Definition Classes
    RDD
  241. def subtractByKey[W](other: RDD[(K, W)], p: Partitioner)(implicit arg0: ClassTag[W]): RDD[(K, V)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  242. def subtractByKey[W](other: RDD[(K, W)], numPartitions: Int)(implicit arg0: ClassTag[W]): RDD[(K, V)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  243. def subtractByKey[W](other: RDD[(K, W)])(implicit arg0: ClassTag[W]): RDD[(K, V)]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  244. final def synchronized[T0](arg0: ⇒ T0): T0

    Definition Classes
    AnyRef
  245. lazy val tableDef: TableDef

  246. val tableName: String

  247. def take(num: Int): Array[R]

    Definition Classes
    CassandraRDD → RDD
  248. def takeAsync(num: Int): FutureAction[Seq[R]]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to AsyncRDDActions[R] performed by method rddToAsyncRDDActions in org.apache.spark.rdd.RDD. This conversion will take place only if R is accompanied by a ClassTag, which is a runtime representation of its type that survives erasure (R: ClassTag).
    Definition Classes
    AsyncRDDActions
  249. def takeOrdered(num: Int)(implicit ord: Ordering[R]): Array[R]

    Definition Classes
    RDD
  250. def takeSample(withReplacement: Boolean, num: Int, seed: Long): Array[R]

    Definition Classes
    RDD
  251. def toDebugString: String

    Definition Classes
    RDD
  252. def toEmptyCassandraRDD: EmptyCassandraRDD[R]

    Definition Classes
    CassandraTableScanRDDCassandraRDD
  253. def toJavaRDD(): JavaRDD[R]

    Definition Classes
    RDD
  254. def toLocalIterator: Iterator[R]

    Definition Classes
    RDD
  255. def toString(): String

    Definition Classes
    RDD → AnyRef → Any
  256. implicit lazy val tokenFactory: TokenFactory[V, T]

    Definition Classes
    SplitSizeEstimator
  257. def top(num: Int)(implicit ord: Ordering[R]): Array[R]

    Definition Classes
    RDD
  258. def treeAggregate[U](zeroValue: U)(seqOp: (U, R) ⇒ U, combOp: (U, U) ⇒ U, depth: Int)(implicit arg0: ClassTag[U]): U

    Definition Classes
    RDD
  259. def treeReduce(f: (R, R) ⇒ R, depth: Int): R

    Definition Classes
    RDD
  260. def union(other: RDD[R]): RDD[R]

    Definition Classes
    RDD
  261. def unpersist(blocking: Boolean): CassandraTableScanRDD.this.type

    Definition Classes
    RDD
  262. def values: RDD[V]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V] performed by method rddToPairRDDFunctions in org.apache.spark.rdd.RDD.
    This conversion will take place only if all of the following constraints are met:
    1. an implicit value of type ClassTag[K] is in scope
    2. an implicit value of type ClassTag[V] is in scope
    3. an implicit value of type Ordering[K] is in scope
    4. R is (K, V) (R =:= (K, V))
    Definition Classes
    PairRDDFunctions
  263. def verify(): RowReader[R]

    Checks for existence of keyspace and table.

    Checks for existence of keyspace and table.

    Definition Classes
    CassandraTableRowReaderProvider
  264. final def wait(): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  265. final def wait(arg0: Long, arg1: Int): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  266. final def wait(arg0: Long): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  267. def where(cql: String, values: Any*): Self

    Adds a CQL WHERE predicate(s) to the query.

    Adds a CQL WHERE predicate(s) to the query. Useful for leveraging secondary indexes in Cassandra. Implicitly adds an ALLOW FILTERING clause to the WHERE clause, however beware that some predicates might be rejected by Cassandra, particularly in cases when they filter on an unindexed, non-clustering column.

    Definition Classes
    CassandraRDD
  268. val where: CqlWhereClause

    Definition Classes
    CassandraTableScanRDDCassandraRDD
  269. def withAscOrder: Self

    Definition Classes
    CassandraRDD
  270. def withConnector(connector: CassandraConnector): Self

    Returns a copy of this Cassandra RDD with specified connector

    Returns a copy of this Cassandra RDD with specified connector

    Definition Classes
    CassandraRDD
  271. def withDescOrder: Self

    Definition Classes
    CassandraRDD
  272. def withReadConf(readConf: ReadConf): Self

    Allows to set custom read configuration, e.

    Allows to set custom read configuration, e.g. consistency level or fetch size.

    Definition Classes
    CassandraRDD
  273. def zip[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(R, U)]

    Definition Classes
    RDD
  274. def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D])(f: (Iterator[R], Iterator[B], Iterator[C], Iterator[D]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  275. def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D], preservesPartitioning: Boolean)(f: (Iterator[R], Iterator[B], Iterator[C], Iterator[D]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  276. def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C])(f: (Iterator[R], Iterator[B], Iterator[C]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  277. def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C], preservesPartitioning: Boolean)(f: (Iterator[R], Iterator[B], Iterator[C]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  278. def zipPartitions[B, V](rdd2: RDD[B])(f: (Iterator[R], Iterator[B]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  279. def zipPartitions[B, V](rdd2: RDD[B], preservesPartitioning: Boolean)(f: (Iterator[R], Iterator[B]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  280. def zipWithIndex(): RDD[(R, Long)]

    Definition Classes
    RDD
  281. def zipWithUniqueId(): RDD[(R, Long)]

    Definition Classes
    RDD
  282. def [B](y: B): (CassandraTableScanRDD[R], B)

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to ArrowAssoc[CassandraTableScanRDD[R]] performed by method any2ArrowAssoc in scala.Predef.
    Definition Classes
    ArrowAssoc

Shadowed Implicit Value Members

  1. def histogram(buckets: Array[Double], evenBuckets: Boolean): Array[Long]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method numericRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is a numeric class, such as Int, Long, Float or Double (R: Numeric).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).histogram(buckets, evenBuckets)
    Definition Classes
    DoubleRDDFunctions
  2. def histogram(bucketCount: Int): (Array[Double], Array[Long])

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method numericRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is a numeric class, such as Int, Long, Float or Double (R: Numeric).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).histogram(bucketCount)
    Definition Classes
    DoubleRDDFunctions
  3. def histogram(buckets: Array[Double], evenBuckets: Boolean): Array[Long]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method doubleRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is Double (R =:= Double).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).histogram(buckets, evenBuckets)
    Definition Classes
    DoubleRDDFunctions
  4. def histogram(bucketCount: Int): (Array[Double], Array[Long])

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method doubleRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is Double (R =:= Double).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).histogram(bucketCount)
    Definition Classes
    DoubleRDDFunctions
  5. def mean(): Double

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method numericRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is a numeric class, such as Int, Long, Float or Double (R: Numeric).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).mean()
    Definition Classes
    DoubleRDDFunctions
  6. def mean(): Double

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method doubleRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is Double (R =:= Double).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).mean()
    Definition Classes
    DoubleRDDFunctions
  7. def meanApprox(timeout: Long, confidence: Double): PartialResult[BoundedDouble]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method numericRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is a numeric class, such as Int, Long, Float or Double (R: Numeric).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).meanApprox(timeout, confidence)
    Definition Classes
    DoubleRDDFunctions
  8. def meanApprox(timeout: Long, confidence: Double): PartialResult[BoundedDouble]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method doubleRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is Double (R =:= Double).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).meanApprox(timeout, confidence)
    Definition Classes
    DoubleRDDFunctions
  9. def sampleStdev(): Double

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method numericRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is a numeric class, such as Int, Long, Float or Double (R: Numeric).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).sampleStdev()
    Definition Classes
    DoubleRDDFunctions
  10. def sampleStdev(): Double

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method doubleRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is Double (R =:= Double).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).sampleStdev()
    Definition Classes
    DoubleRDDFunctions
  11. def sampleVariance(): Double

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method numericRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is a numeric class, such as Int, Long, Float or Double (R: Numeric).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).sampleVariance()
    Definition Classes
    DoubleRDDFunctions
  12. def sampleVariance(): Double

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method doubleRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is Double (R =:= Double).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).sampleVariance()
    Definition Classes
    DoubleRDDFunctions
  13. val self: Any

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to StringAdd performed by method any2stringadd in scala.Predef.
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: StringAdd).self
    Definition Classes
    StringAdd
  14. val self: Any

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to StringFormat performed by method any2stringfmt in scala.Predef.
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: StringFormat).self
    Definition Classes
    StringFormat
  15. val sparkContext: SparkContext

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to RDDFunctions[R] performed by method toRDDFunctions in com.datastax.spark.connector.
    Shadowing
    This implicitly inherited member is shadowed by one or more members in this class.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: RDDFunctions[R]).sparkContext
    Definition Classes
    RDDFunctionsWritableToCassandra
  16. def stats(): StatCounter

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method numericRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is a numeric class, such as Int, Long, Float or Double (R: Numeric).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).stats()
    Definition Classes
    DoubleRDDFunctions
  17. def stats(): StatCounter

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method doubleRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is Double (R =:= Double).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).stats()
    Definition Classes
    DoubleRDDFunctions
  18. def stdev(): Double

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method numericRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is a numeric class, such as Int, Long, Float or Double (R: Numeric).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).stdev()
    Definition Classes
    DoubleRDDFunctions
  19. def stdev(): Double

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method doubleRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is Double (R =:= Double).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).stdev()
    Definition Classes
    DoubleRDDFunctions
  20. def sum(): Double

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method numericRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is a numeric class, such as Int, Long, Float or Double (R: Numeric).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).sum()
    Definition Classes
    DoubleRDDFunctions
  21. def sum(): Double

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method doubleRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is Double (R =:= Double).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).sum()
    Definition Classes
    DoubleRDDFunctions
  22. def sumApprox(timeout: Long, confidence: Double): PartialResult[BoundedDouble]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method numericRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is a numeric class, such as Int, Long, Float or Double (R: Numeric).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).sumApprox(timeout, confidence)
    Definition Classes
    DoubleRDDFunctions
  23. def sumApprox(timeout: Long, confidence: Double): PartialResult[BoundedDouble]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method doubleRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is Double (R =:= Double).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).sumApprox(timeout, confidence)
    Definition Classes
    DoubleRDDFunctions
  24. def variance(): Double

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method numericRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is a numeric class, such as Int, Long, Float or Double (R: Numeric).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).variance()
    Definition Classes
    DoubleRDDFunctions
  25. def variance(): Double

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to DoubleRDDFunctions performed by method doubleRDDToDoubleRDDFunctions in org.apache.spark.rdd.RDD. This conversion will take place only if R is Double (R =:= Double).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: DoubleRDDFunctions).variance()
    Definition Classes
    DoubleRDDFunctions

Deprecated Value Members

  1. def x: CassandraTableScanRDD[R]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to ArrowAssoc[CassandraTableScanRDD[R]] performed by method any2ArrowAssoc in scala.Predef.
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: ArrowAssoc[CassandraTableScanRDD[R]]).x
    Definition Classes
    ArrowAssoc
    Annotations
    @deprecated
    Deprecated

    (Since version 2.10.0) Use leftOfArrow instead

  2. def x: CassandraTableScanRDD[R]

    Implicit information
    This member is added by an implicit conversion from CassandraTableScanRDD[R] to Ensuring[CassandraTableScanRDD[R]] performed by method any2Ensuring in scala.Predef. This conversion will take place only if R is a superclass of Any and a subclass of (Nothing, Nothing) with Double (R >: Any <: (Nothing, Nothing) with Double).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraTableScanRDD: Ensuring[CassandraTableScanRDD[R]]).x
    Definition Classes
    Ensuring
    Annotations
    @deprecated
    Deprecated

    (Since version 2.10.0) Use resultOfEnsuring instead

Inherited from SplitSizeEstimator[R]

Inherited from CassandraRDD[R]

Inherited from RDD[R]

Inherited from Logging

Inherited from Serializable

Inherited from Serializable

Inherited from AnyRef

Inherited from Any

Inherited by implicit conversion rddToPairRDDFunctions from CassandraTableScanRDD[R] to org.apache.spark.rdd.PairRDDFunctions[K, V]

Inherited by implicit conversion numericRDDToDoubleRDDFunctions from CassandraTableScanRDD[R] to DoubleRDDFunctions

Inherited by implicit conversion doubleRDDToDoubleRDDFunctions from CassandraTableScanRDD[R] to DoubleRDDFunctions

Inherited by implicit conversion rddToOrderedRDDFunctions from CassandraTableScanRDD[R] to OrderedRDDFunctions[K, V, (K, V)]

Inherited by implicit conversion rddToSequenceFileRDDFunctions from CassandraTableScanRDD[R] to SequenceFileRDDFunctions[K, V]

Inherited by implicit conversion rddToAsyncRDDActions from CassandraTableScanRDD[R] to AsyncRDDActions[R]

Inherited by implicit conversion toCassandraTableScanRDDPairFunctions from CassandraTableScanRDD[R] to CassandraTableScanPairRDDFunctions[K, V]

Inherited by implicit conversion toPairRDDFunctions from CassandraTableScanRDD[R] to PairRDDFunctions[K, V]

Inherited by implicit conversion toCassandraTableScanFunctions from CassandraTableScanRDD[R] to CassandraTableScanRDDFunctions[R]

Inherited by implicit conversion toRDDFunctions from CassandraTableScanRDD[R] to RDDFunctions[R]

Inherited by implicit conversion any2stringadd from CassandraTableScanRDD[R] to StringAdd

Inherited by implicit conversion any2stringfmt from CassandraTableScanRDD[R] to StringFormat

Inherited by implicit conversion any2ArrowAssoc from CassandraTableScanRDD[R] to ArrowAssoc[CassandraTableScanRDD[R]]

Inherited by implicit conversion any2Ensuring from CassandraTableScanRDD[R] to Ensuring[CassandraTableScanRDD[R]]

Ungrouped