com.datastax.spark.connector.streaming

CassandraStreamingRDD

class CassandraStreamingRDD[R] extends CassandraRDD[R]

RDD representing a Cassandra table for Spark Streaming.

See also

com.datastax.spark.connector.rdd.CassandraRDD

Linear Supertypes
CassandraRDD[R], RDD[R], Logging, Serializable, Serializable, AnyRef, Any
Ordering
  1. Alphabetic
  2. By inheritance
Inherited
  1. CassandraStreamingRDD
  2. CassandraRDD
  3. RDD
  4. Logging
  5. Serializable
  6. Serializable
  7. AnyRef
  8. Any
Implicitly
  1. by toPairRDDFunctions
  2. by toRDDFunctions
  3. by any2stringadd
  4. by any2stringfmt
  5. by any2ArrowAssoc
  6. by any2Ensuring
  1. Hide All
  2. Show all
Learn more about member selection
Visibility
  1. Public
  2. All

Value Members

  1. final def !=(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  2. final def !=(arg0: Any): Boolean

    Definition Classes
    Any
  3. final def ##(): Int

    Definition Classes
    AnyRef → Any
  4. def +(other: String): String

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to StringAdd performed by method any2stringadd in scala.Predef.
    Definition Classes
    StringAdd
  5. def ++(other: RDD[R]): RDD[R]

    Definition Classes
    RDD
  6. def ->[B](y: B): (CassandraStreamingRDD[R], B)

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to ArrowAssoc[CassandraStreamingRDD[R]] performed by method any2ArrowAssoc in scala.Predef.
    Definition Classes
    ArrowAssoc
    Annotations
    @inline()
  7. final def ==(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  8. final def ==(arg0: Any): Boolean

    Definition Classes
    Any
  9. def aggregate[U](zeroValue: U)(seqOp: (U, R) ⇒ U, combOp: (U, U) ⇒ U)(implicit arg0: ClassTag[U]): U

    Definition Classes
    RDD
  10. def as[B, A0, A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11](f: (A0, A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6], arg8: TypeConverter[A7], arg9: TypeConverter[A8], arg10: TypeConverter[A9], arg11: TypeConverter[A10], arg12: TypeConverter[A11]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  11. def as[B, A0, A1, A2, A3, A4, A5, A6, A7, A8, A9, A10](f: (A0, A1, A2, A3, A4, A5, A6, A7, A8, A9, A10) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6], arg8: TypeConverter[A7], arg9: TypeConverter[A8], arg10: TypeConverter[A9], arg11: TypeConverter[A10]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  12. def as[B, A0, A1, A2, A3, A4, A5, A6, A7, A8, A9](f: (A0, A1, A2, A3, A4, A5, A6, A7, A8, A9) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6], arg8: TypeConverter[A7], arg9: TypeConverter[A8], arg10: TypeConverter[A9]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  13. def as[B, A0, A1, A2, A3, A4, A5, A6, A7, A8](f: (A0, A1, A2, A3, A4, A5, A6, A7, A8) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6], arg8: TypeConverter[A7], arg9: TypeConverter[A8]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  14. def as[B, A0, A1, A2, A3, A4, A5, A6, A7](f: (A0, A1, A2, A3, A4, A5, A6, A7) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6], arg8: TypeConverter[A7]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  15. def as[B, A0, A1, A2, A3, A4, A5, A6](f: (A0, A1, A2, A3, A4, A5, A6) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  16. def as[B, A0, A1, A2, A3, A4, A5](f: (A0, A1, A2, A3, A4, A5) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  17. def as[B, A0, A1, A2, A3, A4](f: (A0, A1, A2, A3, A4) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  18. def as[B, A0, A1, A2, A3](f: (A0, A1, A2, A3) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  19. def as[B, A0, A1, A2](f: (A0, A1, A2) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  20. def as[B, A0, A1](f: (A0, A1) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1]): CassandraRDD[B]

    Definition Classes
    CassandraRDD
  21. def as[B, A0](f: (A0) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0]): CassandraRDD[B]

    Maps each row into object of a different type using provided function taking column value(s) as argument(s).

    Maps each row into object of a different type using provided function taking column value(s) as argument(s). Can be used to convert each row to a tuple or a case class object:

    sc.cassandraTable("ks", "table").select("column1").as((s: String) => s)                 // yields CassandraRDD[String]
    sc.cassandraTable("ks", "table").select("column1", "column2").as((_: String, _: Long))  // yields CassandraRDD[(String, Long)]
    
    case class MyRow(key: String, value: Long)
    sc.cassandraTable("ks", "table").select("column1", "column2").as(MyRow)                 // yields CassandraRDD[MyRow]
    Definition Classes
    CassandraRDD
  22. final def asInstanceOf[T0]: T0

    Definition Classes
    Any
  23. def cache(): CassandraStreamingRDD.this.type

    Definition Classes
    RDD
  24. def cartesian[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(R, U)]

    Definition Classes
    RDD
  25. def checkpoint(): Unit

    Definition Classes
    RDD
  26. def clearDependencies(): Unit

    Attributes
    protected
    Definition Classes
    RDD
  27. def clone(): AnyRef

    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  28. def clusteringOrder(order: ClusteringOrder): CassandraRDD[R]

    Adds a CQL ORDER BY clause to the query.

    Adds a CQL ORDER BY clause to the query. It can be applied only in case there are clustering columns and primary key predicate is pushed down in where. It is useful when the default direction of ordering rows within a single Cassandra partition needs to be changed.

    Definition Classes
    CassandraRDD
  29. def coalesce(numPartitions: Int, shuffle: Boolean)(implicit ord: Ordering[R]): RDD[R]

    Definition Classes
    RDD
  30. def collect[U](f: PartialFunction[R, U])(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
  31. def collect(): Array[R]

    Definition Classes
    RDD
  32. val columnNames: ColumnSelector

    Definition Classes
    CassandraRDD
  33. def compute(split: Partition, context: TaskContext): Iterator[R]

    Definition Classes
    CassandraRDD → RDD
  34. def context: SparkContext

    Definition Classes
    RDD
  35. def count(): Long

    Definition Classes
    CassandraRDD → RDD
  36. def countApprox(timeout: Long, confidence: Double): PartialResult[BoundedDouble]

    Definition Classes
    RDD
    Annotations
    @Experimental()
  37. def countApproxDistinct(relativeSD: Double): Long

    Definition Classes
    RDD
  38. def countApproxDistinct(p: Int, sp: Int): Long

    Definition Classes
    RDD
    Annotations
    @Experimental()
  39. def countByValue()(implicit ord: Ordering[R]): Map[R, Long]

    Definition Classes
    RDD
  40. def countByValueApprox(timeout: Long, confidence: Double)(implicit ord: Ordering[R]): PartialResult[Map[R, BoundedDouble]]

    Definition Classes
    RDD
    Annotations
    @Experimental()
  41. final def dependencies: Seq[Dependency[_]]

    Definition Classes
    RDD
  42. def distinct(): RDD[R]

    Definition Classes
    RDD
  43. def distinct(numPartitions: Int)(implicit ord: Ordering[R]): RDD[R]

    Definition Classes
    RDD
  44. def ensuring(cond: (CassandraStreamingRDD[R]) ⇒ Boolean, msg: ⇒ Any): CassandraStreamingRDD[R]

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to Ensuring[CassandraStreamingRDD[R]] performed by method any2Ensuring in scala.Predef. This conversion will take place only if R is (Nothing, Nothing) (R =:= (Nothing, Nothing)).
    Definition Classes
    Ensuring
  45. def ensuring(cond: (CassandraStreamingRDD[R]) ⇒ Boolean): CassandraStreamingRDD[R]

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to Ensuring[CassandraStreamingRDD[R]] performed by method any2Ensuring in scala.Predef. This conversion will take place only if R is (Nothing, Nothing) (R =:= (Nothing, Nothing)).
    Definition Classes
    Ensuring
  46. def ensuring(cond: Boolean, msg: ⇒ Any): CassandraStreamingRDD[R]

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to Ensuring[CassandraStreamingRDD[R]] performed by method any2Ensuring in scala.Predef. This conversion will take place only if R is (Nothing, Nothing) (R =:= (Nothing, Nothing)).
    Definition Classes
    Ensuring
  47. def ensuring(cond: Boolean): CassandraStreamingRDD[R]

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to Ensuring[CassandraStreamingRDD[R]] performed by method any2Ensuring in scala.Predef. This conversion will take place only if R is (Nothing, Nothing) (R =:= (Nothing, Nothing)).
    Definition Classes
    Ensuring
  48. final def eq(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  49. def equals(arg0: Any): Boolean

    Definition Classes
    AnyRef → Any
  50. def filter(f: (R) ⇒ Boolean): RDD[R]

    Definition Classes
    RDD
  51. def finalize(): Unit

    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  52. def first(): R

    Definition Classes
    RDD
  53. def firstParent[U](implicit arg0: ClassTag[U]): RDD[U]

    Attributes
    protected[org.apache.spark]
    Definition Classes
    RDD
  54. def flatMap[U](f: (R) ⇒ TraversableOnce[U])(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
  55. def fold(zeroValue: R)(op: (R, R) ⇒ R): R

    Definition Classes
    RDD
  56. def foreach(f: (R) ⇒ Unit): Unit

    Definition Classes
    RDD
  57. def foreachPartition(f: (Iterator[R]) ⇒ Unit): Unit

    Definition Classes
    RDD
  58. def formatted(fmtstr: String): String

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to StringFormat performed by method any2stringfmt in scala.Predef.
    Definition Classes
    StringFormat
    Annotations
    @inline()
  59. def getCheckpointFile: Option[String]

    Definition Classes
    RDD
  60. final def getClass(): Class[_]

    Definition Classes
    AnyRef → Any
  61. def getDependencies: Seq[Dependency[_]]

    Attributes
    protected
    Definition Classes
    RDD
  62. def getPartitions: Array[Partition]

    Definition Classes
    CassandraRDD → RDD
  63. def getPreferredLocations(split: Partition): Seq[String]

    Definition Classes
    CassandraRDD → RDD
  64. def getStorageLevel: StorageLevel

    Definition Classes
    RDD
  65. def glom(): RDD[Array[R]]

    Definition Classes
    RDD
  66. def groupBy[K](f: (R) ⇒ K, p: Partitioner)(implicit kt: ClassTag[K], ord: Ordering[K]): RDD[(K, Iterable[R])]

    Definition Classes
    RDD
  67. def groupBy[K](f: (R) ⇒ K, numPartitions: Int)(implicit kt: ClassTag[K]): RDD[(K, Iterable[R])]

    Definition Classes
    RDD
  68. def groupBy[K](f: (R) ⇒ K)(implicit kt: ClassTag[K]): RDD[(K, Iterable[R])]

    Definition Classes
    RDD
  69. def hashCode(): Int

    Definition Classes
    AnyRef → Any
  70. val id: Int

    Definition Classes
    RDD
  71. def intersection(other: RDD[R], numPartitions: Int): RDD[R]

    Definition Classes
    RDD
  72. def intersection(other: RDD[R], partitioner: Partitioner)(implicit ord: Ordering[R]): RDD[R]

    Definition Classes
    RDD
  73. def intersection(other: RDD[R]): RDD[R]

    Definition Classes
    RDD
  74. def isCheckpointed: Boolean

    Definition Classes
    RDD
  75. final def isInstanceOf[T0]: Boolean

    Definition Classes
    Any
  76. def isTraceEnabled(): Boolean

    Attributes
    protected
    Definition Classes
    Logging
  77. final def iterator(split: Partition, context: TaskContext): Iterator[R]

    Definition Classes
    RDD
  78. def keyBy[K](f: (R) ⇒ K): RDD[(K, R)]

    Definition Classes
    RDD
  79. val keyspaceName: String

    Definition Classes
    CassandraRDD
  80. def limit(rowLimit: Long): CassandraRDD[R]

    Adds the limit clause to CQL select statement.

    Adds the limit clause to CQL select statement. The limit will be applied for each created Spark partition. In other words, unless the data are fetched from a single Cassandra partition the number of results is unpredictable.

    The main purpose of passing limit clause is to fetch top n rows from a single Cassandra partition when the table is designed so that it uses clustering keys and a partition key predicate is passed to the where clause.

    Definition Classes
    CassandraRDD
  81. def log: Logger

    Attributes
    protected
    Definition Classes
    Logging
  82. def logDebug(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  83. def logDebug(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  84. def logError(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  85. def logError(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  86. def logInfo(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  87. def logInfo(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  88. def logName: String

    Attributes
    protected
    Definition Classes
    Logging
  89. def logTrace(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  90. def logTrace(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  91. def logWarning(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  92. def logWarning(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  93. def map[U](f: (R) ⇒ U)(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
  94. def mapPartitions[U](f: (Iterator[R]) ⇒ Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
  95. def mapPartitionsWithIndex[U](f: (Int, Iterator[R]) ⇒ Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
  96. def max()(implicit ord: Ordering[R]): R

    Definition Classes
    RDD
  97. def min()(implicit ord: Ordering[R]): R

    Definition Classes
    RDD
  98. var name: String

    Definition Classes
    RDD
  99. final def ne(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  100. final def notify(): Unit

    Definition Classes
    AnyRef
  101. final def notifyAll(): Unit

    Definition Classes
    AnyRef
  102. def parent[U](j: Int)(implicit arg0: ClassTag[U]): RDD[U]

    Attributes
    protected[org.apache.spark]
    Definition Classes
    RDD
  103. val partitioner: Option[Partitioner]

    Definition Classes
    RDD
  104. final def partitions: Array[Partition]

    Definition Classes
    RDD
  105. def persist(): CassandraStreamingRDD.this.type

    Definition Classes
    RDD
  106. def persist(newLevel: StorageLevel): CassandraStreamingRDD.this.type

    Definition Classes
    RDD
  107. def pipe(command: Seq[String], env: Map[String, String], printPipeContext: ((String) ⇒ Unit) ⇒ Unit, printRDDElement: (R, (String) ⇒ Unit) ⇒ Unit, separateWorkingDir: Boolean): RDD[String]

    Definition Classes
    RDD
  108. def pipe(command: String, env: Map[String, String]): RDD[String]

    Definition Classes
    RDD
  109. def pipe(command: String): RDD[String]

    Definition Classes
    RDD
  110. final def preferredLocations(split: Partition): Seq[String]

    Definition Classes
    RDD
  111. def protocolVersion(session: Session): ProtocolVersion

    Definition Classes
    CassandraRDD
  112. def randomSplit(weights: Array[Double], seed: Long): Array[RDD[R]]

    Definition Classes
    RDD
  113. def reduce(f: (R, R) ⇒ R): R

    Definition Classes
    RDD
  114. def repartition(numPartitions: Int)(implicit ord: Ordering[R]): RDD[R]

    Definition Classes
    RDD
  115. implicit val rrf: RowReaderFactory[R]

  116. implicit val rtf: RowReaderFactory[R]

    Definition Classes
    CassandraRDD
  117. def sample(withReplacement: Boolean, fraction: Double, seed: Long): RDD[R]

    Definition Classes
    RDD
  118. def saveAsCassandraTable(keyspaceName: String, tableName: String, columns: ColumnSelector = AllColumns, writeConf: WriteConf = ...)(implicit connector: CassandraConnector = ..., rwf: RowWriterFactory[R], columnMapper: ColumnMapper[R]): Unit

    Saves the data from RDD to a new table with definition taken from the ColumnMapper for this class.

    Saves the data from RDD to a new table with definition taken from the ColumnMapper for this class.

    keyspaceName

    keyspace where to create a new table

    tableName

    name of the table to create; the table must not exist

    columns

    Selects the columns to save data to. Uses only the unique column names, and you must select at least all primary key columns. All other fields are discarded. Non-selected property/column names are left unchanged. This parameter does not affect table creation.

    writeConf

    additional configuration object allowing to set consistency level, batch size, etc.

    connector

    optional, implicit connector to Cassandra

    rwf

    factory for obtaining the row writer to be used to extract column values from items of the RDD

    columnMapper

    a column mapper determining the definition of the table

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to RDDFunctions[R] performed by method toRDDFunctions in com.datastax.spark.connector.
    Definition Classes
    RDDFunctions
  119. def saveAsCassandraTableEx(table: TableDef, columns: ColumnSelector = AllColumns, writeConf: WriteConf = ...)(implicit connector: CassandraConnector = ..., rwf: RowWriterFactory[R]): Unit

    Saves the data from RDD to a new table defined by the given TableDef.

    Saves the data from RDD to a new table defined by the given TableDef.

    First it creates a new table with all columns from the TableDef and then it saves RDD content in the same way as saveToCassandra. The table must not exist prior to this call.

    table

    table definition used to create a new table

    columns

    Selects the columns to save data to. Uses only the unique column names, and you must select at least all primary key columns. All other fields are discarded. Non-selected property/column names are left unchanged. This parameter does not affect table creation.

    writeConf

    additional configuration object allowing to set consistency level, batch size, etc.

    connector

    optional, implicit connector to Cassandra

    rwf

    factory for obtaining the row writer to be used to extract column values from items of the RDD

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to RDDFunctions[R] performed by method toRDDFunctions in com.datastax.spark.connector.
    Definition Classes
    RDDFunctions
  120. def saveAsObjectFile(path: String): Unit

    Definition Classes
    RDD
  121. def saveAsTextFile(path: String, codec: Class[_ <: CompressionCodec]): Unit

    Definition Classes
    RDD
  122. def saveAsTextFile(path: String): Unit

    Definition Classes
    RDD
  123. def saveToCassandra(keyspaceName: String, tableName: String, columns: ColumnSelector = AllColumns, writeConf: WriteConf = ...)(implicit connector: CassandraConnector = ..., rwf: RowWriterFactory[R]): Unit

    Saves the data from RDD to a Cassandra table.

    Saves the data from RDD to a Cassandra table. Uses the specified column names.

    keyspaceName

    the name of the Keyspace to use

    tableName

    the name of the Table to use

    writeConf

    additional configuration object allowing to set consistency level, batch size, etc.

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to RDDFunctions[R] performed by method toRDDFunctions in com.datastax.spark.connector.
    Definition Classes
    RDDFunctionsWritableToCassandra
    See also

    com.datastax.spark.connector.writer.WritableToCassandra

  124. val sc: SparkContext

    Definition Classes
    CassandraRDD
  125. def select(columns: SelectableColumnRef*): CassandraRDD[R]

    Narrows down the selected set of columns.

    Narrows down the selected set of columns. Use this for better performance, when you don't need all the columns in the result RDD. When called multiple times, it selects the subset of the already selected columns, so after a column was removed by the previous select call, it is not possible to add it back.

    The selected columns are NamedColumnRef instances. This type allows to specify columns for straightforward retrieval and to read TTL or write time of regular columns as well. Implicit conversions included in com.datastax.spark.connector package make it possible to provide just column names (which is also backward compatible) and optional add .ttl or .writeTime suffix in order to create an appropriate NamedColumnRef instance.

    Definition Classes
    CassandraRDD
  126. lazy val selectedColumnRefs: Seq[SelectableColumnRef]

    Returns the names of columns to be selected from the table.

    Returns the names of columns to be selected from the table.

    Definition Classes
    CassandraRDD
  127. def setName(_name: String): CassandraStreamingRDD.this.type

    Definition Classes
    RDD
  128. def sortBy[K](f: (R) ⇒ K, ascending: Boolean, numPartitions: Int)(implicit ord: Ordering[K], ctag: ClassTag[K]): RDD[R]

    Definition Classes
    RDD
  129. def spanBy[U](f: (R) ⇒ U): RDD[(U, Iterable[R])]

    Applies a function to each item, and groups consecutive items having the same value together.

    Applies a function to each item, and groups consecutive items having the same value together. Contrary to groupBy, items from the same group must be already next to each other in the original collection. Works locally on each partition, so items from different partitions will never be placed in the same group.

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to RDDFunctions[R] performed by method toRDDFunctions in com.datastax.spark.connector.
    Definition Classes
    RDDFunctions
  130. def spanByKey: RDD[(K, Seq[V])]

    Groups items with the same key, assuming the items with the same key are next to each other in the collection.

    Groups items with the same key, assuming the items with the same key are next to each other in the collection. It does not perform shuffle, therefore it is much faster than using much more universal Spark RDD groupByKey. For this method to be useful with Cassandra tables, the key must represent a prefix of the primary key, containing at least the partition key of the Cassandra table.

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to PairRDDFunctions[K, V] performed by method toPairRDDFunctions in com.datastax.spark.connector. This conversion will take place only if R is (K, V) (R =:= (K, V)).
    Definition Classes
    PairRDDFunctions
  131. def sparkContext: SparkContext

    Definition Classes
    RDD
  132. def subtract(other: RDD[R], p: Partitioner)(implicit ord: Ordering[R]): RDD[R]

    Definition Classes
    RDD
  133. def subtract(other: RDD[R], numPartitions: Int): RDD[R]

    Definition Classes
    RDD
  134. def subtract(other: RDD[R]): RDD[R]

    Definition Classes
    RDD
  135. final def synchronized[T0](arg0: ⇒ T0): T0

    Definition Classes
    AnyRef
  136. lazy val tableDef: TableDef

    Definition Classes
    CassandraRDD
  137. val tableName: String

    Definition Classes
    CassandraRDD
  138. def take(num: Int): Array[R]

    Definition Classes
    CassandraRDD → RDD
  139. def takeOrdered(num: Int)(implicit ord: Ordering[R]): Array[R]

    Definition Classes
    RDD
  140. def takeSample(withReplacement: Boolean, num: Int, seed: Long): Array[R]

    Definition Classes
    RDD
  141. def toDebugString: String

    Definition Classes
    RDD
  142. def toEmptyCassandraRDD: CassandraRDD[R]

    Produces the empty CassandraRDD which has the same signature and properties, but it does not perform any validation and it does not even try to return any rows.

    Produces the empty CassandraRDD which has the same signature and properties, but it does not perform any validation and it does not even try to return any rows.

    Definition Classes
    CassandraRDD
  143. def toJavaRDD(): JavaRDD[R]

    Definition Classes
    RDD
  144. def toLocalIterator: Iterator[R]

    Definition Classes
    RDD
  145. def toString(): String

    Definition Classes
    RDD → AnyRef → Any
  146. def top(num: Int)(implicit ord: Ordering[R]): Array[R]

    Definition Classes
    RDD
  147. def union(other: RDD[R]): RDD[R]

    Definition Classes
    RDD
  148. def unpersist(blocking: Boolean): CassandraStreamingRDD.this.type

    Definition Classes
    RDD
  149. lazy val verify: Unit

    Definition Classes
    CassandraRDD
  150. final def wait(): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  151. final def wait(arg0: Long, arg1: Int): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  152. final def wait(arg0: Long): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  153. def where(cql: String, values: Any*): CassandraRDD[R]

    Adds a CQL WHERE predicate(s) to the query.

    Adds a CQL WHERE predicate(s) to the query. Useful for leveraging secondary indexes in Cassandra. Implicitly adds an ALLOW FILTERING clause to the WHERE clause, however beware that some predicates might be rejected by Cassandra, particularly in cases when they filter on an unindexed, non-clustering column.

    Definition Classes
    CassandraRDD
  154. def withAscOrder: CassandraRDD[R]

    Definition Classes
    CassandraRDD
  155. def withConnector(connector: CassandraConnector): CassandraRDD[R]

    Returns a copy of this Cassandra RDD with specified connector

    Returns a copy of this Cassandra RDD with specified connector

    Definition Classes
    CassandraRDD
  156. def withDescOrder: CassandraRDD[R]

    Definition Classes
    CassandraRDD
  157. def withReadConf(readConf: ReadConf): CassandraRDD[R]

    Allows to set custom read configuration, e.

    Allows to set custom read configuration, e.g. consistency level or fetch size.

    Definition Classes
    CassandraRDD
  158. def zip[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(R, U)]

    Definition Classes
    RDD
  159. def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D])(f: (Iterator[R], Iterator[B], Iterator[C], Iterator[D]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  160. def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D], preservesPartitioning: Boolean)(f: (Iterator[R], Iterator[B], Iterator[C], Iterator[D]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  161. def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C])(f: (Iterator[R], Iterator[B], Iterator[C]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  162. def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C], preservesPartitioning: Boolean)(f: (Iterator[R], Iterator[B], Iterator[C]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  163. def zipPartitions[B, V](rdd2: RDD[B])(f: (Iterator[R], Iterator[B]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  164. def zipPartitions[B, V](rdd2: RDD[B], preservesPartitioning: Boolean)(f: (Iterator[R], Iterator[B]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  165. def zipWithIndex(): RDD[(R, Long)]

    Definition Classes
    RDD
  166. def zipWithUniqueId(): RDD[(R, Long)]

    Definition Classes
    RDD
  167. def [B](y: B): (CassandraStreamingRDD[R], B)

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to ArrowAssoc[CassandraStreamingRDD[R]] performed by method any2ArrowAssoc in scala.Predef.
    Definition Classes
    ArrowAssoc

Shadowed Implicit Value Members

  1. val self: Any

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to StringAdd performed by method any2stringadd in scala.Predef.
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraStreamingRDD: StringAdd).self
    Definition Classes
    StringAdd
  2. val self: Any

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to StringFormat performed by method any2stringfmt in scala.Predef.
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraStreamingRDD: StringFormat).self
    Definition Classes
    StringFormat
  3. val sparkContext: SparkContext

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to RDDFunctions[R] performed by method toRDDFunctions in com.datastax.spark.connector.
    Shadowing
    This implicitly inherited member is shadowed by one or more members in this class.
    To access this member you can use a type ascription:
    (cassandraStreamingRDD: RDDFunctions[R]).sparkContext
    Definition Classes
    RDDFunctionsWritableToCassandra

Deprecated Value Members

  1. def filterWith[A](constructA: (Int) ⇒ A)(p: (R, A) ⇒ Boolean): RDD[R]

    Definition Classes
    RDD
    Annotations
    @deprecated
    Deprecated

    (Since version 1.0.0) use mapPartitionsWithIndex and filter

  2. def flatMapWith[A, U](constructA: (Int) ⇒ A, preservesPartitioning: Boolean)(f: (R, A) ⇒ Seq[U])(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
    Annotations
    @deprecated
    Deprecated

    (Since version 1.0.0) use mapPartitionsWithIndex and flatMap

  3. def foreachWith[A](constructA: (Int) ⇒ A)(f: (R, A) ⇒ Unit): Unit

    Definition Classes
    RDD
    Annotations
    @deprecated
    Deprecated

    (Since version 1.0.0) use mapPartitionsWithIndex and foreach

  4. def mapPartitionsWithContext[U](f: (TaskContext, Iterator[R]) ⇒ Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
    Annotations
    @DeveloperApi() @deprecated
    Deprecated

    (Since version 1.2.0) use TaskContext.get

  5. def mapPartitionsWithSplit[U](f: (Int, Iterator[R]) ⇒ Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
    Annotations
    @deprecated
    Deprecated

    (Since version 0.7.0) use mapPartitionsWithIndex

  6. def mapWith[A, U](constructA: (Int) ⇒ A, preservesPartitioning: Boolean)(f: (R, A) ⇒ U)(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
    Annotations
    @deprecated
    Deprecated

    (Since version 1.0.0) use mapPartitionsWithIndex

  7. def toArray(): Array[R]

    Definition Classes
    RDD
    Annotations
    @deprecated
    Deprecated

    (Since version 1.0.0) use collect

  8. def x: CassandraStreamingRDD[R]

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to ArrowAssoc[CassandraStreamingRDD[R]] performed by method any2ArrowAssoc in scala.Predef.
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraStreamingRDD: ArrowAssoc[CassandraStreamingRDD[R]]).x
    Definition Classes
    ArrowAssoc
    Annotations
    @deprecated
    Deprecated

    (Since version 2.10.0) Use leftOfArrow instead

  9. def x: CassandraStreamingRDD[R]

    Implicit information
    This member is added by an implicit conversion from CassandraStreamingRDD[R] to Ensuring[CassandraStreamingRDD[R]] performed by method any2Ensuring in scala.Predef. This conversion will take place only if R is (Nothing, Nothing) (R =:= (Nothing, Nothing)).
    Shadowing
    This implicitly inherited member is ambiguous. One or more implicitly inherited members have similar signatures, so calling this member may produce an ambiguous implicit conversion compiler error.
    To access this member you can use a type ascription:
    (cassandraStreamingRDD: Ensuring[CassandraStreamingRDD[R]]).x
    Definition Classes
    Ensuring
    Annotations
    @deprecated
    Deprecated

    (Since version 2.10.0) Use resultOfEnsuring instead

Inherited from CassandraRDD[R]

Inherited from RDD[R]

Inherited from Logging

Inherited from Serializable

Inherited from Serializable

Inherited from AnyRef

Inherited from Any

Inherited by implicit conversion toPairRDDFunctions from CassandraStreamingRDD[R] to PairRDDFunctions[K, V]

Inherited by implicit conversion toRDDFunctions from CassandraStreamingRDD[R] to RDDFunctions[R]

Inherited by implicit conversion any2stringadd from CassandraStreamingRDD[R] to StringAdd

Inherited by implicit conversion any2stringfmt from CassandraStreamingRDD[R] to StringFormat

Inherited by implicit conversion any2ArrowAssoc from CassandraStreamingRDD[R] to ArrowAssoc[CassandraStreamingRDD[R]]

Inherited by implicit conversion any2Ensuring from CassandraStreamingRDD[R] to Ensuring[CassandraStreamingRDD[R]]

Ungrouped