reduce By Key
fun <K, V> <Error class: unknown class><<Error class: unknown class><K, V>>.reduceByKey(numPartitions: Int = dstream().ssc().sc().defaultParallelism(), reduceFunc: (V, V) -> V): <Error class: unknown class><<Error class: unknown class><K, V>>
Content copied to clipboard
Return a new DStream by applying reduceByKey
to each RDD. The values for each key are merged using the supplied reduce function. Hash partitioning is used to generate the RDDs with numPartitions
partitions.
fun <K, V> <Error class: unknown class><<Error class: unknown class><K, V>>.reduceByKey(partitioner: <Error class: unknown class>, reduceFunc: (V, V) -> V): <Error class: unknown class><<Error class: unknown class><K, V>>
Content copied to clipboard
Return a new DStream by applying reduceByKey
to each RDD. The values for each key are merged using the supplied reduce function. org.apache.spark.Partitioner is used to control the partitioning of each RDD.