Creates ReadConf based on an externally provided map of properties to override those of SparkCon
Creates ReadConf based on an externally provided map of properties to override those of SparkCon
SparkConf of Spark context to be taken as defaults
externally provided map of properties
Creates ReadConf based on properties provided to Spark Conf
Creates ReadConf based on properties provided to Spark Conf
SparkConf of Spark context with Riak-related properties
The docs recommend to have your number of partitions set to 3 or 4 times the number of CPUs in your cluster so that the work gets distributed more evenly among the CPUs.
The docs recommend to have your number of partitions set to 3 or 4 times the number of CPUs in your cluster so that the work gets distributed more evenly among the CPUs. Meaning, if you only have 1 partition per core in the cluster you will have to wait for the one longest running task to complete but if you had broken that down further the workload would be more evenly balanced with fast and slow running tasks evening out.
Since there is no enough information about available Spark resources such as real number of cores, 3x multiplier will be used.