package connector
Tarantool connector for Apache Spark.
Call tarantoolSpace method on the SparkContext object to create a TarantoolRDD exposing Tarantool space as a Spark RDD.
Example:
Execute the following on a Cartridge router node (the tarantool/crud module must be installed):
local crud = require('crud') crud.insert('test_space', {1, nil, 'a1', 'Don Quixote', 'Miguel de Cervantes', 1605}) crud.insert('test_space', {2, nil, 'a2', 'The Great Gatsby', 'F. Scott Fitzgerald', 1925}) crud.insert('test_space', {3, nil, 'a3', 'War and Peace', 'Leo Tolstoy', 1869})
Write the following in your Java client code:
import io.tarantool.spark.connector._ val sparkMasterHost = "127.0.0.1" val tarantoolRouterAddress = "127.0.0.1:3301" val space = "test_space" // Populate the Spark config with the address of a Cartridge router node and credentials: val conf = new SparkConf(true) conf.set ("tarantool.username", "admin") conf.set ("tarantool.password", "testapp-cluster-cookie") conf.set ("tarantool.hosts", tarantoolRouterAddress) // Connect to the Spark cluster: val sc = new SparkContext("spark://" + sparkMasterHost + ":7077", "example", conf) // Read the space and print its contents: val rdd = sc.tarantoolSpace(space) rdd.toArray().foreach(println) sc.stop()
- Alphabetic
- By Inheritance
- connector
- AnyRef
- Any
- Hide All
- Show All
- Public
- Protected
Package Members
Type Members
- trait Logging extends AnyRef
Utility trait for classes that want to log data.
Utility trait for classes that want to log data. Creates a SLF4J logger for the class and allows logging messages at different levels using methods that only evaluate parameters lazily if the log level is enabled.
This is a copy of what Spark Previously held in org.apache.spark.Logging. That class is now private so we will expose similar functionality here.
- class SparkContextFunctions extends Serializable
Spark API for Tarantool.
Spark API for Tarantool. Provides Tarantool-specific methods on SparkContext
- class SparkContextJavaFunctions extends AnyRef
Java API for bridging
SparkContextFunctions
functionality into Java code - final class TarantoolSpark extends AnyRef
Public facade for using the Tarantool Spark API in Java.
Public facade for using the Tarantool Spark API in Java.
Provides static factory methods as entrypoints for building RDDs and other Spark API entities.
- trait TarantoolSparkException extends TarantoolException
Generic type for all module exceptions
Value Members
- implicit def toSparkContextFunctions(sc: SparkContext): SparkContextFunctions
- object TarantoolSparkException extends Serializable