Packages

  • package root
    Definition Classes
    root
  • package io
    Definition Classes
    root
  • package tarantool
    Definition Classes
    io
  • package spark
    Definition Classes
    tarantool
  • package connector

    Tarantool connector for Apache Spark.

    Tarantool connector for Apache Spark.

    Call tarantoolSpace method on the SparkContext object to create a TarantoolRDD exposing Tarantool space as a Spark RDD.

    Example:

    Execute the following on a Cartridge router node (the tarantool/crud module must be installed):

    local crud = require('crud')
    
    crud.insert('test_space', {1, nil, 'a1', 'Don Quixote', 'Miguel de Cervantes', 1605})
    crud.insert('test_space', {2, nil, 'a2', 'The Great Gatsby', 'F. Scott Fitzgerald', 1925})
    crud.insert('test_space', {3, nil, 'a3', 'War and Peace', 'Leo Tolstoy', 1869})

    Write the following in your Java client code:

    import io.tarantool.spark.connector._
    
    val sparkMasterHost = "127.0.0.1"
    val tarantoolRouterAddress = "127.0.0.1:3301"
    val space = "test_space"
    
    // Populate the Spark config with the address of a Cartridge router node and credentials:
    val conf = new SparkConf(true)
    conf.set ("tarantool.username", "admin")
    conf.set ("tarantool.password", "testapp-cluster-cookie")
    conf.set ("tarantool.hosts", tarantoolRouterAddress)
    
    // Connect to the Spark cluster:
    val sc = new SparkContext("spark://" + sparkMasterHost + ":7077", "example", conf)
    
    // Read the space and print its contents:
    val rdd = sc.tarantoolSpace(space)
    rdd.toArray().foreach(println)
    
    sc.stop()
    Definition Classes
    spark
  • package config
    Definition Classes
    connector
  • package connection
    Definition Classes
    connector
  • package partition
    Definition Classes
    connector
  • package rdd
    Definition Classes
    connector
  • package util
    Definition Classes
    connector
  • Logging
  • SparkContextFunctions
  • SparkContextJavaFunctions
  • TarantoolSpark
  • TarantoolSparkException
c

io.tarantool.spark.connector

SparkContextFunctions

class SparkContextFunctions extends Serializable

Spark API for Tarantool. Provides Tarantool-specific methods on SparkContext

Linear Supertypes
Serializable, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. SparkContextFunctions
  2. Serializable
  3. AnyRef
  4. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. Protected

Instance Constructors

  1. new SparkContextFunctions(sc: SparkContext)

Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##: Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  5. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.CloneNotSupportedException]) @native()
  6. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  7. def equals(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef → Any
  8. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.Throwable])
  9. final def getClass(): Class[_ <: AnyRef]
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  10. def hashCode(): Int
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  11. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  12. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  13. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  14. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  15. val sc: SparkContext
  16. final def synchronized[T0](arg0: => T0): T0
    Definition Classes
    AnyRef
  17. def tarantoolSpace[R](space: String, conditions: Conditions)(implicit ct: ClassTag[R], sparkContext: SparkContext = sc, readConfig: ReadConfig = ReadConfig(space).withConditions(conditions), tupleConverterFactory: TupleConverterFactory[R]): TarantoolReadRDD[R]

    Load data from Tarantool space as TarantoolRDD, filtering it with conditions.

    Load data from Tarantool space as TarantoolRDD, filtering it with conditions.

    This method is made available on SparkContext by importing io.tarantool.spark._

    Example:

    local crud = require('crud')
    
    crud.insert('test_space', {1, nil, 'a1', 'Don Quixote', 'Miguel de Cervantes', 1605})
    crud.insert('test_space', {2, nil, 'a2', 'The Great Gatsby', 'F. Scott Fitzgerald', 1925})
    crud.insert('test_space', {3, nil, 'a3', 'War and Peace', 'Leo Tolstoy', 1869})
    ...
    
    val rdd = sc.tarantoolSpace("test_space", Conditions.indexGreaterThan("id", Collections.singletonList(1)));
    rdd.first().getInteger("id"); // 1
    rdd.first().getString("author"); // "Miguel de Cervantes"
    

    space

    space name

    conditions

    filtering conditions

    returns

    TarantoolReadRDD with tuples from the space

  18. def toString(): String
    Definition Classes
    AnyRef → Any
  19. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException])
  20. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException])
  21. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException]) @native()

Inherited from Serializable

Inherited from AnyRef

Inherited from Any

Ungrouped