package datasources

Linear Supertypes
AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. datasources
  2. AnyRef
  3. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Type Members

  1. abstract class AvroException extends Exception
    Annotations
    @Private()
  2. case class Bound(b: Array[Byte], inc: Boolean) extends Product with Serializable

    The Bound represent the boudary for the scan

    The Bound represent the boudary for the scan

    b

    The byte array of the bound

    inc

    inclusive or not.

    Annotations
    @Private()
  3. case class BoundRange(low: Array[Byte], upper: Array[Byte]) extends Product with Serializable

    The ranges for the data type whose size is known.

    The ranges for the data type whose size is known. Whether the bound is inclusive or exclusive is undefind, and upper to the caller to decide.

    Annotations
    @LimitedPrivate() @Evolving()
  4. case class BoundRanges(less: Array[BoundRange], greater: Array[BoundRange], value: Array[Byte]) extends Product with Serializable

    The class identifies the ranges for a java primitive type.

    The class identifies the ranges for a java primitive type. The caller needs to decide the bound is either inclusive or exclusive on its own. information

    Annotations
    @LimitedPrivate() @Evolving()
  5. trait BytesEncoder extends AnyRef

    The trait to support plugin architecture for different encoder/decoder.

    The trait to support plugin architecture for different encoder/decoder. encode is used for serializing the data type to byte array and the filter is used to filter out the unnecessary records.

    Annotations
    @LimitedPrivate() @Evolving()
  6. class DoubleSerDes extends SerDes
  7. case class GetResource(tbr: TableResource, rs: Array[Result]) extends Resource with Product with Serializable
    Annotations
    @Private()
  8. class HBaseTableScanRDD extends RDD[Result] with Logging
    Annotations
    @Private()
  9. type HBaseType = Array[Byte]
  10. class NaiveEncoder extends BytesEncoder with Logging

    This is the naive non-order preserving encoder/decoder.

    This is the naive non-order preserving encoder/decoder. Due to the inconsistency of the order between java primitive types and their bytearray. The data type has to be passed in so that the filter can work correctly, which is done by wrapping the type into the first byte of the serialized array.

    Annotations
    @Private()
  11. case class RDDResources(set: HashSet[Resource]) extends Product with Serializable
  12. case class Range(lower: Option[Bound], upper: Option[Bound]) extends Product with Serializable
    Annotations
    @Private()
  13. trait ReferencedResource extends AnyRef
    Annotations
    @Private()
  14. case class RegionResource(relation: HBaseRelation) extends ReferencedResource with Product with Serializable
    Annotations
    @Private()
  15. trait Resource extends AnyRef
    Annotations
    @Private()
  16. case class ScanResource(tbr: TableResource, rs: ResultScanner) extends Resource with Product with Serializable
    Annotations
    @Private()
  17. case class SchemaConversionException(msg: String) extends AvroException with Product with Serializable
    Annotations
    @Private()
  18. trait SerDes extends AnyRef
  19. class SerializableConfiguration extends Serializable
    Annotations
    @Private()
  20. case class SerializedFilter(b: Option[Array[Byte]]) extends Product with Serializable
  21. case class TableResource(relation: HBaseRelation) extends ReferencedResource with Product with Serializable
    Annotations
    @Private()

Value Members

  1. val ByteMax: Byte
  2. val ByteMin: Byte
  3. def bytesMax: Null
  4. def bytesMin: Array[Byte]
  5. val ord: Ordering[HBaseType]
  6. implicit val order: Ordering[HBaseType]
  7. object AvroSerdes
    Annotations
    @Private()
  8. object HBaseResources
    Annotations
    @Private()
  9. object HBaseSparkConf

    This is the hbase configuration.

    This is the hbase configuration. User can either set them in SparkConf, which will take effect globally, or configure it per table, which will overwrite the value set in SparkConf. If not set, the default value will take effect.

    Annotations
    @Public()
  10. object HBaseTableScanRDD extends Serializable
  11. object JavaBytesEncoder extends Enumeration with Logging
  12. object Points
    Annotations
    @Private()
  13. object Range extends Serializable
    Annotations
    @Private()
  14. object Ranges
    Annotations
    @Private()
  15. object SchemaConverters

    * On top level, the converters provide three high level interface.

    * On top level, the converters provide three high level interface. 1. toSqlType: This function takes an avro schema and returns a sql schema. 2. createConverterToSQL: Returns a function that is used to convert avro types to their corresponding sparkSQL representations. 3. convertTypeToAvro: This function constructs converter function for a given sparkSQL datatype. This is used in writing Avro records out to disk

    Annotations
    @Private()
  16. object SerializedFilter extends Serializable

Inherited from AnyRef

Inherited from Any

Ungrouped