Class

com.ebiznext.comet.job.transform

AutoTaskJob

Related Doc: package transform

Permalink

class AutoTaskJob extends SparkJob

Execute the SQL Task and store it in parquet/orc/.... If Hive support is enabled, also store it as a Hive Table. If analyze support is active, also compute basic statistics for twhe dataset.

Linear Supertypes
SparkJob, JobBase, StrictLogging, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. AutoTaskJob
  2. SparkJob
  3. JobBase
  4. StrictLogging
  5. AnyRef
  6. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new AutoTaskJob(name: String, defaultArea: Option[StorageArea], format: Option[String], coalesce: Boolean, udf: Option[String], views: Views, engine: Engine, task: AutoTaskDesc, storageHandler: StorageHandler, sqlParameters: Map[String, String], schemaHandler: SchemaHandler)(implicit settings: Settings)

    Permalink

    name

    : Job Name as defined in the YML job description file

    defaultArea

    : Where the resulting dataset is stored by default if not specified in the task

    task

    : Task to run

    sqlParameters

    : Sql Parameters to pass to SQL statements

Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  4. def analyze(fullTableName: String): Any

    Permalink
    Attributes
    protected
    Definition Classes
    SparkJob
  5. def appendToFile(storageHandler: StorageHandler, dataToSave: DataFrame, path: Path): Unit

    Permalink

    Saves a dataset.

    Saves a dataset. If the path is empty (the first time we call metrics on the schema) then we can write.

    If there's already parquet files stored in it, then create a temporary directory to compute on, and flush the path to move updated metrics in it

    dataToSave

    : dataset to be saved

    path

    : Path to save the file at

    Attributes
    protected
    Definition Classes
    SparkJob
  6. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  7. def clone(): AnyRef

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  8. val createDisposition: String

    Permalink
  9. def createViews(views: Views, sqlParameters: Map[String, String]): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    SparkJob
  10. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  11. def equals(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  12. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  13. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
  14. def hashCode(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  15. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  16. val logger: Logger

    Permalink
    Attributes
    protected
    Definition Classes
    StrictLogging
  17. val name: String

    Permalink

    : Job Name as defined in the YML job description file

    : Job Name as defined in the YML job description file

    Definition Classes
    AutoTaskJobJobBase
  18. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  19. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
  20. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
  21. def partitionDataset(dataset: DataFrame, partition: List[String]): DataFrame

    Permalink
    Attributes
    protected
    Definition Classes
    SparkJob
  22. def partitionedDatasetWriter(dataset: DataFrame, partition: List[String]): DataFrameWriter[Row]

    Permalink

    Partition a dataset using dataset columns.

    Partition a dataset using dataset columns. To partition the dataset using the ingestion time, use the reserved column names :

    • comet_date
    • comet_year
    • comet_month
    • comet_day
    • comet_hour
    • comet_minute These columns are renamed to "date", "year", "month", "day", "hour", "minute" in the dataset and their values is set to the current date/time.
    dataset

    : Input dataset

    partition

    : list of columns to use for partitioning.

    returns

    The Spark session used to run this job

    Attributes
    protected
    Definition Classes
    SparkJob
  23. def registerUdf(udf: String): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    SparkJob
  24. def run(): Try[JobResult]

    Permalink

    Just to force any job to implement its entry point using within the "run" method

    Just to force any job to implement its entry point using within the "run" method

    returns

    : Spark Dataframe for Spark Jobs None otherwise

    Definition Classes
    AutoTaskJobJobBase
  25. def runBQ(): Try[JobResult]

    Permalink
  26. def runSpark(): Try[SparkJobResult]

    Permalink
  27. def runView(viewName: String, viewDir: Option[String], viewCount: Int): Try[JobResult]

    Permalink
  28. lazy val session: SparkSession

    Permalink
    Definition Classes
    SparkJob
  29. implicit val settings: Settings

    Permalink
    Definition Classes
    AutoTaskJobJobBase
  30. lazy val sparkEnv: SparkEnv

    Permalink
    Definition Classes
    SparkJob
  31. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  32. def toString(): String

    Permalink
    Definition Classes
    AnyRef → Any
  33. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  34. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  35. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  36. val writeDisposition: String

    Permalink

Inherited from SparkJob

Inherited from JobBase

Inherited from StrictLogging

Inherited from AnyRef

Inherited from Any

Ungrouped