Class

com.ebiznext.comet.job.metrics

MetricsJob

Related Doc: package metrics

Permalink

class MetricsJob extends SparkJob

Linear Supertypes
SparkJob, StrictLogging, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. MetricsJob
  2. SparkJob
  3. StrictLogging
  4. AnyRef
  5. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new MetricsJob(domain: Domain, schema: Schema, stage: Stage, storageHandler: StorageHandler, schemaHandler: SchemaHandler)(implicit settings: Settings)

    Permalink

    domain

    : Domain name

    schema

    : Schema

    stage

    : stage

    storageHandler

    : Storage Handler

Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  5. def clone(): AnyRef

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  6. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  7. def equals(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  8. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  9. def generateFullMetric(dataMetric: DataFrame, listAttibutes: List[String], colName: List[Column]): DataFrame

    Permalink

    Function that retrieves full metrics dataframe with both set discrete and continuous metrics

    Function that retrieves full metrics dataframe with both set discrete and continuous metrics

    dataMetric

    : dataframe obtain from computeDiscretMetric( ) or computeContinuiousMetric( )

    listAttibutes

    : list of all variables

    colName

    : list of column

    returns

    Dataframe : that contain the full metrics with all variables and all metrics

  10. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
  11. def hashCode(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  12. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  13. def lockPath(path: String): Path

    Permalink
  14. val logger: Logger

    Permalink
    Attributes
    protected
    Definition Classes
    StrictLogging
  15. def metricsPath(path: String): Path

    Permalink

    Function to build the metrics save path

    Function to build the metrics save path

    path

    : path where metrics are stored

    returns

    : path where the metrics for the specified schema are stored

  16. def name: String

    Permalink
    Definition Classes
    MetricsJobSparkJob
  17. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  18. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
  19. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
  20. def partitionDataset(dataset: DataFrame, partition: List[String]): DataFrame

    Permalink
    Definition Classes
    SparkJob
  21. def partitionedDatasetWriter(dataset: DataFrame, partition: List[String]): DataFrameWriter[Row]

    Permalink

    Partition a dataset using dataset columns.

    Partition a dataset using dataset columns. To partition the dataset using the igestion time, use the reserved column names :

    • comet_year
    • comet_month
    • comet_day
    • comet_hour
    • comet_minute These columsn are renamed to "year", "month", "day", "hour", "minute" in the dataset and their values is set to the current date/time.
    dataset

    : Input dataset

    partition

    : list of columns to use for partitioning.

    returns

    The Spark session used to run this job

    Definition Classes
    SparkJob
  22. def run(dataUse: DataFrame, timestamp: Timestamp): Try[SparkSession]

    Permalink
  23. def run(): Try[SparkSession]

    Permalink

    Just to force any spark job to implement its entry point using within the "run" method

    Just to force any spark job to implement its entry point using within the "run" method

    returns

    : Spark Session used for the job

    Definition Classes
    MetricsJobSparkJob
  24. def save(dataToSave: DataFrame, path: Path): Unit

    Permalink

    Saves a dataset.

    Saves a dataset. If the path is empty (the first time we call metrics on the schema) then we can write.

    If there's already parquet files stored in it, then create a temporary directory to compute on, and flush the path to move updated metrics in it

    dataToSave

    : dataset to be saved

    path

    : Path to save the file at

  25. lazy val session: SparkSession

    Permalink
    Definition Classes
    SparkJob
  26. implicit val settings: Settings

    Permalink
    Definition Classes
    MetricsJobSparkJob
  27. lazy val sparkEnv: SparkEnv

    Permalink
    Definition Classes
    SparkJob
  28. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  29. def toString(): String

    Permalink
    Definition Classes
    AnyRef → Any
  30. def unionDisContMetric(discreteDataset: Option[DataFrame], continuousDataset: Option[DataFrame], domain: Domain, schema: Schema, count: Long, ingestionTime: Timestamp, stageState: Stage): Option[DataFrame]

    Permalink

    Function Function that unifies discrete and continuous metrics dataframe, then write save the result to parquet

    Function Function that unifies discrete and continuous metrics dataframe, then write save the result to parquet

    discreteDataset

    : dataframe that contains all the discrete metrics

    continuousDataset

    : dataframe that contains all the continuous metrics

    domain

    : name of the domain

    schema

    : schema of the initial data

    ingestionTime

    : time which correspond to the ingestion

    stageState

    : stage (unit / global)

  31. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  32. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  33. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )

Inherited from SparkJob

Inherited from StrictLogging

Inherited from AnyRef

Inherited from Any

Ungrouped