Package

com.ebiznext.comet.job

ingest

Permalink

package ingest

Visibility
  1. Public
  2. All

Type Members

  1. case class AuditLog(jobid: String, paths: String, domain: String, schema: String, success: Boolean, count: Long, countAccepted: Long, countRejected: Long, timestamp: Timestamp, duration: Long, message: String, step: String) extends Product with Serializable

    Permalink
  2. case class ContinuousMetricRecord(domain: String, schema: String, attribute: String, min: Option[Double], max: Option[Double], mean: Option[Double], missingValues: Option[Long], standardDev: Option[Double], variance: Option[Double], sum: Option[Double], skewness: Option[Double], kurtosis: Option[Double], percentile25: Option[Double], median: Option[Double], percentile75: Option[Double], count: Long, cometTime: Long, cometStage: String, cometMetric: String, jobId: String) extends Product with Serializable

    Permalink
  3. case class DiscreteMetricRecord(domain: String, schema: String, attribute: String, missingValuesDiscrete: Long, countDistinct: Long, count: Long, cometTime: Long, cometStage: String, cometMetric: String, jobId: String) extends Product with Serializable

    Permalink
  4. class DsvIngestionJob extends IngestionJob

    Permalink

    Main class to ingest delimiter separated values file

  5. case class FrequencyMetricRecord(domain: String, schema: String, attribute: String, category: String, frequency: Long, count: Long, cometTime: Long, cometStage: String, jobId: String) extends Product with Serializable

    Permalink
  6. trait IngestionJob extends SparkJob

    Permalink

  7. class JsonIngestionJob extends IngestionJob

    Permalink

    Main class to complex json delimiter separated values file If your json contains only one level simple attribute aka.

    Main class to complex json delimiter separated values file If your json contains only one level simple attribute aka. kind of dsv but in json format please use SIMPLE_JSON instead. It's way faster

  8. class KafkaIngestionJob extends JsonIngestionJob

    Permalink

    Main class to ingest JSON messages from Kafka

  9. case class LoadConfig(domain: String = "", schema: String = "", paths: List[Path] = Nil, options: Map[String, String] = Map.empty) extends Product with Serializable

    Permalink

    domain

    domain name of the dataset

    schema

    schema name of the dataset

    paths

    Absolute path of the file to ingest (present in the ingesting area of the domain)

  10. class PositionIngestionJob extends DsvIngestionJob

    Permalink

    Main class to ingest delimiter separated values file

  11. case class RejectedRecord(jobid: String, timestamp: Timestamp, domain: String, schema: String, error: String, path: String) extends Product with Serializable

    Permalink
  12. class SimpleJsonIngestionJob extends DsvIngestionJob

    Permalink

    Parse a simple one level json file.

    Parse a simple one level json file. Complex types such as arrays & maps are not supported. Use JsonIngestionJob instead. This class is for simple json only that makes it way faster.

  13. sealed case class Step(value: String) extends Product with Serializable

    Permalink
  14. class XmlIngestionJob extends IngestionJob

    Permalink

    Main class to XML file If your json contains only one level simple attribute aka.

    Main class to XML file If your json contains only one level simple attribute aka. kind of dsv but in json format please use SIMPLE_JSON instead. It's way faster

  15. class XmlSimplePrivacyJob extends IngestionJob

    Permalink

    Used only to apply data masking rules (privacy) on one or more simple elements in XML data.

    Used only to apply data masking rules (privacy) on one or more simple elements in XML data. The input XML file is read as a text file. Privacy rules are applied on the resulting DataFrame and the result is saved accepted area. In the definition of the XML Schema: - schema.metadata.format should be set to TEXT_XML - schema.attributes should only contain the attributes on which privacy should be applied Comet.defaultWriteFormat should be set text in order to have an XML formatted output file Comet.privacyOnly should be set to true to save the result in one file (coalesce 1)

Value Members

  1. object ImprovedDataFrameContext

    Permalink
  2. object IngestionUtil

    Permalink
  3. object LoadConfig extends CliConfig[LoadConfig] with Serializable

    Permalink
  4. object PositionIngestionUtil

    Permalink

    The Spark task that run on each worker

  5. object SparkAuditLogWriter

    Permalink
  6. object Step extends Serializable

    Permalink
  7. object XmlSimplePrivacyJob

    Permalink

Ungrouped