Class/Object

com.ebiznext.comet.job.index.bqload

BigQuerySparkJob

Related Docs: object BigQuerySparkJob | package bqload

Permalink

class BigQuerySparkJob extends SparkJob with BigQueryJobBase

Linear Supertypes
BigQueryJobBase, SparkJob, JobBase, StrictLogging, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. BigQuerySparkJob
  2. BigQueryJobBase
  3. SparkJob
  4. JobBase
  5. StrictLogging
  6. AnyRef
  7. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new BigQuerySparkJob(cliConfig: BigQueryLoadConfig, maybeSchema: Option[Schema] = None)(implicit settings: Settings)

    Permalink

Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  4. def analyze(fullTableName: String): Any

    Permalink
    Attributes
    protected
    Definition Classes
    SparkJob
  5. def appendToFile(storageHandler: StorageHandler, dataToSave: DataFrame, path: Path): Unit

    Permalink

    Saves a dataset.

    Saves a dataset. If the path is empty (the first time we call metrics on the schema) then we can write.

    If there's already parquet files stored in it, then create a temporary directory to compute on, and flush the path to move updated metrics in it

    dataToSave

    : dataset to be saved

    path

    : Path to save the file at

    Attributes
    protected
    Definition Classes
    SparkJob
  6. def applyTableIamPolicy(tableId: TableId, rls: RowLevelSecurity): Policy

    Permalink

    To set access control on a table or view, we can use Identity and Access Management (IAM) policy After you create a table or view, you can set its policy with a set-iam-policy call For each call, we compare if the existing policy is equal to the defined one (in the Yaml file) If it's the case, we do nothing, otherwise we update the Table policy

    To set access control on a table or view, we can use Identity and Access Management (IAM) policy After you create a table or view, you can set its policy with a set-iam-policy call For each call, we compare if the existing policy is equal to the defined one (in the Yaml file) If it's the case, we do nothing, otherwise we update the Table policy

    Definition Classes
    BigQueryJobBase
  7. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  8. val bigquery: BigQuery

    Permalink
    Definition Classes
    BigQueryJobBase
  9. val bqTable: String

    Permalink
    Definition Classes
    BigQueryJobBase
  10. val bucket: String

    Permalink
  11. val cliConfig: BigQueryLoadConfig

    Permalink
    Definition Classes
    BigQuerySparkJobBigQueryJobBase
  12. def clone(): AnyRef

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  13. val conf: Configuration

    Permalink
  14. def createViews(views: Views, sqlParameters: Map[String, String]): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    SparkJob
  15. val datasetId: DatasetId

    Permalink
    Definition Classes
    BigQueryJobBase
  16. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  17. def equals(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  18. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  19. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
  20. def getOrCreateDataset(): Dataset

    Permalink
    Definition Classes
    BigQueryJobBase
  21. def getOrCreateTable(dataFrame: Option[DataFrame], maybeSchema: Option[Schema]): (Table, StandardTableDefinition)

    Permalink
  22. def hashCode(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  23. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  24. val logger: Logger

    Permalink
    Attributes
    protected
    Definition Classes
    StrictLogging
  25. def name: String

    Permalink
    Definition Classes
    BigQuerySparkJobJobBase
  26. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  27. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
  28. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
  29. def partitionDataset(dataset: DataFrame, partition: List[String]): DataFrame

    Permalink
    Attributes
    protected
    Definition Classes
    SparkJob
  30. def partitionedDatasetWriter(dataset: DataFrame, partition: List[String]): DataFrameWriter[Row]

    Permalink

    Partition a dataset using dataset columns.

    Partition a dataset using dataset columns. To partition the dataset using the ingestion time, use the reserved column names :

    • comet_date
    • comet_year
    • comet_month
    • comet_day
    • comet_hour
    • comet_minute These columns are renamed to "date", "year", "month", "day", "hour", "minute" in the dataset and their values is set to the current date/time.
    dataset

    : Input dataset

    partition

    : list of columns to use for partitioning.

    returns

    The Spark session used to run this job

    Attributes
    protected
    Definition Classes
    SparkJob
  31. def prepareConf(): Configuration

    Permalink
  32. def prepareRLS(): List[String]

    Permalink
    Definition Classes
    BigQueryJobBase
  33. val projectId: String

    Permalink
    Definition Classes
    BigQuerySparkJobBigQueryJobBase
  34. def registerUdf(udf: String): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    SparkJob
  35. def run(): Try[JobResult]

    Permalink

    Just to force any spark job to implement its entry point within the "run" method

    Just to force any spark job to implement its entry point within the "run" method

    returns

    : Spark Session used for the job

    Definition Classes
    BigQuerySparkJobJobBase
  36. def runJob(statement: String, location: String): Job

    Permalink
    Definition Classes
    BigQueryJobBase
  37. def runSparkConnector(): Try[SparkJobResult]

    Permalink
  38. lazy val session: SparkSession

    Permalink
    Definition Classes
    SparkJob
  39. implicit val settings: Settings

    Permalink
    Definition Classes
    BigQuerySparkJobJobBase
  40. lazy val sparkEnv: SparkEnv

    Permalink
    Definition Classes
    SparkJob
  41. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  42. val tableId: TableId

    Permalink
    Definition Classes
    BigQueryJobBase
  43. def timePartitioning(partitionField: String, days: Option[Int] = None, requirePartitionFilter: Boolean): Builder

    Permalink
    Definition Classes
    BigQueryJobBase
  44. def toString(): String

    Permalink
    Definition Classes
    AnyRef → Any
  45. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  46. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  47. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )

Inherited from BigQueryJobBase

Inherited from SparkJob

Inherited from JobBase

Inherited from StrictLogging

Inherited from AnyRef

Inherited from Any

Ungrouped