The configuration files definition
The defined execution plans
The defined execution plans
an implicit spark session
A JobName/ExecutionPlan Map
Saves the resulting dataFrame to disk
Saves the resulting dataFrame to disk
The resulting DataFrame
The output database name
The output table name (job name)
An OptionalJobParameters object containing any custom argument/application files we defined through our application.
Extracts data from a provided sources configuration
Extracts data from a provided sources configuration
The Job name
The Job input sources configuration
The different input sources configuration
The environment in which we want to extract the input sources from
An implicit spark session
A Map of sourceName/SourcePipeline containing the extracted sources.
Loads a set of parameters needed for the ETL Operation
Loads a set of parameters needed for the ETL Operation
through : config files loading, argument parsing, execution parameters creation, etc...
The list of arguments to parse
An ExecutionData object containing all the required parameters
The main method containing the logic for running our ETL job
The main method containing the logic for running our ETL job
The ExecutionData object that will be used
Apply transformations to a given set of sources
Apply transformations to a given set of sources
The Job name
The extracted input sources
The execution plan to apply
An optional date column name to tie the result computation date
An implicit spark session
The resulting DataFrame
A class representing a submitted Spark application.