Interface to define custom logic for DataFrame transformations
Interface to define a custom Spark-DataFrame transformation (1:1)
Configuration of a custom Spark-DataFrame transformation between one input and one output (1:1)
Interface to define a custom Spark-DataFrame transformation (n:m) Same trait as CustomDfTransformer, but multiple input and outputs supported.
Configuration of a custom Spark-DataFrame transformation between several inputs and outputs (n:m)
Configuration of a custom Spark-DataFrame transformation between several inputs and outputs (n:m)
Optional class name to load transformer code from
Optional file where scala code for transformation is loaded from
Optional scala code for transformation
Optional map of DataObjectId and corresponding SQL Code. Use tokens %{<key>} to replace with runtimeOptions in SQL code. Example: "select * from test where run = %{runId}"
Options to pass to the transformation
optional tuples of [key, spark sql expression] to be added as additional options when executing transformation. The spark sql expressions are evaluated against an instance of DefaultExpressionData.
Interface to define custom file transformation for CustomFileAction
Configuration of custom file transformation between one input and one output (1:1)
Configuration of custom file transformation between one input and one output (1:1)
Optional class name to load transformer code from
Optional file where scala code for transformation is loaded from
Optional scala code for transformation
Options to pass to the transformation
Configuration of a custom Spark-DataFrame transformation between one input and one output (1:1)
Note about Python transformation: Environment with Python and PySpark needed. PySpark session is initialize and available under variables
sc
,session
,sqlContext
. Input DataFrame is available asinputDf
. Output DataFrame must be set withsetOutputDf(df)
.Optional class name to load transformer code from
Optional file where scala code for transformation is loaded from
Optional scala code for transformation
Optional SQL code for transformation. Use tokens %{<key>} to replace with runtimeOptions in SQL code. Example: "select * from test where run = %{runId}"
Optional pythonFile to use for python transformation
Optional pythonCode to user for python transformation
Options to pass to the transformation
optional tuples of [key, spark sql expression] to be added as additional options when executing transformation. The spark sql expressions are evaluated against an instance of DefaultExpressionData.