String containing the path to the copybook in a given file system.
Sequence containing the paths to the copybooks.
String containing the actual content of the copybook. Either this, the copybookPath, or multiCopybookPath parameter must be specified.
The list of source file paths.
The record format (F, V, VB, D)
[deprecated by recordFormat] If true the input data consists of text files where records are separated by a line ending character
If true the input data file encoding is EBCDIC, otherwise it is ASCII
Specifies what code page to use for EBCDIC to ASCII/Unicode conversions
An optional custom code page conversion class provided by a user
A charset for ASCII data
Specifies a mapping between a field name and the code page
If true UTF-16 is considered big-endian.
A format of floating-point numbers
A number of bytes to skip at the beginning of the record before parsing a record according to a copybook
A number of bytes to skip at the end of each record
Specifies the length of the record disregarding the copybook record size. Implied the file has fixed record length.
Minium record length for which the record is considered valid.
Maximum record length for which the record is considered valid.
VariableLengthParameters containing the specifications for the consumption of variable-length Cobol records.
If true, OCCURS DEPENDING ON data size will depend on the number of elements
Generate 'record_bytes' field containing raw bytes of the original record
A copybook usually has a root group struct element that acts like a rowtag in XML. This can be retained in Spark schema or can be collapsed
Specify if and how strings should be trimmed when parsed
If true, partial ASCII records can be parsed (in cases when LF character is missing for example)
Parameters for reading multisegment mainframe files
A comment truncation policy
If true, string values that contain only zero bytes (0x0) will be considered null.
Decode binary fields as HEX strings
If true the parser will drop all FILLER fields, even GROUP FILLERS that have non-FILLER nested fields
If true the parser will drop all value FILLER fields
A list of non-terminals (GROUPS) to combine and parse as primitive fields
Specifies if debugging fields need to be added and what should they contain (false, hex, raw).
If true the fixed length file reader won't check file size divisibility. Useful for debugging binary file / copybook mismatches.
Specifies the policy of metadat fields to be added to the Spark schema
If true, partial ASCII records can be parsed (in cases when LF character is missing for example)
A charset for ASCII data
A comment truncation policy
String containing the actual content of the copybook.
String containing the actual content of the copybook. Either this, the copybookPath, or multiCopybookPath parameter must be specified.
String containing the path to the copybook in a given file system.
Specifies if debugging fields need to be added and what should they contain (false, hex, raw).
If true the fixed length file reader won't check file size divisibility.
If true the fixed length file reader won't check file size divisibility. Useful for debugging binary file / copybook mismatches.
Decode binary fields as HEX strings
If true the parser will drop all FILLER fields, even GROUP FILLERS that have non-FILLER nested fields
If true the parser will drop all value FILLER fields
Specifies what code page to use for EBCDIC to ASCII/Unicode conversions
An optional custom code page conversion class provided by a user
Specifies a mapping between a field name and the code page
A format of floating-point numbers
Generate 'record_bytes' field containing raw bytes of the original record
If true, string values that contain only zero bytes (0x0) will be considered null.
If true the input data file encoding is EBCDIC, otherwise it is ASCII
[deprecated by recordFormat] If true the input data consists of text files where records are separated by a line ending character
If true UTF-16 is considered big-endian.
Maximum record length for which the record is considered valid.
Specifies the policy of metadat fields to be added to the Spark schema
Minium record length for which the record is considered valid.
Sequence containing the paths to the copybooks.
Parameters for reading multisegment mainframe files
A list of non-terminals (GROUPS) to combine and parse as primitive fields
A number of bytes to skip at the end of each record
The record format (F, V, VB, D)
Specifies the length of the record disregarding the copybook record size.
Specifies the length of the record disregarding the copybook record size. Implied the file has fixed record length.
A number of bytes to skip at the beginning of the record before parsing a record according to a copybook
A copybook usually has a root group struct element that acts like a rowtag in XML.
A copybook usually has a root group struct element that acts like a rowtag in XML. This can be retained in Spark schema or can be collapsed
The list of source file paths.
Specify if and how strings should be trimmed when parsed
VariableLengthParameters containing the specifications for the consumption of variable-length Cobol records.
If true, OCCURS DEPENDING ON data size will depend on the number of elements
This class holds parameters for the job.
String containing the path to the copybook in a given file system.
Sequence containing the paths to the copybooks.
String containing the actual content of the copybook. Either this, the copybookPath, or multiCopybookPath parameter must be specified.
The list of source file paths.
The record format (F, V, VB, D)
[deprecated by recordFormat] If true the input data consists of text files where records are separated by a line ending character
If true the input data file encoding is EBCDIC, otherwise it is ASCII
Specifies what code page to use for EBCDIC to ASCII/Unicode conversions
An optional custom code page conversion class provided by a user
A charset for ASCII data
Specifies a mapping between a field name and the code page
If true UTF-16 is considered big-endian.
A format of floating-point numbers
A number of bytes to skip at the beginning of the record before parsing a record according to a copybook
A number of bytes to skip at the end of each record
Specifies the length of the record disregarding the copybook record size. Implied the file has fixed record length.
Minium record length for which the record is considered valid.
Maximum record length for which the record is considered valid.
VariableLengthParameters containing the specifications for the consumption of variable-length Cobol records.
If true, OCCURS DEPENDING ON data size will depend on the number of elements
Generate 'record_bytes' field containing raw bytes of the original record
A copybook usually has a root group struct element that acts like a rowtag in XML. This can be retained in Spark schema or can be collapsed
Specify if and how strings should be trimmed when parsed
If true, partial ASCII records can be parsed (in cases when LF character is missing for example)
Parameters for reading multisegment mainframe files
A comment truncation policy
If true, string values that contain only zero bytes (0x0) will be considered null.
Decode binary fields as HEX strings
If true the parser will drop all FILLER fields, even GROUP FILLERS that have non-FILLER nested fields
If true the parser will drop all value FILLER fields
A list of non-terminals (GROUPS) to combine and parse as primitive fields
Specifies if debugging fields need to be added and what should they contain (false, hex, raw).
If true the fixed length file reader won't check file size divisibility. Useful for debugging binary file / copybook mismatches.
Specifies the policy of metadat fields to be added to the Spark schema