If true the input data file encoding is EBCDIC, otherwise it is ASCII
Specifies what code page to use for EBCDIC to ASCII/Unicode conversions
An optional custom code page conversion class provided by a user
A charset for ASCII data
If true UTF-16 strings are considered big-endian.
A format of floating-point numbers
If true, OCCURS DEPENDING ON data size will depend on the number of elements
A name of a field that contains record length. Optional. If not set the copybook record length will be used.
Does input files have 4 byte record length headers
Is RDW big endian? It may depend on flavor of mainframe and/or mainframe to PC transfer method
Does RDW count itself as part of record length itself
Controls a mismatch between RDW and record length
Is indexing input file before processing is requested
The number of records to include in each partition. Notice mainframe records may have variable size, inputSplitMB is the recommended option
A partition size to target. In certain circumstances this size may not be exactly that, but the library will do the best effort to target that size
Default HDFS block size for the HDFS filesystem used. This value is used as the default split size if inputSplitSizeMB is not specified
An offset to the start of the record in each binary data block.
An offset from the end of the record to the end of the binary data block.
A number of bytes to skip at the beginning of each file
A number of bytes to skip at the end of each file
If true, a record id field will be prepended to each record.
Specifies a policy to transform the input schema. The default policy is to keep the schema exactly as it is in the copybook.
Specifies if and how strings should be trimmed when parsed.
Parameters specific to reading multisegment files
A comment truncation policy
If true the parser will drop all FILLER fields, even GROUP FILLERS that have non-FILLER nested fields
A list of non-terminals (GROUPS) to combine and parse as primitive fields
If true, additional debugging fields will be added
A parser used to parse data field record headers
An optional additional option string passed to a custom record header parser
A column name to add to the dataframe. The column will contain input file name for each record similar to 'input_file_name()' function
A charset for ASCII data
A comment truncation policy
If true the parser will drop all FILLER fields, even GROUP FILLERS that have non-FILLER nested fields
Specifies what code page to use for EBCDIC to ASCII/Unicode conversions
An optional custom code page conversion class provided by a user
An offset from the end of the record to the end of the binary data block.
A number of bytes to skip at the end of each file
A number of bytes to skip at the beginning of each file
A format of floating-point numbers
If true, a record id field will be prepended to each record.
Default HDFS block size for the HDFS filesystem used.
Default HDFS block size for the HDFS filesystem used. This value is used as the default split size if inputSplitSizeMB is not specified
A column name to add to the dataframe.
A column name to add to the dataframe. The column will contain input file name for each record similar to 'input_file_name()' function
The number of records to include in each partition.
The number of records to include in each partition. Notice mainframe records may have variable size, inputSplitMB is the recommended option
A partition size to target.
A partition size to target. In certain circumstances this size may not be exactly that, but the library will do the best effort to target that size
If true, additional debugging fields will be added
If true the input data file encoding is EBCDIC, otherwise it is ASCII
Is indexing input file before processing is requested
Is RDW big endian? It may depend on flavor of mainframe and/or mainframe to PC transfer method
Does RDW count itself as part of record length itself
Does input files have 4 byte record length headers
If true UTF-16 strings are considered big-endian.
A name of a field that contains record length.
A name of a field that contains record length. Optional. If not set the copybook record length will be used.
Parameters specific to reading multisegment files
A list of non-terminals (GROUPS) to combine and parse as primitive fields
Controls a mismatch between RDW and record length
A parser used to parse data field record headers
An optional additional option string passed to a custom record header parser
Specifies a policy to transform the input schema.
Specifies a policy to transform the input schema. The default policy is to keep the schema exactly as it is in the copybook.
An offset to the start of the record in each binary data block.
Specifies if and how strings should be trimmed when parsed.
If true, OCCURS DEPENDING ON data size will depend on the number of elements
These are properties for customizing mainframe binary data reader.
If true the input data file encoding is EBCDIC, otherwise it is ASCII
Specifies what code page to use for EBCDIC to ASCII/Unicode conversions
An optional custom code page conversion class provided by a user
A charset for ASCII data
If true UTF-16 strings are considered big-endian.
A format of floating-point numbers
If true, OCCURS DEPENDING ON data size will depend on the number of elements
A name of a field that contains record length. Optional. If not set the copybook record length will be used.
Does input files have 4 byte record length headers
Is RDW big endian? It may depend on flavor of mainframe and/or mainframe to PC transfer method
Does RDW count itself as part of record length itself
Controls a mismatch between RDW and record length
Is indexing input file before processing is requested
The number of records to include in each partition. Notice mainframe records may have variable size, inputSplitMB is the recommended option
A partition size to target. In certain circumstances this size may not be exactly that, but the library will do the best effort to target that size
Default HDFS block size for the HDFS filesystem used. This value is used as the default split size if inputSplitSizeMB is not specified
An offset to the start of the record in each binary data block.
An offset from the end of the record to the end of the binary data block.
A number of bytes to skip at the beginning of each file
A number of bytes to skip at the end of each file
If true, a record id field will be prepended to each record.
Specifies a policy to transform the input schema. The default policy is to keep the schema exactly as it is in the copybook.
Specifies if and how strings should be trimmed when parsed.
Parameters specific to reading multisegment files
A comment truncation policy
If true the parser will drop all FILLER fields, even GROUP FILLERS that have non-FILLER nested fields
A list of non-terminals (GROUPS) to combine and parse as primitive fields
If true, additional debugging fields will be added
A parser used to parse data field record headers
An optional additional option string passed to a custom record header parser
A column name to add to the dataframe. The column will contain input file name for each record similar to 'input_file_name()' function