Does input files have 4 byte record length headers
Block descriptor word (if specified), for FB and VB record formats
Is RDW big endian? It may depend on flavor of mainframe and/or mainframe to PC transfer method
Does RDW count itself as part of record length itself
Controls a mismatch between RDW and record length
An optional custom record header parser for non-standard RDWs
An optional custom raw record parser class non-standard record types
An optional additional option string passed to a custom record header parser
An optional additional option string passed to a custom record extractor
A field that stores record length
A mapping between field value and record size.
A number of bytes to skip at the beginning of each file
A number of bytes to skip at the end of each file
Generate a sequential record number for each record to be able to retain the order of the original data
Is indexing input file before processing is requested
The number of records to include in each partition. Notice mainframe records may have variable size, inputSplitMB is the recommended option
A partition size to target. In certain circumstances this size may not be exactly that, but the library will do the best effort to target that size
Tries to improve locality by extracting preferred locations for variable-length records
Optimizes cluster usage in case of optimization for locality in the presence of new nodes (nodes that do not contain any blocks of the files being processed)
A column name to add to the dataframe. The column will contain input file name for each record similar to 'input_file_name()' function
Block descriptor word (if specified), for FB and VB record formats
A number of bytes to skip at the end of each file
A number of bytes to skip at the beginning of each file
Generate a sequential record number for each record to be able to retain the order of the original data
Tries to improve locality by extracting preferred locations for variable-length records
A column name to add to the dataframe.
A column name to add to the dataframe. The column will contain input file name for each record similar to 'input_file_name()' function
The number of records to include in each partition.
The number of records to include in each partition. Notice mainframe records may have variable size, inputSplitMB is the recommended option
A partition size to target.
A partition size to target. In certain circumstances this size may not be exactly that, but the library will do the best effort to target that size
Is RDW big endian? It may depend on flavor of mainframe and/or mainframe to PC transfer method
Does RDW count itself as part of record length itself
Does input files have 4 byte record length headers
Is indexing input file before processing is requested
Optimizes cluster usage in case of optimization for locality in the presence of new nodes (nodes that do not contain any blocks of the files being processed)
Controls a mismatch between RDW and record length
An optional additional option string passed to a custom record extractor
An optional custom raw record parser class non-standard record types
An optional custom record header parser for non-standard RDWs
A field that stores record length
A mapping between field value and record size.
An optional additional option string passed to a custom record header parser
This class holds the parameters currently used for parsing variable-length records.
Does input files have 4 byte record length headers
Block descriptor word (if specified), for FB and VB record formats
Is RDW big endian? It may depend on flavor of mainframe and/or mainframe to PC transfer method
Does RDW count itself as part of record length itself
Controls a mismatch between RDW and record length
An optional custom record header parser for non-standard RDWs
An optional custom raw record parser class non-standard record types
An optional additional option string passed to a custom record header parser
An optional additional option string passed to a custom record extractor
A field that stores record length
A mapping between field value and record size.
A number of bytes to skip at the beginning of each file
A number of bytes to skip at the end of each file
Generate a sequential record number for each record to be able to retain the order of the original data
Is indexing input file before processing is requested
The number of records to include in each partition. Notice mainframe records may have variable size, inputSplitMB is the recommended option
A partition size to target. In certain circumstances this size may not be exactly that, but the library will do the best effort to target that size
Tries to improve locality by extracting preferred locations for variable-length records
Optimizes cluster usage in case of optimization for locality in the presence of new nodes (nodes that do not contain any blocks of the files being processed)
A column name to add to the dataframe. The column will contain input file name for each record similar to 'input_file_name()' function