@Generated(value="com.amazonaws:aws-java-sdk-code-generator") public class TransformInput extends Object implements Serializable, Cloneable, StructuredPojo
Describes the input source of a transform job and the way the transform job consumes it.
| Constructor and Description | 
|---|
| TransformInput() | 
| Modifier and Type | Method and Description | 
|---|---|
| TransformInput | clone() | 
| boolean | equals(Object obj) | 
| String | getCompressionType()
 If your transform data is compressed, specify the compression type. | 
| String | getContentType()
 The multipurpose internet mail extension (MIME) type of the data. | 
| TransformDataSource | getDataSource()
 Describes the location of the channel data, which is, the S3 location of the input data that the model can
 consume. | 
| String | getSplitType()
 The method to use to split the transform job's data files into smaller batches. | 
| int | hashCode() | 
| void | marshall(ProtocolMarshaller protocolMarshaller)Marshalls this structured data using the given  ProtocolMarshaller. | 
| void | setCompressionType(String compressionType)
 If your transform data is compressed, specify the compression type. | 
| void | setContentType(String contentType)
 The multipurpose internet mail extension (MIME) type of the data. | 
| void | setDataSource(TransformDataSource dataSource)
 Describes the location of the channel data, which is, the S3 location of the input data that the model can
 consume. | 
| void | setSplitType(String splitType)
 The method to use to split the transform job's data files into smaller batches. | 
| String | toString()Returns a string representation of this object. | 
| TransformInput | withCompressionType(CompressionType compressionType)
 If your transform data is compressed, specify the compression type. | 
| TransformInput | withCompressionType(String compressionType)
 If your transform data is compressed, specify the compression type. | 
| TransformInput | withContentType(String contentType)
 The multipurpose internet mail extension (MIME) type of the data. | 
| TransformInput | withDataSource(TransformDataSource dataSource)
 Describes the location of the channel data, which is, the S3 location of the input data that the model can
 consume. | 
| TransformInput | withSplitType(SplitType splitType)
 The method to use to split the transform job's data files into smaller batches. | 
| TransformInput | withSplitType(String splitType)
 The method to use to split the transform job's data files into smaller batches. | 
public void setDataSource(TransformDataSource dataSource)
Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.
dataSource - Describes the location of the channel data, which is, the S3 location of the input data that the model can
        consume.public TransformDataSource getDataSource()
Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.
public TransformInput withDataSource(TransformDataSource dataSource)
Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.
dataSource - Describes the location of the channel data, which is, the S3 location of the input data that the model can
        consume.public void setContentType(String contentType)
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
contentType - The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with
        each http call to transfer data to the transform job.public String getContentType()
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
public TransformInput withContentType(String contentType)
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
contentType - The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with
        each http call to transfer data to the transform job.public void setCompressionType(String compressionType)
 If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses
 the data for the transform job accordingly. The default value is None.
 
compressionType - If your transform data is compressed, specify the compression type. Amazon SageMaker automatically
        decompresses the data for the transform job accordingly. The default value is None.CompressionTypepublic String getCompressionType()
 If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses
 the data for the transform job accordingly. The default value is None.
 
None.CompressionTypepublic TransformInput withCompressionType(String compressionType)
 If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses
 the data for the transform job accordingly. The default value is None.
 
compressionType - If your transform data is compressed, specify the compression type. Amazon SageMaker automatically
        decompresses the data for the transform job accordingly. The default value is None.CompressionTypepublic TransformInput withCompressionType(CompressionType compressionType)
 If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses
 the data for the transform job accordingly. The default value is None.
 
compressionType - If your transform data is compressed, specify the compression type. Amazon SageMaker automatically
        decompresses the data for the transform job accordingly. The default value is None.CompressionTypepublic void setSplitType(String splitType)
 The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
 total size of each object is too large to fit in a single request. You can also use data splitting to improve
 performance by processing multiple concurrent mini-batches. The default value for SplitType is
 None, which indicates that input data files are not split, and request payloads contain the entire
 contents of an input object. Set the value of this parameter to Line to split records on a newline
 character boundary. SplitType also supports a number of record-oriented binary data formats.
 Currently, the supported record formats are:
 
RecordIO
TFRecord
 When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and
 MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord,
 Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
 limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual
 records in each request.
 
 Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
 applied to a binary data format, padding is removed if the value of BatchStrategy is set to
 SingleRecord. Padding is not removed if the value of BatchStrategy is set to
 MultiRecord.
 
 For more information about RecordIO, see Create
 a Dataset Using RecordIO in the MXNet documentation. For more information about TFRecord, see Consuming TFRecord data in the
 TensorFlow documentation.
 
splitType - The method to use to split the transform job's data files into smaller batches. Splitting is necessary
        when the total size of each object is too large to fit in a single request. You can also use data
        splitting to improve performance by processing multiple concurrent mini-batches. The default value for
        SplitType is None, which indicates that input data files are not split, and
        request payloads contain the entire contents of an input object. Set the value of this parameter to
        Line to split records on a newline character boundary. SplitType also supports a
        number of record-oriented binary data formats. Currently, the supported record formats are:
        RecordIO
TFRecord
        When splitting is enabled, the size of a mini-batch depends on the values of the
        BatchStrategy and MaxPayloadInMB parameters. When the value of
        BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of
        records in each request, up to the MaxPayloadInMB limit. If the value of
        BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in each
        request.
        
        Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
        is applied to a binary data format, padding is removed if the value of BatchStrategy is set
        to SingleRecord. Padding is not removed if the value of BatchStrategy is set to
        MultiRecord.
        
        For more information about RecordIO, see Create a Dataset Using RecordIO in the MXNet
        documentation. For more information about TFRecord, see Consuming TFRecord data in
        the TensorFlow documentation.
        
SplitTypepublic String getSplitType()
 The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
 total size of each object is too large to fit in a single request. You can also use data splitting to improve
 performance by processing multiple concurrent mini-batches. The default value for SplitType is
 None, which indicates that input data files are not split, and request payloads contain the entire
 contents of an input object. Set the value of this parameter to Line to split records on a newline
 character boundary. SplitType also supports a number of record-oriented binary data formats.
 Currently, the supported record formats are:
 
RecordIO
TFRecord
 When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and
 MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord,
 Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
 limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual
 records in each request.
 
 Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
 applied to a binary data format, padding is removed if the value of BatchStrategy is set to
 SingleRecord. Padding is not removed if the value of BatchStrategy is set to
 MultiRecord.
 
 For more information about RecordIO, see Create
 a Dataset Using RecordIO in the MXNet documentation. For more information about TFRecord, see Consuming TFRecord data in the
 TensorFlow documentation.
 
SplitType is None, which indicates that input data files are not split, and
         request payloads contain the entire contents of an input object. Set the value of this parameter to
         Line to split records on a newline character boundary. SplitType also supports
         a number of record-oriented binary data formats. Currently, the supported record formats are:
         RecordIO
TFRecord
         When splitting is enabled, the size of a mini-batch depends on the values of the
         BatchStrategy and MaxPayloadInMB parameters. When the value of
         BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of
         records in each request, up to the MaxPayloadInMB limit. If the value of
         BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in
         each request.
         
         Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
         is applied to a binary data format, padding is removed if the value of BatchStrategy is set
         to SingleRecord. Padding is not removed if the value of BatchStrategy is set to
         MultiRecord.
         
         For more information about RecordIO, see Create a Dataset Using RecordIO in the MXNet
         documentation. For more information about TFRecord, see Consuming TFRecord data in
         the TensorFlow documentation.
         
SplitTypepublic TransformInput withSplitType(String splitType)
 The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
 total size of each object is too large to fit in a single request. You can also use data splitting to improve
 performance by processing multiple concurrent mini-batches. The default value for SplitType is
 None, which indicates that input data files are not split, and request payloads contain the entire
 contents of an input object. Set the value of this parameter to Line to split records on a newline
 character boundary. SplitType also supports a number of record-oriented binary data formats.
 Currently, the supported record formats are:
 
RecordIO
TFRecord
 When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and
 MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord,
 Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
 limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual
 records in each request.
 
 Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
 applied to a binary data format, padding is removed if the value of BatchStrategy is set to
 SingleRecord. Padding is not removed if the value of BatchStrategy is set to
 MultiRecord.
 
 For more information about RecordIO, see Create
 a Dataset Using RecordIO in the MXNet documentation. For more information about TFRecord, see Consuming TFRecord data in the
 TensorFlow documentation.
 
splitType - The method to use to split the transform job's data files into smaller batches. Splitting is necessary
        when the total size of each object is too large to fit in a single request. You can also use data
        splitting to improve performance by processing multiple concurrent mini-batches. The default value for
        SplitType is None, which indicates that input data files are not split, and
        request payloads contain the entire contents of an input object. Set the value of this parameter to
        Line to split records on a newline character boundary. SplitType also supports a
        number of record-oriented binary data formats. Currently, the supported record formats are:
        RecordIO
TFRecord
        When splitting is enabled, the size of a mini-batch depends on the values of the
        BatchStrategy and MaxPayloadInMB parameters. When the value of
        BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of
        records in each request, up to the MaxPayloadInMB limit. If the value of
        BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in each
        request.
        
        Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
        is applied to a binary data format, padding is removed if the value of BatchStrategy is set
        to SingleRecord. Padding is not removed if the value of BatchStrategy is set to
        MultiRecord.
        
        For more information about RecordIO, see Create a Dataset Using RecordIO in the MXNet
        documentation. For more information about TFRecord, see Consuming TFRecord data in
        the TensorFlow documentation.
        
SplitTypepublic TransformInput withSplitType(SplitType splitType)
 The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
 total size of each object is too large to fit in a single request. You can also use data splitting to improve
 performance by processing multiple concurrent mini-batches. The default value for SplitType is
 None, which indicates that input data files are not split, and request payloads contain the entire
 contents of an input object. Set the value of this parameter to Line to split records on a newline
 character boundary. SplitType also supports a number of record-oriented binary data formats.
 Currently, the supported record formats are:
 
RecordIO
TFRecord
 When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and
 MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord,
 Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
 limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual
 records in each request.
 
 Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
 applied to a binary data format, padding is removed if the value of BatchStrategy is set to
 SingleRecord. Padding is not removed if the value of BatchStrategy is set to
 MultiRecord.
 
 For more information about RecordIO, see Create
 a Dataset Using RecordIO in the MXNet documentation. For more information about TFRecord, see Consuming TFRecord data in the
 TensorFlow documentation.
 
splitType - The method to use to split the transform job's data files into smaller batches. Splitting is necessary
        when the total size of each object is too large to fit in a single request. You can also use data
        splitting to improve performance by processing multiple concurrent mini-batches. The default value for
        SplitType is None, which indicates that input data files are not split, and
        request payloads contain the entire contents of an input object. Set the value of this parameter to
        Line to split records on a newline character boundary. SplitType also supports a
        number of record-oriented binary data formats. Currently, the supported record formats are:
        RecordIO
TFRecord
        When splitting is enabled, the size of a mini-batch depends on the values of the
        BatchStrategy and MaxPayloadInMB parameters. When the value of
        BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of
        records in each request, up to the MaxPayloadInMB limit. If the value of
        BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in each
        request.
        
        Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
        is applied to a binary data format, padding is removed if the value of BatchStrategy is set
        to SingleRecord. Padding is not removed if the value of BatchStrategy is set to
        MultiRecord.
        
        For more information about RecordIO, see Create a Dataset Using RecordIO in the MXNet
        documentation. For more information about TFRecord, see Consuming TFRecord data in
        the TensorFlow documentation.
        
SplitTypepublic String toString()
toString in class ObjectObject.toString()public TransformInput clone()
public void marshall(ProtocolMarshaller protocolMarshaller)
StructuredPojoProtocolMarshaller.marshall in interface StructuredPojoprotocolMarshaller - Implementation of ProtocolMarshaller used to marshall this object's data.