@Generated(value="com.amazonaws:aws-java-sdk-code-generator") public class TransformInput extends Object implements Serializable, Cloneable, StructuredPojo
Describes the input source of a transform job and the way the transform job consumes it.
| Constructor and Description |
|---|
TransformInput() |
| Modifier and Type | Method and Description |
|---|---|
TransformInput |
clone() |
boolean |
equals(Object obj) |
String |
getCompressionType()
Compressing data helps save on storage space.
|
String |
getContentType()
The multipurpose internet mail extension (MIME) type of the data.
|
TransformDataSource |
getDataSource()
Describes the location of the channel data, meaning the S3 location of the input data that the model can consume.
|
String |
getSplitType()
The method to use to split the transform job's data files into smaller batches.
|
int |
hashCode() |
void |
marshall(ProtocolMarshaller protocolMarshaller)
Marshalls this structured data using the given
ProtocolMarshaller. |
void |
setCompressionType(String compressionType)
Compressing data helps save on storage space.
|
void |
setContentType(String contentType)
The multipurpose internet mail extension (MIME) type of the data.
|
void |
setDataSource(TransformDataSource dataSource)
Describes the location of the channel data, meaning the S3 location of the input data that the model can consume.
|
void |
setSplitType(String splitType)
The method to use to split the transform job's data files into smaller batches.
|
String |
toString()
Returns a string representation of this object.
|
TransformInput |
withCompressionType(CompressionType compressionType)
Compressing data helps save on storage space.
|
TransformInput |
withCompressionType(String compressionType)
Compressing data helps save on storage space.
|
TransformInput |
withContentType(String contentType)
The multipurpose internet mail extension (MIME) type of the data.
|
TransformInput |
withDataSource(TransformDataSource dataSource)
Describes the location of the channel data, meaning the S3 location of the input data that the model can consume.
|
TransformInput |
withSplitType(SplitType splitType)
The method to use to split the transform job's data files into smaller batches.
|
TransformInput |
withSplitType(String splitType)
The method to use to split the transform job's data files into smaller batches.
|
public void setDataSource(TransformDataSource dataSource)
Describes the location of the channel data, meaning the S3 location of the input data that the model can consume.
dataSource - Describes the location of the channel data, meaning the S3 location of the input data that the model can
consume.public TransformDataSource getDataSource()
Describes the location of the channel data, meaning the S3 location of the input data that the model can consume.
public TransformInput withDataSource(TransformDataSource dataSource)
Describes the location of the channel data, meaning the S3 location of the input data that the model can consume.
dataSource - Describes the location of the channel data, meaning the S3 location of the input data that the model can
consume.public void setContentType(String contentType)
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
contentType - The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with
each http call to transfer data to the transform job.public String getContentType()
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
public TransformInput withContentType(String contentType)
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
contentType - The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with
each http call to transfer data to the transform job.public void setCompressionType(String compressionType)
Compressing data helps save on storage space. If your transform data is compressed, specify the compression type.
Amazon SageMaker automatically decompresses the data for the transform job accordingly. The default value is
None.
compressionType - Compressing data helps save on storage space. If your transform data is compressed, specify the
compression type. Amazon SageMaker automatically decompresses the data for the transform job accordingly.
The default value is None.CompressionTypepublic String getCompressionType()
Compressing data helps save on storage space. If your transform data is compressed, specify the compression type.
Amazon SageMaker automatically decompresses the data for the transform job accordingly. The default value is
None.
None.CompressionTypepublic TransformInput withCompressionType(String compressionType)
Compressing data helps save on storage space. If your transform data is compressed, specify the compression type.
Amazon SageMaker automatically decompresses the data for the transform job accordingly. The default value is
None.
compressionType - Compressing data helps save on storage space. If your transform data is compressed, specify the
compression type. Amazon SageMaker automatically decompresses the data for the transform job accordingly.
The default value is None.CompressionTypepublic TransformInput withCompressionType(CompressionType compressionType)
Compressing data helps save on storage space. If your transform data is compressed, specify the compression type.
Amazon SageMaker automatically decompresses the data for the transform job accordingly. The default value is
None.
compressionType - Compressing data helps save on storage space. If your transform data is compressed, specify the
compression type. Amazon SageMaker automatically decompresses the data for the transform job accordingly.
The default value is None.CompressionTypepublic void setSplitType(String splitType)
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
total size of each object is too large to fit in a single request. You can also use data splitting to improve
performance by processing multiple concurrent mini-batches. The default value for SplitType is
None, which indicates that input data files are not split, and request payloads contain the entire
contents of an input object. Set the value of this parameter to Line to split records on a newline
character boundary. SplitType also supports a number of record-oriented binary data formats.
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and
MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord,
Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual
records in each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
applied to a binary data format, padding is removed if the value of BatchStrategy is set to
SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about the RecordIO data format, see Data Format in the MXNet documentation. For more information about the TFRecord fofmat, see Consuming TFRecord data in the TensorFlow documentation.
splitType - The method to use to split the transform job's data files into smaller batches. Splitting is necessary
when the total size of each object is too large to fit in a single request. You can also use data
splitting to improve performance by processing multiple concurrent mini-batches. The default value for
SplitType is None, which indicates that input data files are not split, and
request payloads contain the entire contents of an input object. Set the value of this parameter to
Line to split records on a newline character boundary. SplitType also supports a
number of record-oriented binary data formats.
When splitting is enabled, the size of a mini-batch depends on the values of the
BatchStrategy and MaxPayloadInMB parameters. When the value of
BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of
records in each request, up to the MaxPayloadInMB limit. If the value of
BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in each
request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
is applied to a binary data format, padding is removed if the value of BatchStrategy is set
to SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about the RecordIO data format, see Data Format in the MXNet documentation. For more information about the TFRecord fofmat, see Consuming TFRecord data in the TensorFlow documentation.
SplitTypepublic String getSplitType()
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
total size of each object is too large to fit in a single request. You can also use data splitting to improve
performance by processing multiple concurrent mini-batches. The default value for SplitType is
None, which indicates that input data files are not split, and request payloads contain the entire
contents of an input object. Set the value of this parameter to Line to split records on a newline
character boundary. SplitType also supports a number of record-oriented binary data formats.
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and
MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord,
Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual
records in each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
applied to a binary data format, padding is removed if the value of BatchStrategy is set to
SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about the RecordIO data format, see Data Format in the MXNet documentation. For more information about the TFRecord fofmat, see Consuming TFRecord data in the TensorFlow documentation.
SplitType is None, which indicates that input data files are not split, and
request payloads contain the entire contents of an input object. Set the value of this parameter to
Line to split records on a newline character boundary. SplitType also supports
a number of record-oriented binary data formats.
When splitting is enabled, the size of a mini-batch depends on the values of the
BatchStrategy and MaxPayloadInMB parameters. When the value of
BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of
records in each request, up to the MaxPayloadInMB limit. If the value of
BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in
each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
is applied to a binary data format, padding is removed if the value of BatchStrategy is set
to SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about the RecordIO data format, see Data Format in the MXNet documentation. For more information about the TFRecord fofmat, see Consuming TFRecord data in the TensorFlow documentation.
SplitTypepublic TransformInput withSplitType(String splitType)
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
total size of each object is too large to fit in a single request. You can also use data splitting to improve
performance by processing multiple concurrent mini-batches. The default value for SplitType is
None, which indicates that input data files are not split, and request payloads contain the entire
contents of an input object. Set the value of this parameter to Line to split records on a newline
character boundary. SplitType also supports a number of record-oriented binary data formats.
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and
MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord,
Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual
records in each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
applied to a binary data format, padding is removed if the value of BatchStrategy is set to
SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about the RecordIO data format, see Data Format in the MXNet documentation. For more information about the TFRecord fofmat, see Consuming TFRecord data in the TensorFlow documentation.
splitType - The method to use to split the transform job's data files into smaller batches. Splitting is necessary
when the total size of each object is too large to fit in a single request. You can also use data
splitting to improve performance by processing multiple concurrent mini-batches. The default value for
SplitType is None, which indicates that input data files are not split, and
request payloads contain the entire contents of an input object. Set the value of this parameter to
Line to split records on a newline character boundary. SplitType also supports a
number of record-oriented binary data formats.
When splitting is enabled, the size of a mini-batch depends on the values of the
BatchStrategy and MaxPayloadInMB parameters. When the value of
BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of
records in each request, up to the MaxPayloadInMB limit. If the value of
BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in each
request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
is applied to a binary data format, padding is removed if the value of BatchStrategy is set
to SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about the RecordIO data format, see Data Format in the MXNet documentation. For more information about the TFRecord fofmat, see Consuming TFRecord data in the TensorFlow documentation.
SplitTypepublic TransformInput withSplitType(SplitType splitType)
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
total size of each object is too large to fit in a single request. You can also use data splitting to improve
performance by processing multiple concurrent mini-batches. The default value for SplitType is
None, which indicates that input data files are not split, and request payloads contain the entire
contents of an input object. Set the value of this parameter to Line to split records on a newline
character boundary. SplitType also supports a number of record-oriented binary data formats.
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and
MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord,
Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual
records in each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
applied to a binary data format, padding is removed if the value of BatchStrategy is set to
SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about the RecordIO data format, see Data Format in the MXNet documentation. For more information about the TFRecord fofmat, see Consuming TFRecord data in the TensorFlow documentation.
splitType - The method to use to split the transform job's data files into smaller batches. Splitting is necessary
when the total size of each object is too large to fit in a single request. You can also use data
splitting to improve performance by processing multiple concurrent mini-batches. The default value for
SplitType is None, which indicates that input data files are not split, and
request payloads contain the entire contents of an input object. Set the value of this parameter to
Line to split records on a newline character boundary. SplitType also supports a
number of record-oriented binary data formats.
When splitting is enabled, the size of a mini-batch depends on the values of the
BatchStrategy and MaxPayloadInMB parameters. When the value of
BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of
records in each request, up to the MaxPayloadInMB limit. If the value of
BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in each
request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
is applied to a binary data format, padding is removed if the value of BatchStrategy is set
to SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about the RecordIO data format, see Data Format in the MXNet documentation. For more information about the TFRecord fofmat, see Consuming TFRecord data in the TensorFlow documentation.
SplitTypepublic String toString()
toString in class ObjectObject.toString()public TransformInput clone()
public void marshall(ProtocolMarshaller protocolMarshaller)
StructuredPojoProtocolMarshaller.marshall in interface StructuredPojoprotocolMarshaller - Implementation of ProtocolMarshaller used to marshall this object's data.Copyright © 2013 Amazon Web Services, Inc. All Rights Reserved.