@Generated(value="com.amazonaws:aws-java-sdk-code-generator") public class S3Settings extends Object implements Serializable, Cloneable, StructuredPojo
Settings for exporting data to Amazon S3.
Constructor and Description |
---|
S3Settings() |
Modifier and Type | Method and Description |
---|---|
S3Settings |
clone() |
boolean |
equals(Object obj) |
String |
getBucketFolder()
An optional parameter to set a folder name in the S3 bucket.
|
String |
getBucketName()
The name of the S3 bucket.
|
Boolean |
getCdcInsertsOnly()
Option to write only
INSERT operations to the comma-separated value (CSV) output files. |
String |
getCompressionType()
An optional parameter to use GZIP to compress the target files.
|
String |
getCsvDelimiter()
The delimiter used to separate columns in the source files.
|
String |
getCsvRowDelimiter()
The delimiter used to separate rows in the source files.
|
String |
getDataFormat()
The format of the data which you want to use for output.
|
Integer |
getDataPageSize()
The size of one data page in bytes.
|
Integer |
getDictPageSizeLimit()
The maximum size of an encoded dictionary page of a column.
|
Boolean |
getEnableStatistics()
Enables statistics for Parquet pages and rowGroups.
|
String |
getEncodingType()
The type of encoding you are using:
RLE_DICTIONARY (default), PLAIN , or
PLAIN_DICTIONARY . |
String |
getEncryptionMode()
The type of server side encryption you want to use for your data.
|
String |
getExternalTableDefinition()
The external table definition.
|
String |
getParquetVersion()
The version of Apache Parquet format you want to use:
PARQUET_1_0 (default) or
PARQUET_2_0 . |
Integer |
getRowGroupLength()
The number of rows in a row group.
|
String |
getServerSideEncryptionKmsKeyId()
If you are using SSE_KMS for the
EncryptionMode , provide the KMS Key ID. |
String |
getServiceAccessRoleArn()
The Amazon Resource Name (ARN) used by the service access IAM role.
|
int |
hashCode() |
Boolean |
isCdcInsertsOnly()
Option to write only
INSERT operations to the comma-separated value (CSV) output files. |
Boolean |
isEnableStatistics()
Enables statistics for Parquet pages and rowGroups.
|
void |
marshall(ProtocolMarshaller protocolMarshaller)
Marshalls this structured data using the given
ProtocolMarshaller . |
void |
setBucketFolder(String bucketFolder)
An optional parameter to set a folder name in the S3 bucket.
|
void |
setBucketName(String bucketName)
The name of the S3 bucket.
|
void |
setCdcInsertsOnly(Boolean cdcInsertsOnly)
Option to write only
INSERT operations to the comma-separated value (CSV) output files. |
void |
setCompressionType(CompressionTypeValue compressionType)
An optional parameter to use GZIP to compress the target files.
|
void |
setCompressionType(String compressionType)
An optional parameter to use GZIP to compress the target files.
|
void |
setCsvDelimiter(String csvDelimiter)
The delimiter used to separate columns in the source files.
|
void |
setCsvRowDelimiter(String csvRowDelimiter)
The delimiter used to separate rows in the source files.
|
void |
setDataFormat(DataFormatValue dataFormat)
The format of the data which you want to use for output.
|
void |
setDataFormat(String dataFormat)
The format of the data which you want to use for output.
|
void |
setDataPageSize(Integer dataPageSize)
The size of one data page in bytes.
|
void |
setDictPageSizeLimit(Integer dictPageSizeLimit)
The maximum size of an encoded dictionary page of a column.
|
void |
setEnableStatistics(Boolean enableStatistics)
Enables statistics for Parquet pages and rowGroups.
|
void |
setEncodingType(EncodingTypeValue encodingType)
The type of encoding you are using:
RLE_DICTIONARY (default), PLAIN , or
PLAIN_DICTIONARY . |
void |
setEncodingType(String encodingType)
The type of encoding you are using:
RLE_DICTIONARY (default), PLAIN , or
PLAIN_DICTIONARY . |
void |
setEncryptionMode(EncryptionModeValue encryptionMode)
The type of server side encryption you want to use for your data.
|
void |
setEncryptionMode(String encryptionMode)
The type of server side encryption you want to use for your data.
|
void |
setExternalTableDefinition(String externalTableDefinition)
The external table definition.
|
void |
setParquetVersion(ParquetVersionValue parquetVersion)
The version of Apache Parquet format you want to use:
PARQUET_1_0 (default) or
PARQUET_2_0 . |
void |
setParquetVersion(String parquetVersion)
The version of Apache Parquet format you want to use:
PARQUET_1_0 (default) or
PARQUET_2_0 . |
void |
setRowGroupLength(Integer rowGroupLength)
The number of rows in a row group.
|
void |
setServerSideEncryptionKmsKeyId(String serverSideEncryptionKmsKeyId)
If you are using SSE_KMS for the
EncryptionMode , provide the KMS Key ID. |
void |
setServiceAccessRoleArn(String serviceAccessRoleArn)
The Amazon Resource Name (ARN) used by the service access IAM role.
|
String |
toString()
Returns a string representation of this object.
|
S3Settings |
withBucketFolder(String bucketFolder)
An optional parameter to set a folder name in the S3 bucket.
|
S3Settings |
withBucketName(String bucketName)
The name of the S3 bucket.
|
S3Settings |
withCdcInsertsOnly(Boolean cdcInsertsOnly)
Option to write only
INSERT operations to the comma-separated value (CSV) output files. |
S3Settings |
withCompressionType(CompressionTypeValue compressionType)
An optional parameter to use GZIP to compress the target files.
|
S3Settings |
withCompressionType(String compressionType)
An optional parameter to use GZIP to compress the target files.
|
S3Settings |
withCsvDelimiter(String csvDelimiter)
The delimiter used to separate columns in the source files.
|
S3Settings |
withCsvRowDelimiter(String csvRowDelimiter)
The delimiter used to separate rows in the source files.
|
S3Settings |
withDataFormat(DataFormatValue dataFormat)
The format of the data which you want to use for output.
|
S3Settings |
withDataFormat(String dataFormat)
The format of the data which you want to use for output.
|
S3Settings |
withDataPageSize(Integer dataPageSize)
The size of one data page in bytes.
|
S3Settings |
withDictPageSizeLimit(Integer dictPageSizeLimit)
The maximum size of an encoded dictionary page of a column.
|
S3Settings |
withEnableStatistics(Boolean enableStatistics)
Enables statistics for Parquet pages and rowGroups.
|
S3Settings |
withEncodingType(EncodingTypeValue encodingType)
The type of encoding you are using:
RLE_DICTIONARY (default), PLAIN , or
PLAIN_DICTIONARY . |
S3Settings |
withEncodingType(String encodingType)
The type of encoding you are using:
RLE_DICTIONARY (default), PLAIN , or
PLAIN_DICTIONARY . |
S3Settings |
withEncryptionMode(EncryptionModeValue encryptionMode)
The type of server side encryption you want to use for your data.
|
S3Settings |
withEncryptionMode(String encryptionMode)
The type of server side encryption you want to use for your data.
|
S3Settings |
withExternalTableDefinition(String externalTableDefinition)
The external table definition.
|
S3Settings |
withParquetVersion(ParquetVersionValue parquetVersion)
The version of Apache Parquet format you want to use:
PARQUET_1_0 (default) or
PARQUET_2_0 . |
S3Settings |
withParquetVersion(String parquetVersion)
The version of Apache Parquet format you want to use:
PARQUET_1_0 (default) or
PARQUET_2_0 . |
S3Settings |
withRowGroupLength(Integer rowGroupLength)
The number of rows in a row group.
|
S3Settings |
withServerSideEncryptionKmsKeyId(String serverSideEncryptionKmsKeyId)
If you are using SSE_KMS for the
EncryptionMode , provide the KMS Key ID. |
S3Settings |
withServiceAccessRoleArn(String serviceAccessRoleArn)
The Amazon Resource Name (ARN) used by the service access IAM role.
|
public void setServiceAccessRoleArn(String serviceAccessRoleArn)
The Amazon Resource Name (ARN) used by the service access IAM role.
serviceAccessRoleArn
- The Amazon Resource Name (ARN) used by the service access IAM role.public String getServiceAccessRoleArn()
The Amazon Resource Name (ARN) used by the service access IAM role.
public S3Settings withServiceAccessRoleArn(String serviceAccessRoleArn)
The Amazon Resource Name (ARN) used by the service access IAM role.
serviceAccessRoleArn
- The Amazon Resource Name (ARN) used by the service access IAM role.public void setExternalTableDefinition(String externalTableDefinition)
The external table definition.
externalTableDefinition
- The external table definition.public String getExternalTableDefinition()
The external table definition.
public S3Settings withExternalTableDefinition(String externalTableDefinition)
The external table definition.
externalTableDefinition
- The external table definition.public void setCsvRowDelimiter(String csvRowDelimiter)
The delimiter used to separate rows in the source files. The default is a carriage return (\n
).
csvRowDelimiter
- The delimiter used to separate rows in the source files. The default is a carriage return (\n
).public String getCsvRowDelimiter()
The delimiter used to separate rows in the source files. The default is a carriage return (\n
).
\n
).public S3Settings withCsvRowDelimiter(String csvRowDelimiter)
The delimiter used to separate rows in the source files. The default is a carriage return (\n
).
csvRowDelimiter
- The delimiter used to separate rows in the source files. The default is a carriage return (\n
).public void setCsvDelimiter(String csvDelimiter)
The delimiter used to separate columns in the source files. The default is a comma.
csvDelimiter
- The delimiter used to separate columns in the source files. The default is a comma.public String getCsvDelimiter()
The delimiter used to separate columns in the source files. The default is a comma.
public S3Settings withCsvDelimiter(String csvDelimiter)
The delimiter used to separate columns in the source files. The default is a comma.
csvDelimiter
- The delimiter used to separate columns in the source files. The default is a comma.public void setBucketFolder(String bucketFolder)
An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
<bucketFolder>/<schema_name>/<table_name>/
. If this parameter is not specified,
then the path used is <schema_name>/<table_name>/
.
bucketFolder
- An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
<bucketFolder>/<schema_name>/<table_name>/
. If this parameter is not
specified, then the path used is <schema_name>/<table_name>/
.public String getBucketFolder()
An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
<bucketFolder>/<schema_name>/<table_name>/
. If this parameter is not specified,
then the path used is <schema_name>/<table_name>/
.
<bucketFolder>/<schema_name>/<table_name>/
. If this parameter is not
specified, then the path used is <schema_name>/<table_name>/
.public S3Settings withBucketFolder(String bucketFolder)
An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
<bucketFolder>/<schema_name>/<table_name>/
. If this parameter is not specified,
then the path used is <schema_name>/<table_name>/
.
bucketFolder
- An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
<bucketFolder>/<schema_name>/<table_name>/
. If this parameter is not
specified, then the path used is <schema_name>/<table_name>/
.public void setBucketName(String bucketName)
The name of the S3 bucket.
bucketName
- The name of the S3 bucket.public String getBucketName()
The name of the S3 bucket.
public S3Settings withBucketName(String bucketName)
The name of the S3 bucket.
bucketName
- The name of the S3 bucket.public void setCompressionType(String compressionType)
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET data formats.
compressionType
- An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files.
Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET
data formats.CompressionTypeValue
public String getCompressionType()
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET data formats.
CompressionTypeValue
public S3Settings withCompressionType(String compressionType)
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET data formats.
compressionType
- An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files.
Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET
data formats.CompressionTypeValue
public void setCompressionType(CompressionTypeValue compressionType)
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET data formats.
compressionType
- An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files.
Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET
data formats.CompressionTypeValue
public S3Settings withCompressionType(CompressionTypeValue compressionType)
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET data formats.
compressionType
- An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files.
Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET
data formats.CompressionTypeValue
public void setEncryptionMode(String encryptionMode)
The type of server side encryption you want to use for your data. This is part of the endpoint settings or the
extra connections attributes for Amazon S3. You can choose either SSE_S3
(default) or
SSE_KMS
. To use SSE_S3
, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*"
to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
encryptionMode
- The type of server side encryption you want to use for your data. This is part of the endpoint settings or
the extra connections attributes for Amazon S3. You can choose either SSE_S3
(default) or
SSE_KMS
. To use SSE_S3
, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*"
to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
EncryptionModeValue
public String getEncryptionMode()
The type of server side encryption you want to use for your data. This is part of the endpoint settings or the
extra connections attributes for Amazon S3. You can choose either SSE_S3
(default) or
SSE_KMS
. To use SSE_S3
, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*"
to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
SSE_S3
(default) or
SSE_KMS
. To use SSE_S3
, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*"
to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
EncryptionModeValue
public S3Settings withEncryptionMode(String encryptionMode)
The type of server side encryption you want to use for your data. This is part of the endpoint settings or the
extra connections attributes for Amazon S3. You can choose either SSE_S3
(default) or
SSE_KMS
. To use SSE_S3
, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*"
to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
encryptionMode
- The type of server side encryption you want to use for your data. This is part of the endpoint settings or
the extra connections attributes for Amazon S3. You can choose either SSE_S3
(default) or
SSE_KMS
. To use SSE_S3
, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*"
to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
EncryptionModeValue
public void setEncryptionMode(EncryptionModeValue encryptionMode)
The type of server side encryption you want to use for your data. This is part of the endpoint settings or the
extra connections attributes for Amazon S3. You can choose either SSE_S3
(default) or
SSE_KMS
. To use SSE_S3
, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*"
to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
encryptionMode
- The type of server side encryption you want to use for your data. This is part of the endpoint settings or
the extra connections attributes for Amazon S3. You can choose either SSE_S3
(default) or
SSE_KMS
. To use SSE_S3
, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*"
to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
EncryptionModeValue
public S3Settings withEncryptionMode(EncryptionModeValue encryptionMode)
The type of server side encryption you want to use for your data. This is part of the endpoint settings or the
extra connections attributes for Amazon S3. You can choose either SSE_S3
(default) or
SSE_KMS
. To use SSE_S3
, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*"
to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
encryptionMode
- The type of server side encryption you want to use for your data. This is part of the endpoint settings or
the extra connections attributes for Amazon S3. You can choose either SSE_S3
(default) or
SSE_KMS
. To use SSE_S3
, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*"
to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
EncryptionModeValue
public void setServerSideEncryptionKmsKeyId(String serverSideEncryptionKmsKeyId)
If you are using SSE_KMS for the EncryptionMode
, provide the KMS Key ID. The key you use needs an
attached policy that enables IAM user permissions and allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier <value> --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=<value>,BucketFolder=<value>,BucketName=<value>,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=<value>
serverSideEncryptionKmsKeyId
- If you are using SSE_KMS for the EncryptionMode
, provide the KMS Key ID. The key you use
needs an attached policy that enables IAM user permissions and allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier <value> --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=<value>,BucketFolder=<value>,BucketName=<value>,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=<value>
public String getServerSideEncryptionKmsKeyId()
If you are using SSE_KMS for the EncryptionMode
, provide the KMS Key ID. The key you use needs an
attached policy that enables IAM user permissions and allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier <value> --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=<value>,BucketFolder=<value>,BucketName=<value>,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=<value>
EncryptionMode
, provide the KMS Key ID. The key you use
needs an attached policy that enables IAM user permissions and allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier <value> --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=<value>,BucketFolder=<value>,BucketName=<value>,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=<value>
public S3Settings withServerSideEncryptionKmsKeyId(String serverSideEncryptionKmsKeyId)
If you are using SSE_KMS for the EncryptionMode
, provide the KMS Key ID. The key you use needs an
attached policy that enables IAM user permissions and allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier <value> --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=<value>,BucketFolder=<value>,BucketName=<value>,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=<value>
serverSideEncryptionKmsKeyId
- If you are using SSE_KMS for the EncryptionMode
, provide the KMS Key ID. The key you use
needs an attached policy that enables IAM user permissions and allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier <value> --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=<value>,BucketFolder=<value>,BucketName=<value>,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=<value>
public void setDataFormat(String dataFormat)
The format of the data which you want to use for output. You can choose one of the following:
CSV
: This is a row-based format with comma-separated values.
PARQUET
: Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
dataFormat
- The format of the data which you want to use for output. You can choose one of the following:
CSV
: This is a row-based format with comma-separated values.
PARQUET
: Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
DataFormatValue
public String getDataFormat()
The format of the data which you want to use for output. You can choose one of the following:
CSV
: This is a row-based format with comma-separated values.
PARQUET
: Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
CSV
: This is a row-based format with comma-separated values.
PARQUET
: Apache Parquet is a columnar storage format that features efficient compression
and provides faster query response.
DataFormatValue
public S3Settings withDataFormat(String dataFormat)
The format of the data which you want to use for output. You can choose one of the following:
CSV
: This is a row-based format with comma-separated values.
PARQUET
: Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
dataFormat
- The format of the data which you want to use for output. You can choose one of the following:
CSV
: This is a row-based format with comma-separated values.
PARQUET
: Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
DataFormatValue
public void setDataFormat(DataFormatValue dataFormat)
The format of the data which you want to use for output. You can choose one of the following:
CSV
: This is a row-based format with comma-separated values.
PARQUET
: Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
dataFormat
- The format of the data which you want to use for output. You can choose one of the following:
CSV
: This is a row-based format with comma-separated values.
PARQUET
: Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
DataFormatValue
public S3Settings withDataFormat(DataFormatValue dataFormat)
The format of the data which you want to use for output. You can choose one of the following:
CSV
: This is a row-based format with comma-separated values.
PARQUET
: Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
dataFormat
- The format of the data which you want to use for output. You can choose one of the following:
CSV
: This is a row-based format with comma-separated values.
PARQUET
: Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
DataFormatValue
public void setEncodingType(String encodingType)
The type of encoding you are using: RLE_DICTIONARY
(default), PLAIN
, or
PLAIN_DICTIONARY
.
RLE_DICTIONARY
uses a combination of bit-packing and run-length encoding to store repeated values
more efficiently.
PLAIN
does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY
builds a dictionary of the values encountered in a given column. The dictionary is
stored in a dictionary page for each column chunk.
encodingType
- The type of encoding you are using: RLE_DICTIONARY
(default), PLAIN
, or
PLAIN_DICTIONARY
.
RLE_DICTIONARY
uses a combination of bit-packing and run-length encoding to store repeated
values more efficiently.
PLAIN
does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY
builds a dictionary of the values encountered in a given column. The
dictionary is stored in a dictionary page for each column chunk.
EncodingTypeValue
public String getEncodingType()
The type of encoding you are using: RLE_DICTIONARY
(default), PLAIN
, or
PLAIN_DICTIONARY
.
RLE_DICTIONARY
uses a combination of bit-packing and run-length encoding to store repeated values
more efficiently.
PLAIN
does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY
builds a dictionary of the values encountered in a given column. The dictionary is
stored in a dictionary page for each column chunk.
RLE_DICTIONARY
(default), PLAIN
, or
PLAIN_DICTIONARY
.
RLE_DICTIONARY
uses a combination of bit-packing and run-length encoding to store repeated
values more efficiently.
PLAIN
does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY
builds a dictionary of the values encountered in a given column. The
dictionary is stored in a dictionary page for each column chunk.
EncodingTypeValue
public S3Settings withEncodingType(String encodingType)
The type of encoding you are using: RLE_DICTIONARY
(default), PLAIN
, or
PLAIN_DICTIONARY
.
RLE_DICTIONARY
uses a combination of bit-packing and run-length encoding to store repeated values
more efficiently.
PLAIN
does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY
builds a dictionary of the values encountered in a given column. The dictionary is
stored in a dictionary page for each column chunk.
encodingType
- The type of encoding you are using: RLE_DICTIONARY
(default), PLAIN
, or
PLAIN_DICTIONARY
.
RLE_DICTIONARY
uses a combination of bit-packing and run-length encoding to store repeated
values more efficiently.
PLAIN
does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY
builds a dictionary of the values encountered in a given column. The
dictionary is stored in a dictionary page for each column chunk.
EncodingTypeValue
public void setEncodingType(EncodingTypeValue encodingType)
The type of encoding you are using: RLE_DICTIONARY
(default), PLAIN
, or
PLAIN_DICTIONARY
.
RLE_DICTIONARY
uses a combination of bit-packing and run-length encoding to store repeated values
more efficiently.
PLAIN
does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY
builds a dictionary of the values encountered in a given column. The dictionary is
stored in a dictionary page for each column chunk.
encodingType
- The type of encoding you are using: RLE_DICTIONARY
(default), PLAIN
, or
PLAIN_DICTIONARY
.
RLE_DICTIONARY
uses a combination of bit-packing and run-length encoding to store repeated
values more efficiently.
PLAIN
does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY
builds a dictionary of the values encountered in a given column. The
dictionary is stored in a dictionary page for each column chunk.
EncodingTypeValue
public S3Settings withEncodingType(EncodingTypeValue encodingType)
The type of encoding you are using: RLE_DICTIONARY
(default), PLAIN
, or
PLAIN_DICTIONARY
.
RLE_DICTIONARY
uses a combination of bit-packing and run-length encoding to store repeated values
more efficiently.
PLAIN
does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY
builds a dictionary of the values encountered in a given column. The dictionary is
stored in a dictionary page for each column chunk.
encodingType
- The type of encoding you are using: RLE_DICTIONARY
(default), PLAIN
, or
PLAIN_DICTIONARY
.
RLE_DICTIONARY
uses a combination of bit-packing and run-length encoding to store repeated
values more efficiently.
PLAIN
does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY
builds a dictionary of the values encountered in a given column. The
dictionary is stored in a dictionary page for each column chunk.
EncodingTypeValue
public void setDictPageSizeLimit(Integer dictPageSizeLimit)
The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this column is
stored using an encoding type of PLAIN
. Defaults to 1024 * 1024 bytes (1MiB), the maximum size of a
dictionary page before it reverts to PLAIN
encoding. For PARQUET
format only.
dictPageSizeLimit
- The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this
column is stored using an encoding type of PLAIN
. Defaults to 1024 * 1024 bytes (1MiB), the
maximum size of a dictionary page before it reverts to PLAIN
encoding. For
PARQUET
format only.public Integer getDictPageSizeLimit()
The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this column is
stored using an encoding type of PLAIN
. Defaults to 1024 * 1024 bytes (1MiB), the maximum size of a
dictionary page before it reverts to PLAIN
encoding. For PARQUET
format only.
PLAIN
. Defaults to 1024 * 1024 bytes (1MiB), the
maximum size of a dictionary page before it reverts to PLAIN
encoding. For
PARQUET
format only.public S3Settings withDictPageSizeLimit(Integer dictPageSizeLimit)
The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this column is
stored using an encoding type of PLAIN
. Defaults to 1024 * 1024 bytes (1MiB), the maximum size of a
dictionary page before it reverts to PLAIN
encoding. For PARQUET
format only.
dictPageSizeLimit
- The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this
column is stored using an encoding type of PLAIN
. Defaults to 1024 * 1024 bytes (1MiB), the
maximum size of a dictionary page before it reverts to PLAIN
encoding. For
PARQUET
format only.public void setRowGroupLength(Integer rowGroupLength)
The number of rows in a row group. A smaller row group size provides faster reads. But as the number of row
groups grows, the slower writes become. Defaults to 10,000 (ten thousand) rows. For PARQUET
format
only.
If you choose a value larger than the maximum, RowGroupLength
is set to the max row group length in
bytes (64 * 1024 * 1024).
rowGroupLength
- The number of rows in a row group. A smaller row group size provides faster reads. But as the number of
row groups grows, the slower writes become. Defaults to 10,000 (ten thousand) rows. For
PARQUET
format only.
If you choose a value larger than the maximum, RowGroupLength
is set to the max row group
length in bytes (64 * 1024 * 1024).
public Integer getRowGroupLength()
The number of rows in a row group. A smaller row group size provides faster reads. But as the number of row
groups grows, the slower writes become. Defaults to 10,000 (ten thousand) rows. For PARQUET
format
only.
If you choose a value larger than the maximum, RowGroupLength
is set to the max row group length in
bytes (64 * 1024 * 1024).
PARQUET
format only.
If you choose a value larger than the maximum, RowGroupLength
is set to the max row group
length in bytes (64 * 1024 * 1024).
public S3Settings withRowGroupLength(Integer rowGroupLength)
The number of rows in a row group. A smaller row group size provides faster reads. But as the number of row
groups grows, the slower writes become. Defaults to 10,000 (ten thousand) rows. For PARQUET
format
only.
If you choose a value larger than the maximum, RowGroupLength
is set to the max row group length in
bytes (64 * 1024 * 1024).
rowGroupLength
- The number of rows in a row group. A smaller row group size provides faster reads. But as the number of
row groups grows, the slower writes become. Defaults to 10,000 (ten thousand) rows. For
PARQUET
format only.
If you choose a value larger than the maximum, RowGroupLength
is set to the max row group
length in bytes (64 * 1024 * 1024).
public void setDataPageSize(Integer dataPageSize)
The size of one data page in bytes. Defaults to 1024 * 1024 bytes (1MiB). For PARQUET
format only.
dataPageSize
- The size of one data page in bytes. Defaults to 1024 * 1024 bytes (1MiB). For PARQUET
format
only.public Integer getDataPageSize()
The size of one data page in bytes. Defaults to 1024 * 1024 bytes (1MiB). For PARQUET
format only.
PARQUET
format
only.public S3Settings withDataPageSize(Integer dataPageSize)
The size of one data page in bytes. Defaults to 1024 * 1024 bytes (1MiB). For PARQUET
format only.
dataPageSize
- The size of one data page in bytes. Defaults to 1024 * 1024 bytes (1MiB). For PARQUET
format
only.public void setParquetVersion(String parquetVersion)
The version of Apache Parquet format you want to use: PARQUET_1_0
(default) or
PARQUET_2_0
.
parquetVersion
- The version of Apache Parquet format you want to use: PARQUET_1_0
(default) or
PARQUET_2_0
.ParquetVersionValue
public String getParquetVersion()
The version of Apache Parquet format you want to use: PARQUET_1_0
(default) or
PARQUET_2_0
.
PARQUET_1_0
(default) or
PARQUET_2_0
.ParquetVersionValue
public S3Settings withParquetVersion(String parquetVersion)
The version of Apache Parquet format you want to use: PARQUET_1_0
(default) or
PARQUET_2_0
.
parquetVersion
- The version of Apache Parquet format you want to use: PARQUET_1_0
(default) or
PARQUET_2_0
.ParquetVersionValue
public void setParquetVersion(ParquetVersionValue parquetVersion)
The version of Apache Parquet format you want to use: PARQUET_1_0
(default) or
PARQUET_2_0
.
parquetVersion
- The version of Apache Parquet format you want to use: PARQUET_1_0
(default) or
PARQUET_2_0
.ParquetVersionValue
public S3Settings withParquetVersion(ParquetVersionValue parquetVersion)
The version of Apache Parquet format you want to use: PARQUET_1_0
(default) or
PARQUET_2_0
.
parquetVersion
- The version of Apache Parquet format you want to use: PARQUET_1_0
(default) or
PARQUET_2_0
.ParquetVersionValue
public void setEnableStatistics(Boolean enableStatistics)
Enables statistics for Parquet pages and rowGroups. Choose TRUE
to enable statistics, choose
FALSE
to disable. Statistics include NULL
, DISTINCT
, MAX
, and
MIN
values. Defaults to TRUE
. For PARQUET
format only.
enableStatistics
- Enables statistics for Parquet pages and rowGroups. Choose TRUE
to enable statistics, choose
FALSE
to disable. Statistics include NULL
, DISTINCT
,
MAX
, and MIN
values. Defaults to TRUE
. For PARQUET
format only.public Boolean getEnableStatistics()
Enables statistics for Parquet pages and rowGroups. Choose TRUE
to enable statistics, choose
FALSE
to disable. Statistics include NULL
, DISTINCT
, MAX
, and
MIN
values. Defaults to TRUE
. For PARQUET
format only.
TRUE
to enable statistics, choose
FALSE
to disable. Statistics include NULL
, DISTINCT
,
MAX
, and MIN
values. Defaults to TRUE
. For PARQUET
format only.public S3Settings withEnableStatistics(Boolean enableStatistics)
Enables statistics for Parquet pages and rowGroups. Choose TRUE
to enable statistics, choose
FALSE
to disable. Statistics include NULL
, DISTINCT
, MAX
, and
MIN
values. Defaults to TRUE
. For PARQUET
format only.
enableStatistics
- Enables statistics for Parquet pages and rowGroups. Choose TRUE
to enable statistics, choose
FALSE
to disable. Statistics include NULL
, DISTINCT
,
MAX
, and MIN
values. Defaults to TRUE
. For PARQUET
format only.public Boolean isEnableStatistics()
Enables statistics for Parquet pages and rowGroups. Choose TRUE
to enable statistics, choose
FALSE
to disable. Statistics include NULL
, DISTINCT
, MAX
, and
MIN
values. Defaults to TRUE
. For PARQUET
format only.
TRUE
to enable statistics, choose
FALSE
to disable. Statistics include NULL
, DISTINCT
,
MAX
, and MIN
values. Defaults to TRUE
. For PARQUET
format only.public void setCdcInsertsOnly(Boolean cdcInsertsOnly)
Option to write only INSERT
operations to the comma-separated value (CSV) output files. By default,
the first field in a CSV record contains the letter I
(insert), U
(update) or
D
(delete) to indicate whether the row was inserted, updated, or deleted at the source database. If
cdcInsertsOnly
is set to true, then only INSERT
s are recorded in the CSV file, without
the I
annotation on each line. Valid values are TRUE
and FALSE
.
cdcInsertsOnly
- Option to write only INSERT
operations to the comma-separated value (CSV) output files. By
default, the first field in a CSV record contains the letter I
(insert), U
(update) or D
(delete) to indicate whether the row was inserted, updated, or deleted at the
source database. If cdcInsertsOnly
is set to true, then only INSERT
s are
recorded in the CSV file, without the I
annotation on each line. Valid values are
TRUE
and FALSE
.public Boolean getCdcInsertsOnly()
Option to write only INSERT
operations to the comma-separated value (CSV) output files. By default,
the first field in a CSV record contains the letter I
(insert), U
(update) or
D
(delete) to indicate whether the row was inserted, updated, or deleted at the source database. If
cdcInsertsOnly
is set to true, then only INSERT
s are recorded in the CSV file, without
the I
annotation on each line. Valid values are TRUE
and FALSE
.
INSERT
operations to the comma-separated value (CSV) output files. By
default, the first field in a CSV record contains the letter I
(insert), U
(update) or D
(delete) to indicate whether the row was inserted, updated, or deleted at the
source database. If cdcInsertsOnly
is set to true, then only INSERT
s are
recorded in the CSV file, without the I
annotation on each line. Valid values are
TRUE
and FALSE
.public S3Settings withCdcInsertsOnly(Boolean cdcInsertsOnly)
Option to write only INSERT
operations to the comma-separated value (CSV) output files. By default,
the first field in a CSV record contains the letter I
(insert), U
(update) or
D
(delete) to indicate whether the row was inserted, updated, or deleted at the source database. If
cdcInsertsOnly
is set to true, then only INSERT
s are recorded in the CSV file, without
the I
annotation on each line. Valid values are TRUE
and FALSE
.
cdcInsertsOnly
- Option to write only INSERT
operations to the comma-separated value (CSV) output files. By
default, the first field in a CSV record contains the letter I
(insert), U
(update) or D
(delete) to indicate whether the row was inserted, updated, or deleted at the
source database. If cdcInsertsOnly
is set to true, then only INSERT
s are
recorded in the CSV file, without the I
annotation on each line. Valid values are
TRUE
and FALSE
.public Boolean isCdcInsertsOnly()
Option to write only INSERT
operations to the comma-separated value (CSV) output files. By default,
the first field in a CSV record contains the letter I
(insert), U
(update) or
D
(delete) to indicate whether the row was inserted, updated, or deleted at the source database. If
cdcInsertsOnly
is set to true, then only INSERT
s are recorded in the CSV file, without
the I
annotation on each line. Valid values are TRUE
and FALSE
.
INSERT
operations to the comma-separated value (CSV) output files. By
default, the first field in a CSV record contains the letter I
(insert), U
(update) or D
(delete) to indicate whether the row was inserted, updated, or deleted at the
source database. If cdcInsertsOnly
is set to true, then only INSERT
s are
recorded in the CSV file, without the I
annotation on each line. Valid values are
TRUE
and FALSE
.public String toString()
toString
in class Object
Object.toString()
public S3Settings clone()
public void marshall(ProtocolMarshaller protocolMarshaller)
StructuredPojo
ProtocolMarshaller
.marshall
in interface StructuredPojo
protocolMarshaller
- Implementation of ProtocolMarshaller
used to marshall this object's data.Copyright © 2013 Amazon Web Services, Inc. All Rights Reserved.