@Generated(value="com.amazonaws:aws-java-sdk-code-generator") public class S3Settings extends Object implements Serializable, Cloneable, StructuredPojo
Settings for exporting data to Amazon S3.
| Constructor and Description |
|---|
S3Settings() |
| Modifier and Type | Method and Description |
|---|---|
S3Settings |
clone() |
boolean |
equals(Object obj) |
String |
getBucketFolder()
An optional parameter to set a folder name in the S3 bucket.
|
String |
getBucketName()
The name of the S3 bucket.
|
Boolean |
getCdcInsertsOnly()
Option to write only
INSERT operations to the comma-separated value (CSV) output files. |
String |
getCompressionType()
An optional parameter to use GZIP to compress the target files.
|
String |
getCsvDelimiter()
The delimiter used to separate columns in the source files.
|
String |
getCsvRowDelimiter()
The delimiter used to separate rows in the source files.
|
String |
getDataFormat()
The format of the data which you want to use for output.
|
Integer |
getDataPageSize()
The size of one data page in bytes.
|
Integer |
getDictPageSizeLimit()
The maximum size of an encoded dictionary page of a column.
|
Boolean |
getEnableStatistics()
Enables statistics for Parquet pages and rowGroups.
|
String |
getEncodingType()
The type of encoding you are using:
RLE_DICTIONARY (default), PLAIN, or
PLAIN_DICTIONARY. |
String |
getEncryptionMode()
The type of server side encryption you want to use for your data.
|
String |
getExternalTableDefinition()
The external table definition.
|
String |
getParquetVersion()
The version of Apache Parquet format you want to use:
PARQUET_1_0 (default) or
PARQUET_2_0. |
Integer |
getRowGroupLength()
The number of rows in a row group.
|
String |
getServerSideEncryptionKmsKeyId()
If you are using SSE_KMS for the
EncryptionMode, provide the KMS Key ID. |
String |
getServiceAccessRoleArn()
The Amazon Resource Name (ARN) used by the service access IAM role.
|
int |
hashCode() |
Boolean |
isCdcInsertsOnly()
Option to write only
INSERT operations to the comma-separated value (CSV) output files. |
Boolean |
isEnableStatistics()
Enables statistics for Parquet pages and rowGroups.
|
void |
marshall(ProtocolMarshaller protocolMarshaller) |
void |
setBucketFolder(String bucketFolder)
An optional parameter to set a folder name in the S3 bucket.
|
void |
setBucketName(String bucketName)
The name of the S3 bucket.
|
void |
setCdcInsertsOnly(Boolean cdcInsertsOnly)
Option to write only
INSERT operations to the comma-separated value (CSV) output files. |
void |
setCompressionType(CompressionTypeValue compressionType)
An optional parameter to use GZIP to compress the target files.
|
void |
setCompressionType(String compressionType)
An optional parameter to use GZIP to compress the target files.
|
void |
setCsvDelimiter(String csvDelimiter)
The delimiter used to separate columns in the source files.
|
void |
setCsvRowDelimiter(String csvRowDelimiter)
The delimiter used to separate rows in the source files.
|
void |
setDataFormat(DataFormatValue dataFormat)
The format of the data which you want to use for output.
|
void |
setDataFormat(String dataFormat)
The format of the data which you want to use for output.
|
void |
setDataPageSize(Integer dataPageSize)
The size of one data page in bytes.
|
void |
setDictPageSizeLimit(Integer dictPageSizeLimit)
The maximum size of an encoded dictionary page of a column.
|
void |
setEnableStatistics(Boolean enableStatistics)
Enables statistics for Parquet pages and rowGroups.
|
void |
setEncodingType(EncodingTypeValue encodingType)
The type of encoding you are using:
RLE_DICTIONARY (default), PLAIN, or
PLAIN_DICTIONARY. |
void |
setEncodingType(String encodingType)
The type of encoding you are using:
RLE_DICTIONARY (default), PLAIN, or
PLAIN_DICTIONARY. |
void |
setEncryptionMode(EncryptionModeValue encryptionMode)
The type of server side encryption you want to use for your data.
|
void |
setEncryptionMode(String encryptionMode)
The type of server side encryption you want to use for your data.
|
void |
setExternalTableDefinition(String externalTableDefinition)
The external table definition.
|
void |
setParquetVersion(ParquetVersionValue parquetVersion)
The version of Apache Parquet format you want to use:
PARQUET_1_0 (default) or
PARQUET_2_0. |
void |
setParquetVersion(String parquetVersion)
The version of Apache Parquet format you want to use:
PARQUET_1_0 (default) or
PARQUET_2_0. |
void |
setRowGroupLength(Integer rowGroupLength)
The number of rows in a row group.
|
void |
setServerSideEncryptionKmsKeyId(String serverSideEncryptionKmsKeyId)
If you are using SSE_KMS for the
EncryptionMode, provide the KMS Key ID. |
void |
setServiceAccessRoleArn(String serviceAccessRoleArn)
The Amazon Resource Name (ARN) used by the service access IAM role.
|
String |
toString()
Returns a string representation of this object.
|
S3Settings |
withBucketFolder(String bucketFolder)
An optional parameter to set a folder name in the S3 bucket.
|
S3Settings |
withBucketName(String bucketName)
The name of the S3 bucket.
|
S3Settings |
withCdcInsertsOnly(Boolean cdcInsertsOnly)
Option to write only
INSERT operations to the comma-separated value (CSV) output files. |
S3Settings |
withCompressionType(CompressionTypeValue compressionType)
An optional parameter to use GZIP to compress the target files.
|
S3Settings |
withCompressionType(String compressionType)
An optional parameter to use GZIP to compress the target files.
|
S3Settings |
withCsvDelimiter(String csvDelimiter)
The delimiter used to separate columns in the source files.
|
S3Settings |
withCsvRowDelimiter(String csvRowDelimiter)
The delimiter used to separate rows in the source files.
|
S3Settings |
withDataFormat(DataFormatValue dataFormat)
The format of the data which you want to use for output.
|
S3Settings |
withDataFormat(String dataFormat)
The format of the data which you want to use for output.
|
S3Settings |
withDataPageSize(Integer dataPageSize)
The size of one data page in bytes.
|
S3Settings |
withDictPageSizeLimit(Integer dictPageSizeLimit)
The maximum size of an encoded dictionary page of a column.
|
S3Settings |
withEnableStatistics(Boolean enableStatistics)
Enables statistics for Parquet pages and rowGroups.
|
S3Settings |
withEncodingType(EncodingTypeValue encodingType)
The type of encoding you are using:
RLE_DICTIONARY (default), PLAIN, or
PLAIN_DICTIONARY. |
S3Settings |
withEncodingType(String encodingType)
The type of encoding you are using:
RLE_DICTIONARY (default), PLAIN, or
PLAIN_DICTIONARY. |
S3Settings |
withEncryptionMode(EncryptionModeValue encryptionMode)
The type of server side encryption you want to use for your data.
|
S3Settings |
withEncryptionMode(String encryptionMode)
The type of server side encryption you want to use for your data.
|
S3Settings |
withExternalTableDefinition(String externalTableDefinition)
The external table definition.
|
S3Settings |
withParquetVersion(ParquetVersionValue parquetVersion)
The version of Apache Parquet format you want to use:
PARQUET_1_0 (default) or
PARQUET_2_0. |
S3Settings |
withParquetVersion(String parquetVersion)
The version of Apache Parquet format you want to use:
PARQUET_1_0 (default) or
PARQUET_2_0. |
S3Settings |
withRowGroupLength(Integer rowGroupLength)
The number of rows in a row group.
|
S3Settings |
withServerSideEncryptionKmsKeyId(String serverSideEncryptionKmsKeyId)
If you are using SSE_KMS for the
EncryptionMode, provide the KMS Key ID. |
S3Settings |
withServiceAccessRoleArn(String serviceAccessRoleArn)
The Amazon Resource Name (ARN) used by the service access IAM role.
|
public void setServiceAccessRoleArn(String serviceAccessRoleArn)
The Amazon Resource Name (ARN) used by the service access IAM role.
serviceAccessRoleArn - The Amazon Resource Name (ARN) used by the service access IAM role.public String getServiceAccessRoleArn()
The Amazon Resource Name (ARN) used by the service access IAM role.
public S3Settings withServiceAccessRoleArn(String serviceAccessRoleArn)
The Amazon Resource Name (ARN) used by the service access IAM role.
serviceAccessRoleArn - The Amazon Resource Name (ARN) used by the service access IAM role.public void setExternalTableDefinition(String externalTableDefinition)
The external table definition.
externalTableDefinition - The external table definition.public String getExternalTableDefinition()
The external table definition.
public S3Settings withExternalTableDefinition(String externalTableDefinition)
The external table definition.
externalTableDefinition - The external table definition.public void setCsvRowDelimiter(String csvRowDelimiter)
The delimiter used to separate rows in the source files. The default is a carriage return (\n).
csvRowDelimiter - The delimiter used to separate rows in the source files. The default is a carriage return (\n
).public String getCsvRowDelimiter()
The delimiter used to separate rows in the source files. The default is a carriage return (\n).
\n).public S3Settings withCsvRowDelimiter(String csvRowDelimiter)
The delimiter used to separate rows in the source files. The default is a carriage return (\n).
csvRowDelimiter - The delimiter used to separate rows in the source files. The default is a carriage return (\n
).public void setCsvDelimiter(String csvDelimiter)
The delimiter used to separate columns in the source files. The default is a comma.
csvDelimiter - The delimiter used to separate columns in the source files. The default is a comma.public String getCsvDelimiter()
The delimiter used to separate columns in the source files. The default is a comma.
public S3Settings withCsvDelimiter(String csvDelimiter)
The delimiter used to separate columns in the source files. The default is a comma.
csvDelimiter - The delimiter used to separate columns in the source files. The default is a comma.public void setBucketFolder(String bucketFolder)
An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
<bucketFolder>/<schema_name>/<table_name>/. If this parameter is not specified,
then the path used is <schema_name>/<table_name>/.
bucketFolder - An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
<bucketFolder>/<schema_name>/<table_name>/. If this parameter is not
specified, then the path used is <schema_name>/<table_name>/.public String getBucketFolder()
An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
<bucketFolder>/<schema_name>/<table_name>/. If this parameter is not specified,
then the path used is <schema_name>/<table_name>/.
<bucketFolder>/<schema_name>/<table_name>/. If this parameter is not
specified, then the path used is <schema_name>/<table_name>/.public S3Settings withBucketFolder(String bucketFolder)
An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
<bucketFolder>/<schema_name>/<table_name>/. If this parameter is not specified,
then the path used is <schema_name>/<table_name>/.
bucketFolder - An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
<bucketFolder>/<schema_name>/<table_name>/. If this parameter is not
specified, then the path used is <schema_name>/<table_name>/.public void setBucketName(String bucketName)
The name of the S3 bucket.
bucketName - The name of the S3 bucket.public String getBucketName()
The name of the S3 bucket.
public S3Settings withBucketName(String bucketName)
The name of the S3 bucket.
bucketName - The name of the S3 bucket.public void setCompressionType(String compressionType)
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET data formats.
compressionType - An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files.
Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET
data formats.CompressionTypeValuepublic String getCompressionType()
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET data formats.
CompressionTypeValuepublic S3Settings withCompressionType(String compressionType)
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET data formats.
compressionType - An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files.
Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET
data formats.CompressionTypeValuepublic void setCompressionType(CompressionTypeValue compressionType)
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET data formats.
compressionType - An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files.
Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET
data formats.CompressionTypeValuepublic S3Settings withCompressionType(CompressionTypeValue compressionType)
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET data formats.
compressionType - An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files.
Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both CSV and PARQUET
data formats.CompressionTypeValuepublic void setEncryptionMode(String encryptionMode)
The type of server side encryption you want to use for your data. This is part of the endpoint settings or the
extra connections attributes for Amazon S3. You can choose either SSE_S3 (default) or
SSE_KMS. To use SSE_S3, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*" to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
encryptionMode - The type of server side encryption you want to use for your data. This is part of the endpoint settings or
the extra connections attributes for Amazon S3. You can choose either SSE_S3 (default) or
SSE_KMS. To use SSE_S3, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*" to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
EncryptionModeValuepublic String getEncryptionMode()
The type of server side encryption you want to use for your data. This is part of the endpoint settings or the
extra connections attributes for Amazon S3. You can choose either SSE_S3 (default) or
SSE_KMS. To use SSE_S3, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*" to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
SSE_S3 (default) or
SSE_KMS. To use SSE_S3, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*" to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
EncryptionModeValuepublic S3Settings withEncryptionMode(String encryptionMode)
The type of server side encryption you want to use for your data. This is part of the endpoint settings or the
extra connections attributes for Amazon S3. You can choose either SSE_S3 (default) or
SSE_KMS. To use SSE_S3, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*" to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
encryptionMode - The type of server side encryption you want to use for your data. This is part of the endpoint settings or
the extra connections attributes for Amazon S3. You can choose either SSE_S3 (default) or
SSE_KMS. To use SSE_S3, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*" to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
EncryptionModeValuepublic void setEncryptionMode(EncryptionModeValue encryptionMode)
The type of server side encryption you want to use for your data. This is part of the endpoint settings or the
extra connections attributes for Amazon S3. You can choose either SSE_S3 (default) or
SSE_KMS. To use SSE_S3, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*" to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
encryptionMode - The type of server side encryption you want to use for your data. This is part of the endpoint settings or
the extra connections attributes for Amazon S3. You can choose either SSE_S3 (default) or
SSE_KMS. To use SSE_S3, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*" to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
EncryptionModeValuepublic S3Settings withEncryptionMode(EncryptionModeValue encryptionMode)
The type of server side encryption you want to use for your data. This is part of the endpoint settings or the
extra connections attributes for Amazon S3. You can choose either SSE_S3 (default) or
SSE_KMS. To use SSE_S3, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*" to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
encryptionMode - The type of server side encryption you want to use for your data. This is part of the endpoint settings or
the extra connections attributes for Amazon S3. You can choose either SSE_S3 (default) or
SSE_KMS. To use SSE_S3, you need an IAM role with permission to allow
"arn:aws:s3:::dms-*" to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
EncryptionModeValuepublic void setServerSideEncryptionKmsKeyId(String serverSideEncryptionKmsKeyId)
If you are using SSE_KMS for the EncryptionMode, provide the KMS Key ID. The key you use needs an
attached policy that enables IAM user permissions and allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier <value> --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=<value>,BucketFolder=<value>,BucketName=<value>,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=<value>
serverSideEncryptionKmsKeyId - If you are using SSE_KMS for the EncryptionMode, provide the KMS Key ID. The key you use
needs an attached policy that enables IAM user permissions and allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier <value> --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=<value>,BucketFolder=<value>,BucketName=<value>,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=<value>
public String getServerSideEncryptionKmsKeyId()
If you are using SSE_KMS for the EncryptionMode, provide the KMS Key ID. The key you use needs an
attached policy that enables IAM user permissions and allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier <value> --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=<value>,BucketFolder=<value>,BucketName=<value>,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=<value>
EncryptionMode, provide the KMS Key ID. The key you use
needs an attached policy that enables IAM user permissions and allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier <value> --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=<value>,BucketFolder=<value>,BucketName=<value>,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=<value>
public S3Settings withServerSideEncryptionKmsKeyId(String serverSideEncryptionKmsKeyId)
If you are using SSE_KMS for the EncryptionMode, provide the KMS Key ID. The key you use needs an
attached policy that enables IAM user permissions and allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier <value> --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=<value>,BucketFolder=<value>,BucketName=<value>,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=<value>
serverSideEncryptionKmsKeyId - If you are using SSE_KMS for the EncryptionMode, provide the KMS Key ID. The key you use
needs an attached policy that enables IAM user permissions and allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier <value> --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=<value>,BucketFolder=<value>,BucketName=<value>,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=<value>
public void setDataFormat(String dataFormat)
The format of the data which you want to use for output. You can choose one of the following:
CSV : This is a row-based format with comma-separated values.
PARQUET : Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
dataFormat - The format of the data which you want to use for output. You can choose one of the following:
CSV : This is a row-based format with comma-separated values.
PARQUET : Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
DataFormatValuepublic String getDataFormat()
The format of the data which you want to use for output. You can choose one of the following:
CSV : This is a row-based format with comma-separated values.
PARQUET : Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
CSV : This is a row-based format with comma-separated values.
PARQUET : Apache Parquet is a columnar storage format that features efficient compression
and provides faster query response.
DataFormatValuepublic S3Settings withDataFormat(String dataFormat)
The format of the data which you want to use for output. You can choose one of the following:
CSV : This is a row-based format with comma-separated values.
PARQUET : Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
dataFormat - The format of the data which you want to use for output. You can choose one of the following:
CSV : This is a row-based format with comma-separated values.
PARQUET : Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
DataFormatValuepublic void setDataFormat(DataFormatValue dataFormat)
The format of the data which you want to use for output. You can choose one of the following:
CSV : This is a row-based format with comma-separated values.
PARQUET : Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
dataFormat - The format of the data which you want to use for output. You can choose one of the following:
CSV : This is a row-based format with comma-separated values.
PARQUET : Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
DataFormatValuepublic S3Settings withDataFormat(DataFormatValue dataFormat)
The format of the data which you want to use for output. You can choose one of the following:
CSV : This is a row-based format with comma-separated values.
PARQUET : Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
dataFormat - The format of the data which you want to use for output. You can choose one of the following:
CSV : This is a row-based format with comma-separated values.
PARQUET : Apache Parquet is a columnar storage format that features efficient compression and
provides faster query response.
DataFormatValuepublic void setEncodingType(String encodingType)
The type of encoding you are using: RLE_DICTIONARY (default), PLAIN, or
PLAIN_DICTIONARY.
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated values
more efficiently.
PLAIN does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The dictionary is
stored in a dictionary page for each column chunk.
encodingType - The type of encoding you are using: RLE_DICTIONARY (default), PLAIN, or
PLAIN_DICTIONARY.
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated
values more efficiently.
PLAIN does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The
dictionary is stored in a dictionary page for each column chunk.
EncodingTypeValuepublic String getEncodingType()
The type of encoding you are using: RLE_DICTIONARY (default), PLAIN, or
PLAIN_DICTIONARY.
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated values
more efficiently.
PLAIN does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The dictionary is
stored in a dictionary page for each column chunk.
RLE_DICTIONARY (default), PLAIN, or
PLAIN_DICTIONARY.
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated
values more efficiently.
PLAIN does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The
dictionary is stored in a dictionary page for each column chunk.
EncodingTypeValuepublic S3Settings withEncodingType(String encodingType)
The type of encoding you are using: RLE_DICTIONARY (default), PLAIN, or
PLAIN_DICTIONARY.
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated values
more efficiently.
PLAIN does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The dictionary is
stored in a dictionary page for each column chunk.
encodingType - The type of encoding you are using: RLE_DICTIONARY (default), PLAIN, or
PLAIN_DICTIONARY.
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated
values more efficiently.
PLAIN does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The
dictionary is stored in a dictionary page for each column chunk.
EncodingTypeValuepublic void setEncodingType(EncodingTypeValue encodingType)
The type of encoding you are using: RLE_DICTIONARY (default), PLAIN, or
PLAIN_DICTIONARY.
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated values
more efficiently.
PLAIN does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The dictionary is
stored in a dictionary page for each column chunk.
encodingType - The type of encoding you are using: RLE_DICTIONARY (default), PLAIN, or
PLAIN_DICTIONARY.
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated
values more efficiently.
PLAIN does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The
dictionary is stored in a dictionary page for each column chunk.
EncodingTypeValuepublic S3Settings withEncodingType(EncodingTypeValue encodingType)
The type of encoding you are using: RLE_DICTIONARY (default), PLAIN, or
PLAIN_DICTIONARY.
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated values
more efficiently.
PLAIN does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The dictionary is
stored in a dictionary page for each column chunk.
encodingType - The type of encoding you are using: RLE_DICTIONARY (default), PLAIN, or
PLAIN_DICTIONARY.
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated
values more efficiently.
PLAIN does not use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The
dictionary is stored in a dictionary page for each column chunk.
EncodingTypeValuepublic void setDictPageSizeLimit(Integer dictPageSizeLimit)
The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this column is
stored using an encoding type of PLAIN. Defaults to 1024 * 1024 bytes (1MiB), the maximum size of a
dictionary page before it reverts to PLAIN encoding. For PARQUET format only.
dictPageSizeLimit - The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this
column is stored using an encoding type of PLAIN. Defaults to 1024 * 1024 bytes (1MiB), the
maximum size of a dictionary page before it reverts to PLAIN encoding. For
PARQUET format only.public Integer getDictPageSizeLimit()
The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this column is
stored using an encoding type of PLAIN. Defaults to 1024 * 1024 bytes (1MiB), the maximum size of a
dictionary page before it reverts to PLAIN encoding. For PARQUET format only.
PLAIN. Defaults to 1024 * 1024 bytes (1MiB), the
maximum size of a dictionary page before it reverts to PLAIN encoding. For
PARQUET format only.public S3Settings withDictPageSizeLimit(Integer dictPageSizeLimit)
The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this column is
stored using an encoding type of PLAIN. Defaults to 1024 * 1024 bytes (1MiB), the maximum size of a
dictionary page before it reverts to PLAIN encoding. For PARQUET format only.
dictPageSizeLimit - The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this
column is stored using an encoding type of PLAIN. Defaults to 1024 * 1024 bytes (1MiB), the
maximum size of a dictionary page before it reverts to PLAIN encoding. For
PARQUET format only.public void setRowGroupLength(Integer rowGroupLength)
The number of rows in a row group. A smaller row group size provides faster reads. But as the number of row
groups grows, the slower writes become. Defaults to 10,000 (ten thousand) rows. For PARQUET format
only.
If you choose a value larger than the maximum, RowGroupLength is set to the max row group length in
bytes (64 * 1024 * 1024).
rowGroupLength - The number of rows in a row group. A smaller row group size provides faster reads. But as the number of
row groups grows, the slower writes become. Defaults to 10,000 (ten thousand) rows. For
PARQUET format only.
If you choose a value larger than the maximum, RowGroupLength is set to the max row group
length in bytes (64 * 1024 * 1024).
public Integer getRowGroupLength()
The number of rows in a row group. A smaller row group size provides faster reads. But as the number of row
groups grows, the slower writes become. Defaults to 10,000 (ten thousand) rows. For PARQUET format
only.
If you choose a value larger than the maximum, RowGroupLength is set to the max row group length in
bytes (64 * 1024 * 1024).
PARQUET format only.
If you choose a value larger than the maximum, RowGroupLength is set to the max row group
length in bytes (64 * 1024 * 1024).
public S3Settings withRowGroupLength(Integer rowGroupLength)
The number of rows in a row group. A smaller row group size provides faster reads. But as the number of row
groups grows, the slower writes become. Defaults to 10,000 (ten thousand) rows. For PARQUET format
only.
If you choose a value larger than the maximum, RowGroupLength is set to the max row group length in
bytes (64 * 1024 * 1024).
rowGroupLength - The number of rows in a row group. A smaller row group size provides faster reads. But as the number of
row groups grows, the slower writes become. Defaults to 10,000 (ten thousand) rows. For
PARQUET format only.
If you choose a value larger than the maximum, RowGroupLength is set to the max row group
length in bytes (64 * 1024 * 1024).
public void setDataPageSize(Integer dataPageSize)
The size of one data page in bytes. Defaults to 1024 * 1024 bytes (1MiB). For PARQUET format only.
dataPageSize - The size of one data page in bytes. Defaults to 1024 * 1024 bytes (1MiB). For PARQUET format
only.public Integer getDataPageSize()
The size of one data page in bytes. Defaults to 1024 * 1024 bytes (1MiB). For PARQUET format only.
PARQUET format
only.public S3Settings withDataPageSize(Integer dataPageSize)
The size of one data page in bytes. Defaults to 1024 * 1024 bytes (1MiB). For PARQUET format only.
dataPageSize - The size of one data page in bytes. Defaults to 1024 * 1024 bytes (1MiB). For PARQUET format
only.public void setParquetVersion(String parquetVersion)
The version of Apache Parquet format you want to use: PARQUET_1_0 (default) or
PARQUET_2_0.
parquetVersion - The version of Apache Parquet format you want to use: PARQUET_1_0 (default) or
PARQUET_2_0.ParquetVersionValuepublic String getParquetVersion()
The version of Apache Parquet format you want to use: PARQUET_1_0 (default) or
PARQUET_2_0.
PARQUET_1_0 (default) or
PARQUET_2_0.ParquetVersionValuepublic S3Settings withParquetVersion(String parquetVersion)
The version of Apache Parquet format you want to use: PARQUET_1_0 (default) or
PARQUET_2_0.
parquetVersion - The version of Apache Parquet format you want to use: PARQUET_1_0 (default) or
PARQUET_2_0.ParquetVersionValuepublic void setParquetVersion(ParquetVersionValue parquetVersion)
The version of Apache Parquet format you want to use: PARQUET_1_0 (default) or
PARQUET_2_0.
parquetVersion - The version of Apache Parquet format you want to use: PARQUET_1_0 (default) or
PARQUET_2_0.ParquetVersionValuepublic S3Settings withParquetVersion(ParquetVersionValue parquetVersion)
The version of Apache Parquet format you want to use: PARQUET_1_0 (default) or
PARQUET_2_0.
parquetVersion - The version of Apache Parquet format you want to use: PARQUET_1_0 (default) or
PARQUET_2_0.ParquetVersionValuepublic void setEnableStatistics(Boolean enableStatistics)
Enables statistics for Parquet pages and rowGroups. Choose TRUE to enable statistics, choose
FALSE to disable. Statistics include NULL, DISTINCT, MAX, and
MIN values. Defaults to TRUE. For PARQUET format only.
enableStatistics - Enables statistics for Parquet pages and rowGroups. Choose TRUE to enable statistics, choose
FALSE to disable. Statistics include NULL, DISTINCT,
MAX, and MIN values. Defaults to TRUE. For PARQUET
format only.public Boolean getEnableStatistics()
Enables statistics for Parquet pages and rowGroups. Choose TRUE to enable statistics, choose
FALSE to disable. Statistics include NULL, DISTINCT, MAX, and
MIN values. Defaults to TRUE. For PARQUET format only.
TRUE to enable statistics, choose
FALSE to disable. Statistics include NULL, DISTINCT,
MAX, and MIN values. Defaults to TRUE. For PARQUET
format only.public S3Settings withEnableStatistics(Boolean enableStatistics)
Enables statistics for Parquet pages and rowGroups. Choose TRUE to enable statistics, choose
FALSE to disable. Statistics include NULL, DISTINCT, MAX, and
MIN values. Defaults to TRUE. For PARQUET format only.
enableStatistics - Enables statistics for Parquet pages and rowGroups. Choose TRUE to enable statistics, choose
FALSE to disable. Statistics include NULL, DISTINCT,
MAX, and MIN values. Defaults to TRUE. For PARQUET
format only.public Boolean isEnableStatistics()
Enables statistics for Parquet pages and rowGroups. Choose TRUE to enable statistics, choose
FALSE to disable. Statistics include NULL, DISTINCT, MAX, and
MIN values. Defaults to TRUE. For PARQUET format only.
TRUE to enable statistics, choose
FALSE to disable. Statistics include NULL, DISTINCT,
MAX, and MIN values. Defaults to TRUE. For PARQUET
format only.public void setCdcInsertsOnly(Boolean cdcInsertsOnly)
Option to write only INSERT operations to the comma-separated value (CSV) output files. By default,
the first field in a CSV record contains the letter I (insert), U (update) or
D (delete) to indicate whether the row was inserted, updated, or deleted at the source database. If
cdcInsertsOnly is set to true, then only INSERTs are recorded in the CSV file, without
the I annotation on each line. Valid values are TRUE and FALSE.
cdcInsertsOnly - Option to write only INSERT operations to the comma-separated value (CSV) output files. By
default, the first field in a CSV record contains the letter I (insert), U
(update) or D (delete) to indicate whether the row was inserted, updated, or deleted at the
source database. If cdcInsertsOnly is set to true, then only INSERTs are
recorded in the CSV file, without the I annotation on each line. Valid values are
TRUE and FALSE.public Boolean getCdcInsertsOnly()
Option to write only INSERT operations to the comma-separated value (CSV) output files. By default,
the first field in a CSV record contains the letter I (insert), U (update) or
D (delete) to indicate whether the row was inserted, updated, or deleted at the source database. If
cdcInsertsOnly is set to true, then only INSERTs are recorded in the CSV file, without
the I annotation on each line. Valid values are TRUE and FALSE.
INSERT operations to the comma-separated value (CSV) output files. By
default, the first field in a CSV record contains the letter I (insert), U
(update) or D (delete) to indicate whether the row was inserted, updated, or deleted at the
source database. If cdcInsertsOnly is set to true, then only INSERTs are
recorded in the CSV file, without the I annotation on each line. Valid values are
TRUE and FALSE.public S3Settings withCdcInsertsOnly(Boolean cdcInsertsOnly)
Option to write only INSERT operations to the comma-separated value (CSV) output files. By default,
the first field in a CSV record contains the letter I (insert), U (update) or
D (delete) to indicate whether the row was inserted, updated, or deleted at the source database. If
cdcInsertsOnly is set to true, then only INSERTs are recorded in the CSV file, without
the I annotation on each line. Valid values are TRUE and FALSE.
cdcInsertsOnly - Option to write only INSERT operations to the comma-separated value (CSV) output files. By
default, the first field in a CSV record contains the letter I (insert), U
(update) or D (delete) to indicate whether the row was inserted, updated, or deleted at the
source database. If cdcInsertsOnly is set to true, then only INSERTs are
recorded in the CSV file, without the I annotation on each line. Valid values are
TRUE and FALSE.public Boolean isCdcInsertsOnly()
Option to write only INSERT operations to the comma-separated value (CSV) output files. By default,
the first field in a CSV record contains the letter I (insert), U (update) or
D (delete) to indicate whether the row was inserted, updated, or deleted at the source database. If
cdcInsertsOnly is set to true, then only INSERTs are recorded in the CSV file, without
the I annotation on each line. Valid values are TRUE and FALSE.
INSERT operations to the comma-separated value (CSV) output files. By
default, the first field in a CSV record contains the letter I (insert), U
(update) or D (delete) to indicate whether the row was inserted, updated, or deleted at the
source database. If cdcInsertsOnly is set to true, then only INSERTs are
recorded in the CSV file, without the I annotation on each line. Valid values are
TRUE and FALSE.public String toString()
toString in class ObjectObject.toString()public S3Settings clone()
public void marshall(ProtocolMarshaller protocolMarshaller)
marshall in interface StructuredPojoCopyright © 2019. All rights reserved.