Struct aws_sdk_firehose::model::ParquetSerDe
source ·
[−]#[non_exhaustive]pub struct ParquetSerDe {
pub block_size_bytes: Option<i32>,
pub page_size_bytes: Option<i32>,
pub compression: Option<ParquetCompression>,
pub enable_dictionary_compression: Option<bool>,
pub max_padding_bytes: Option<i32>,
pub writer_version: Option<ParquetWriterVersion>,
}
Expand description
A serializer to use for converting data to the Parquet format before storing it in Amazon S3. For more information, see Apache Parquet.
Fields (Non-exhaustive)
This struct is marked as non-exhaustive
Struct { .. }
syntax; cannot be matched against without a wildcard ..
; and struct update syntax will not work.block_size_bytes: Option<i32>
The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Kinesis Data Firehose uses this value for padding calculations.
page_size_bytes: Option<i32>
The Parquet page size. Column chunks are divided into pages. A page is conceptually an indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and the default is 1 MiB.
compression: Option<ParquetCompression>
The compression code to use over data blocks. The possible values are UNCOMPRESSED
, SNAPPY
, and GZIP
, with the default being SNAPPY
. Use SNAPPY
for higher decompression speed. Use GZIP
if the compression ratio is more important than speed.
enable_dictionary_compression: Option<bool>
Indicates whether to enable dictionary compression.
max_padding_bytes: Option<i32>
The maximum amount of padding to apply. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 0.
writer_version: Option<ParquetWriterVersion>
Indicates the version of row format to output. The possible values are V1
and V2
. The default is V1
.
Implementations
The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Kinesis Data Firehose uses this value for padding calculations.
The Parquet page size. Column chunks are divided into pages. A page is conceptually an indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and the default is 1 MiB.
The compression code to use over data blocks. The possible values are UNCOMPRESSED
, SNAPPY
, and GZIP
, with the default being SNAPPY
. Use SNAPPY
for higher decompression speed. Use GZIP
if the compression ratio is more important than speed.
Indicates whether to enable dictionary compression.
The maximum amount of padding to apply. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 0.
Indicates the version of row format to output. The possible values are V1
and V2
. The default is V1
.
Creates a new builder-style object to manufacture ParquetSerDe
Trait Implementations
This method tests for self
and other
values to be equal, and is used
by ==
. Read more
This method tests for !=
.
Auto Trait Implementations
impl RefUnwindSafe for ParquetSerDe
impl Send for ParquetSerDe
impl Sync for ParquetSerDe
impl Unpin for ParquetSerDe
impl UnwindSafe for ParquetSerDe
Blanket Implementations
Mutably borrows from an owned value. Read more
Attaches the provided Subscriber
to this type, returning a
WithDispatch
wrapper. Read more
Attaches the current default Subscriber
to this type, returning a
WithDispatch
wrapper. Read more