Class: Aws::Firehose::Types::ParquetSerDe
- Inherits:
-
Struct
- Object
- Struct
- Aws::Firehose::Types::ParquetSerDe
- Includes:
- Structure
- Defined in:
- lib/aws-sdk-firehose/types.rb
Overview
When making an API call, you may pass ParquetSerDe data as a hash:
{
block_size_bytes: 1,
page_size_bytes: 1,
compression: "UNCOMPRESSED", # accepts UNCOMPRESSED, GZIP, SNAPPY
enable_dictionary_compression: false,
max_padding_bytes: 1,
writer_version: "V1", # accepts V1, V2
}
A serializer to use for converting data to the Parquet format before storing it in Amazon S3. For more information, see [Apache Parquet].
Instance Attribute Summary collapse
-
#block_size_bytes ⇒ Integer
The Hadoop Distributed File System (HDFS) block size.
-
#compression ⇒ String
The compression code to use over data blocks.
-
#enable_dictionary_compression ⇒ Boolean
Indicates whether to enable dictionary compression.
-
#max_padding_bytes ⇒ Integer
The maximum amount of padding to apply.
-
#page_size_bytes ⇒ Integer
The Parquet page size.
-
#writer_version ⇒ String
Indicates the version of row format to output.
Instance Attribute Details
#block_size_bytes ⇒ Integer
The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Kinesis Data Firehose uses this value for padding calculations.
2436 2437 2438 2439 2440 2441 2442 2443 2444 |
# File 'lib/aws-sdk-firehose/types.rb', line 2436 class ParquetSerDe < Struct.new( :block_size_bytes, :page_size_bytes, :compression, :enable_dictionary_compression, :max_padding_bytes, :writer_version) include Aws::Structure end |
#compression ⇒ String
The compression code to use over data blocks. The possible values are ‘UNCOMPRESSED`, `SNAPPY`, and `GZIP`, with the default being `SNAPPY`. Use `SNAPPY` for higher decompression speed. Use `GZIP` if the compression ration is more important than speed.
2436 2437 2438 2439 2440 2441 2442 2443 2444 |
# File 'lib/aws-sdk-firehose/types.rb', line 2436 class ParquetSerDe < Struct.new( :block_size_bytes, :page_size_bytes, :compression, :enable_dictionary_compression, :max_padding_bytes, :writer_version) include Aws::Structure end |
#enable_dictionary_compression ⇒ Boolean
Indicates whether to enable dictionary compression.
2436 2437 2438 2439 2440 2441 2442 2443 2444 |
# File 'lib/aws-sdk-firehose/types.rb', line 2436 class ParquetSerDe < Struct.new( :block_size_bytes, :page_size_bytes, :compression, :enable_dictionary_compression, :max_padding_bytes, :writer_version) include Aws::Structure end |
#max_padding_bytes ⇒ Integer
The maximum amount of padding to apply. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 0.
2436 2437 2438 2439 2440 2441 2442 2443 2444 |
# File 'lib/aws-sdk-firehose/types.rb', line 2436 class ParquetSerDe < Struct.new( :block_size_bytes, :page_size_bytes, :compression, :enable_dictionary_compression, :max_padding_bytes, :writer_version) include Aws::Structure end |
#page_size_bytes ⇒ Integer
The Parquet page size. Column chunks are divided into pages. A page is conceptually an indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and the default is 1 MiB.
2436 2437 2438 2439 2440 2441 2442 2443 2444 |
# File 'lib/aws-sdk-firehose/types.rb', line 2436 class ParquetSerDe < Struct.new( :block_size_bytes, :page_size_bytes, :compression, :enable_dictionary_compression, :max_padding_bytes, :writer_version) include Aws::Structure end |
#writer_version ⇒ String
Indicates the version of row format to output. The possible values are ‘V1` and `V2`. The default is `V1`.
2436 2437 2438 2439 2440 2441 2442 2443 2444 |
# File 'lib/aws-sdk-firehose/types.rb', line 2436 class ParquetSerDe < Struct.new( :block_size_bytes, :page_size_bytes, :compression, :enable_dictionary_compression, :max_padding_bytes, :writer_version) include Aws::Structure end |