io_parquet only.Expand description
APIs to write to Parquet format.
Structs
A descriptor for leaf-level primitive columns. This encapsulates information such as definition and repetition levels and is used to re-assemble nested data.
A CompressedDataPage is compressed, encoded representation of a Parquet data page.
It holds actual data and thus cloning it is expensive.
A FallibleStreamingIterator that consumes EncodedPage and yields CompressedPage
holding a reusable buffer (Vec<u8>) for compression.
Dynamically-typed FallibleStreamingIterator.
An interface to write a parquet to a Write
An iterator adapter that converts an iterator over Chunk into an iterator
of row groups.
Use it to create an iterator consumable by the parquet’s API.
A schema descriptor. This encapsulates the top-level schemas for all the columns, as well as all descriptors for all the primitive columns.
Write options of different interfaces on this crate
Enums
A CompressedPage is a compressed, encoded representation of a Parquet page. It holds actual data
and thus cloning it is expensive.
A EncodedPage is an uncompressed, encoded representation of a Parquet page. It may hold actual data
and thus cloning it may be expensive.
Representation of a Parquet type.
Used to describe primitive leaf fields and structs, including top-level schema.
Note that the top-level schema type is represented using GroupType whose
repetition is None.
The parquet version to use
Traits
A fallible, streaming iterator.
Functions
Converts an Array to a CompressedPage based on options, descriptor and encoding.
Returns an iterator of EncodedPage.
Checks whether the data_type can be encoded as encoding.
Note that this is whether this implementation supports it, which is a subset of
what the parquet spec allows.
Maps a Chunk and parquet-specific options to an RowGroupIter used to
write to parquet
Creates a parquet SchemaDescriptor from a Schema.
Creates a ParquetType from a Field.