io_parquet
only.Expand description
APIs to write to Parquet format.
Structs
A descriptor for leaf-level primitive columns. This encapsulates information such as definition and repetition levels and is used to re-assemble nested data.
A CompressedDataPage
is compressed, encoded representation of a Parquet data page.
It holds actual data and thus cloning it is expensive.
A FallibleStreamingIterator
that consumes EncodedPage
and yields CompressedPage
holding a reusable buffer (Vec<u8>
) for compression.
Dynamically-typed FallibleStreamingIterator
.
An interface to write a parquet to a Write
An iterator adapter that converts an iterator over Chunk
into an iterator
of row groups.
Use it to create an iterator consumable by the parquet’s API.
A schema descriptor. This encapsulates the top-level schemas for all the columns, as well as all descriptors for all the primitive columns.
Write options of different interfaces on this crate
Enums
A CompressedPage
is a compressed, encoded representation of a Parquet page. It holds actual data
and thus cloning it is expensive.
A EncodedPage
is an uncompressed, encoded representation of a Parquet page. It may hold actual data
and thus cloning it may be expensive.
Representation of a Parquet type.
Used to describe primitive leaf fields and structs, including top-level schema.
Note that the top-level schema type is represented using GroupType
whose
repetition is None
.
The parquet version to use
Traits
A fallible, streaming iterator.
Functions
Converts an Array
to a CompressedPage
based on options, descriptor and encoding
.
Returns an iterator of EncodedPage
.
Checks whether the data_type
can be encoded as encoding
.
Note that this is whether this implementation supports it, which is a subset of
what the parquet spec allows.
Maps a Chunk
and parquet-specific options to an RowGroupIter
used to
write to parquet
Creates a parquet SchemaDescriptor
from a Schema
.
Creates a ParquetType
from a Field
.