public abstract class ParquetVectorizedInputFormat<T,SplitT extends org.apache.flink.connector.file.src.FileSourceSplit> extends Object implements org.apache.flink.connector.file.src.reader.BulkFormat<T,SplitT>
BulkFormat that reads data from the file to VectorizedColumnBatch in
vectorized mode.| Modifier and Type | Class and Description |
|---|---|
protected static class |
ParquetVectorizedInputFormat.ParquetReaderBatch<T>
Reader batch that provides writing and reading capabilities.
|
| Constructor and Description |
|---|
ParquetVectorizedInputFormat(SerializableConfiguration hadoopConfig,
org.apache.flink.table.types.logical.RowType projectedType,
ColumnBatchFactory<SplitT> batchFactory,
int batchSize,
boolean isUtcTimestamp,
boolean isCaseSensitive) |
| Modifier and Type | Method and Description |
|---|---|
org.apache.flink.formats.parquet.ParquetVectorizedInputFormat.ParquetReader |
createReader(org.apache.flink.configuration.Configuration config,
SplitT split) |
protected abstract ParquetVectorizedInputFormat.ParquetReaderBatch<T> |
createReaderBatch(org.apache.flink.table.data.vector.writable.WritableColumnVector[] writableVectors,
org.apache.flink.table.data.vector.VectorizedColumnBatch columnarBatch,
org.apache.flink.connector.file.src.util.Pool.Recycler<ParquetVectorizedInputFormat.ParquetReaderBatch<T>> recycler) |
boolean |
isSplittable() |
protected int |
numBatchesToCirculate(org.apache.flink.configuration.Configuration config) |
org.apache.flink.formats.parquet.ParquetVectorizedInputFormat.ParquetReader |
restoreReader(org.apache.flink.configuration.Configuration config,
SplitT split) |
public ParquetVectorizedInputFormat(SerializableConfiguration hadoopConfig, org.apache.flink.table.types.logical.RowType projectedType, ColumnBatchFactory<SplitT> batchFactory, int batchSize, boolean isUtcTimestamp, boolean isCaseSensitive)
public org.apache.flink.formats.parquet.ParquetVectorizedInputFormat.ParquetReader createReader(org.apache.flink.configuration.Configuration config,
SplitT split)
throws IOException
createReader in interface org.apache.flink.connector.file.src.reader.BulkFormat<T,SplitT extends org.apache.flink.connector.file.src.FileSourceSplit>IOExceptionprotected int numBatchesToCirculate(org.apache.flink.configuration.Configuration config)
public org.apache.flink.formats.parquet.ParquetVectorizedInputFormat.ParquetReader restoreReader(org.apache.flink.configuration.Configuration config,
SplitT split)
throws IOException
restoreReader in interface org.apache.flink.connector.file.src.reader.BulkFormat<T,SplitT extends org.apache.flink.connector.file.src.FileSourceSplit>IOExceptionpublic boolean isSplittable()
protected abstract ParquetVectorizedInputFormat.ParquetReaderBatch<T> createReaderBatch(org.apache.flink.table.data.vector.writable.WritableColumnVector[] writableVectors, org.apache.flink.table.data.vector.VectorizedColumnBatch columnarBatch, org.apache.flink.connector.file.src.util.Pool.Recycler<ParquetVectorizedInputFormat.ParquetReaderBatch<T>> recycler)
writableVectors - vectors to be writecolumnarBatch - vectors to be readrecycler - batch recyclerCopyright © 2014–2022 The Apache Software Foundation. All rights reserved.