public class ParquetReader extends Object implements Closeable
| Modifier and Type | Field and Description |
|---|---|
protected org.apache.parquet.hadoop.metadata.BlockMetaData |
currentBlockMetadata |
protected ParquetDataSource |
dataSource |
protected ColumnReader[] |
verificationColumnReaders |
| Constructor and Description |
|---|
ParquetReader(org.apache.parquet.io.MessageColumnIO messageColumnIO,
List<org.apache.parquet.hadoop.metadata.BlockMetaData> blocks,
Optional<List<Long>> firstRowsOfBlocks,
ParquetDataSource dataSource,
AggregatedMemoryContext systemMemoryContext,
io.airlift.units.DataSize maxReadBlockSize,
boolean batchReadEnabled,
boolean enableVerification,
Predicate parquetPredicate,
List<org.apache.parquet.internal.filter2.columnindex.ColumnIndexStore> blockIndexStores,
boolean columnIndexFilterEnabled,
Optional<org.apache.parquet.crypto.InternalFileDecryptor> fileDecryptor) |
| Modifier and Type | Method and Description |
|---|---|
protected byte[] |
allocateBlock(long length) |
void |
close() |
protected PageReader |
createPageReader(byte[] buffer,
int bufferSize,
org.apache.parquet.hadoop.metadata.ColumnChunkMetaData metadata,
org.apache.parquet.column.ColumnDescriptor columnDescriptor) |
protected PageReader |
createPageReader(List<ByteBuffer> buffers,
int bufferSize,
org.apache.parquet.hadoop.metadata.ColumnChunkMetaData metadata,
org.apache.parquet.column.ColumnDescriptor columnDescriptor,
org.apache.parquet.internal.column.columnindex.OffsetIndex offsetIndex) |
ParquetDataSource |
getDataSource() |
long |
getPosition() |
AggregatedMemoryContext |
getSystemMemoryContext() |
long |
lastBatchStartRow()
Get the global row index of the first row in the last batch.
|
int |
nextBatch() |
Block |
readBlock(Field field) |
protected final ColumnReader[] verificationColumnReaders
protected final ParquetDataSource dataSource
protected org.apache.parquet.hadoop.metadata.BlockMetaData currentBlockMetadata
public ParquetReader(org.apache.parquet.io.MessageColumnIO messageColumnIO,
List<org.apache.parquet.hadoop.metadata.BlockMetaData> blocks,
Optional<List<Long>> firstRowsOfBlocks,
ParquetDataSource dataSource,
AggregatedMemoryContext systemMemoryContext,
io.airlift.units.DataSize maxReadBlockSize,
boolean batchReadEnabled,
boolean enableVerification,
Predicate parquetPredicate,
List<org.apache.parquet.internal.filter2.columnindex.ColumnIndexStore> blockIndexStores,
boolean columnIndexFilterEnabled,
Optional<org.apache.parquet.crypto.InternalFileDecryptor> fileDecryptor)
public void close()
throws IOException
close in interface Closeableclose in interface AutoCloseableIOExceptionpublic long getPosition()
public long lastBatchStartRow()
public int nextBatch()
protected PageReader createPageReader(List<ByteBuffer> buffers, int bufferSize, org.apache.parquet.hadoop.metadata.ColumnChunkMetaData metadata, org.apache.parquet.column.ColumnDescriptor columnDescriptor, org.apache.parquet.internal.column.columnindex.OffsetIndex offsetIndex) throws IOException
IOExceptionprotected PageReader createPageReader(byte[] buffer, int bufferSize, org.apache.parquet.hadoop.metadata.ColumnChunkMetaData metadata, org.apache.parquet.column.ColumnDescriptor columnDescriptor) throws IOException
IOExceptionprotected byte[] allocateBlock(long length)
public Block readBlock(Field field) throws IOException
IOExceptionpublic ParquetDataSource getDataSource()
public AggregatedMemoryContext getSystemMemoryContext()
Copyright © 2012–2022. All rights reserved.