R - target type of the object being written into Parquet files (for ex,
IndexedRecord, InternalRow)public abstract class HoodieBaseParquetWriter<R> extends Object implements Closeable
ParquetWriter implementations| Modifier and Type | Field and Description |
|---|---|
static String |
BLOOM_FILTER_ENABLED |
static String |
BLOOM_FILTER_EXPECTED_NDV |
| Constructor and Description |
|---|
HoodieBaseParquetWriter(StoragePath file,
HoodieParquetConfig<? extends org.apache.parquet.hadoop.api.WriteSupport<R>> parquetConfig) |
| Modifier and Type | Method and Description |
|---|---|
boolean |
canWrite() |
void |
close() |
long |
getDataSize() |
protected long |
getRecordCountForNextSizeCheck() |
protected long |
getWrittenRecordCount() |
protected void |
handleParquetBloomFilters(org.apache.parquet.hadoop.ParquetWriter.Builder parquetWriterbuilder,
org.apache.hadoop.conf.Configuration hadoopConf)
Once we get parquet version >= 1.12 among all engines we can cleanup the reflexion hack.
|
void |
write(R object) |
public static final String BLOOM_FILTER_EXPECTED_NDV
public static final String BLOOM_FILTER_ENABLED
public HoodieBaseParquetWriter(StoragePath file, HoodieParquetConfig<? extends org.apache.parquet.hadoop.api.WriteSupport<R>> parquetConfig) throws IOException
IOExceptionprotected void handleParquetBloomFilters(org.apache.parquet.hadoop.ParquetWriter.Builder parquetWriterbuilder,
org.apache.hadoop.conf.Configuration hadoopConf)
parquetWriterbuilder - hadoopConf - public boolean canWrite()
public long getDataSize()
public void write(R object) throws IOException
IOExceptionprotected long getWrittenRecordCount()
protected long getRecordCountForNextSizeCheck()
public void close()
throws IOException
close in interface Closeableclose in interface AutoCloseableIOExceptionCopyright © 2024 The Apache Software Foundation. All rights reserved.