public abstract class HoodieParquetInputFormatBase
extends org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat
implements org.apache.hadoop.conf.Configurable
FileInputFormat
implementation inherits from MapredParquetInputFormat.
To make sure that Hudi implementations are leveraging these optimizations to the fullest, this class
serves as a base-class for every FileInputFormat implementations working with Parquet file-format.
However, this class serves as a simple delegate to the actual implementation hierarchy: it expects
either HoodieCopyOnWriteTableInputFormat or HoodieMergeOnReadTableInputFormat to be supplied
to which it delegates all of its necessary methods.| Modifier | Constructor and Description |
|---|---|
protected |
HoodieParquetInputFormatBase(HoodieCopyOnWriteTableInputFormat inputFormatDelegate) |
| Modifier and Type | Method and Description |
|---|---|
org.apache.hadoop.conf.Configuration |
getConf() |
org.apache.hadoop.mapred.InputSplit[] |
getSplits(org.apache.hadoop.mapred.JobConf job,
int numSplits) |
protected boolean |
isSplitable(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path filename) |
org.apache.hadoop.fs.FileStatus[] |
listStatus(org.apache.hadoop.mapred.JobConf job) |
protected org.apache.hadoop.mapred.FileSplit |
makeSplit(org.apache.hadoop.fs.Path file,
long start,
long length,
String[] hosts) |
protected org.apache.hadoop.mapred.FileSplit |
makeSplit(org.apache.hadoop.fs.Path file,
long start,
long length,
String[] hosts,
String[] inMemoryHosts) |
void |
setConf(org.apache.hadoop.conf.Configuration conf) |
getRecordReaderprotected HoodieParquetInputFormatBase(HoodieCopyOnWriteTableInputFormat inputFormatDelegate)
public final void setConf(org.apache.hadoop.conf.Configuration conf)
setConf in interface org.apache.hadoop.conf.Configurablepublic final org.apache.hadoop.conf.Configuration getConf()
getConf in interface org.apache.hadoop.conf.Configurablepublic final org.apache.hadoop.mapred.InputSplit[] getSplits(org.apache.hadoop.mapred.JobConf job,
int numSplits)
throws IOException
getSplits in interface org.apache.hadoop.mapred.InputFormat<org.apache.hadoop.io.NullWritable,org.apache.hadoop.io.ArrayWritable>getSplits in class org.apache.hadoop.mapred.FileInputFormat<org.apache.hadoop.io.NullWritable,org.apache.hadoop.io.ArrayWritable>IOExceptionprotected final boolean isSplitable(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path filename)
isSplitable in class org.apache.hadoop.mapred.FileInputFormat<org.apache.hadoop.io.NullWritable,org.apache.hadoop.io.ArrayWritable>protected final org.apache.hadoop.mapred.FileSplit makeSplit(org.apache.hadoop.fs.Path file,
long start,
long length,
String[] hosts)
makeSplit in class org.apache.hadoop.mapred.FileInputFormat<org.apache.hadoop.io.NullWritable,org.apache.hadoop.io.ArrayWritable>protected final org.apache.hadoop.mapred.FileSplit makeSplit(org.apache.hadoop.fs.Path file,
long start,
long length,
String[] hosts,
String[] inMemoryHosts)
makeSplit in class org.apache.hadoop.mapred.FileInputFormat<org.apache.hadoop.io.NullWritable,org.apache.hadoop.io.ArrayWritable>public final org.apache.hadoop.fs.FileStatus[] listStatus(org.apache.hadoop.mapred.JobConf job)
throws IOException
listStatus in class org.apache.hadoop.mapred.FileInputFormat<org.apache.hadoop.io.NullWritable,org.apache.hadoop.io.ArrayWritable>IOExceptionCopyright © 2022 The Apache Software Foundation. All rights reserved.