public class HoodieMergeOnReadTableInputFormat extends HoodieCopyOnWriteTableInputFormat implements org.apache.hadoop.conf.Configurable
FileInputFormat allowing for reading of Hudi's
Merge-on-Read (COW) tables in various configurations:
NOTE: This class is invariant of the underlying file-format of the files being read
conf| Constructor and Description |
|---|
HoodieMergeOnReadTableInputFormat() |
| Modifier and Type | Method and Description |
|---|---|
protected boolean |
checkIfValidFileSlice(FileSlice fileSlice) |
protected org.apache.hadoop.fs.FileStatus |
createFileStatusUnchecked(FileSlice fileSlice,
Option<HoodieInstant> latestCompletedInstantOpt,
String tableBasePath,
Option<HoodieVirtualKeyInfo> virtualKeyInfoOpt) |
org.apache.hadoop.mapred.InputSplit[] |
getSplits(org.apache.hadoop.mapred.JobConf job,
int numSplits) |
protected boolean |
isSplitable(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path filename) |
protected List<org.apache.hadoop.fs.FileStatus> |
listStatusForIncrementalMode(org.apache.hadoop.mapred.JobConf job,
HoodieTableMetaClient tableMetaClient,
List<org.apache.hadoop.fs.Path> inputPaths,
String incrementalTableName)
Keep the logic of mor_incr_view as same as spark datasource.
|
protected org.apache.hadoop.mapred.FileSplit |
makeSplit(org.apache.hadoop.fs.Path file,
long start,
long length,
String[] hosts) |
protected org.apache.hadoop.mapred.FileSplit |
makeSplit(org.apache.hadoop.fs.Path file,
long start,
long length,
String[] hosts,
String[] inMemoryHosts) |
doListStatus, getFileStatusUnchecked, getHoodieVirtualKeyInfo, getRecordReader, listStatusgetConf, setConfaddInputPath, addInputPathRecursively, addInputPaths, computeSplitSize, getBlockIndex, getInputPathFilter, getInputPaths, getSplitHosts, setInputPathFilter, setInputPaths, setInputPaths, setMinSplitSizepublic org.apache.hadoop.mapred.InputSplit[] getSplits(org.apache.hadoop.mapred.JobConf job,
int numSplits)
throws IOException
getSplits in interface org.apache.hadoop.mapred.InputFormat<org.apache.hadoop.io.NullWritable,org.apache.hadoop.io.ArrayWritable>getSplits in class org.apache.hadoop.mapred.FileInputFormat<org.apache.hadoop.io.NullWritable,org.apache.hadoop.io.ArrayWritable>IOExceptionprotected org.apache.hadoop.fs.FileStatus createFileStatusUnchecked(FileSlice fileSlice, Option<HoodieInstant> latestCompletedInstantOpt, String tableBasePath, Option<HoodieVirtualKeyInfo> virtualKeyInfoOpt)
createFileStatusUnchecked in class HoodieCopyOnWriteTableInputFormatprotected boolean checkIfValidFileSlice(FileSlice fileSlice)
checkIfValidFileSlice in class HoodieCopyOnWriteTableInputFormatprotected List<org.apache.hadoop.fs.FileStatus> listStatusForIncrementalMode(org.apache.hadoop.mapred.JobConf job, HoodieTableMetaClient tableMetaClient, List<org.apache.hadoop.fs.Path> inputPaths, String incrementalTableName) throws IOException
listStatusForIncrementalMode in class HoodieCopyOnWriteTableInputFormatIOExceptionprotected boolean isSplitable(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path filename)
isSplitable in class HoodieCopyOnWriteTableInputFormatprotected org.apache.hadoop.mapred.FileSplit makeSplit(org.apache.hadoop.fs.Path file,
long start,
long length,
String[] hosts)
makeSplit in class HoodieCopyOnWriteTableInputFormatprotected org.apache.hadoop.mapred.FileSplit makeSplit(org.apache.hadoop.fs.Path file,
long start,
long length,
String[] hosts,
String[] inMemoryHosts)
makeSplit in class HoodieCopyOnWriteTableInputFormatCopyright © 2022 The Apache Software Foundation. All rights reserved.