public class RDFInputFormat extends org.apache.hadoop.mapreduce.lib.input.FileInputFormat<org.apache.hadoop.io.NullWritable,FaunusElement> implements MapReduceFormat
| Constructor and Description |
|---|
RDFInputFormat() |
| Modifier and Type | Method and Description |
|---|---|
void |
addMapReduceJobs(HadoopCompiler compiler) |
org.apache.hadoop.mapreduce.RecordReader<org.apache.hadoop.io.NullWritable,FaunusElement> |
createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context) |
protected boolean |
isSplitable(org.apache.hadoop.mapreduce.JobContext context,
org.apache.hadoop.fs.Path file) |
addInputPath, addInputPathRecursively, addInputPaths, computeSplitSize, getBlockIndex, getFormatMinSplitSize, getInputDirRecursive, getInputPathFilter, getInputPaths, getMaxSplitSize, getMinSplitSize, getSplits, listStatus, makeSplit, setInputDirRecursive, setInputPathFilter, setInputPaths, setInputPaths, setMaxInputSplitSize, setMinInputSplitSizepublic org.apache.hadoop.mapreduce.RecordReader<org.apache.hadoop.io.NullWritable,FaunusElement> createRecordReader(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context) throws IOException
createRecordReader in class org.apache.hadoop.mapreduce.InputFormat<org.apache.hadoop.io.NullWritable,FaunusElement>IOExceptionprotected boolean isSplitable(org.apache.hadoop.mapreduce.JobContext context,
org.apache.hadoop.fs.Path file)
isSplitable in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat<org.apache.hadoop.io.NullWritable,FaunusElement>public void addMapReduceJobs(HadoopCompiler compiler)
addMapReduceJobs in interface MapReduceFormatCopyright © 2012–2016. All rights reserved.