public static class BtJob.BtMapper extends org.apache.hadoop.mapreduce.Mapper<org.apache.hadoop.io.Writable,VectorWritable,org.apache.hadoop.io.LongWritable,SparseRowBlockWritable>
| Constructor and Description |
|---|
BtMapper() |
| Modifier and Type | Method and Description |
|---|---|
protected void |
cleanup(org.apache.hadoop.mapreduce.Mapper.Context context) |
protected void |
map(org.apache.hadoop.io.Writable key,
VectorWritable value,
org.apache.hadoop.mapreduce.Mapper.Context context)
We maintain A and QtHat inputs partitioned the same way, so we
essentially are performing map-side merge here of A and QtHats except
QtHat is stored not row-wise but block-wise.
|
protected void |
setup(org.apache.hadoop.mapreduce.Mapper.Context context) |
protected void map(org.apache.hadoop.io.Writable key,
VectorWritable value,
org.apache.hadoop.mapreduce.Mapper.Context context)
throws IOException,
InterruptedException
map in class org.apache.hadoop.mapreduce.Mapper<org.apache.hadoop.io.Writable,VectorWritable,org.apache.hadoop.io.LongWritable,SparseRowBlockWritable>IOExceptionInterruptedExceptionprotected void setup(org.apache.hadoop.mapreduce.Mapper.Context context)
throws IOException,
InterruptedException
setup in class org.apache.hadoop.mapreduce.Mapper<org.apache.hadoop.io.Writable,VectorWritable,org.apache.hadoop.io.LongWritable,SparseRowBlockWritable>IOExceptionInterruptedExceptionprotected void cleanup(org.apache.hadoop.mapreduce.Mapper.Context context)
throws IOException,
InterruptedException
cleanup in class org.apache.hadoop.mapreduce.Mapper<org.apache.hadoop.io.Writable,VectorWritable,org.apache.hadoop.io.LongWritable,SparseRowBlockWritable>IOExceptionInterruptedExceptionCopyright © 2008–2017 The Apache Software Foundation. All rights reserved.