T - public abstract class CarbonInputFormat<T> extends org.apache.hadoop.mapreduce.lib.input.FileInputFormat<Void,T>
| Modifier and Type | Field and Description |
|---|---|
static String |
DATABASE_NAME |
protected int |
hitedStreamFiles |
static String |
INPUT_SEGMENT_NUMBERS |
protected int |
numBlocks |
protected int |
numSegments |
protected int |
numStreamFiles |
protected int |
numStreamSegments |
static String |
TABLE_NAME |
| Constructor and Description |
|---|
CarbonInputFormat() |
| Modifier and Type | Method and Description |
|---|---|
org.apache.carbondata.core.scan.model.QueryModel |
createQueryModel(org.apache.hadoop.mapreduce.InputSplit inputSplit,
org.apache.hadoop.mapreduce.TaskAttemptContext taskAttemptContext) |
org.apache.hadoop.mapreduce.RecordReader<Void,T> |
createRecordReader(org.apache.hadoop.mapreduce.InputSplit inputSplit,
org.apache.hadoop.mapreduce.TaskAttemptContext taskAttemptContext) |
org.apache.carbondata.core.metadata.AbsoluteTableIdentifier |
getAbsoluteTableIdentifier(org.apache.hadoop.conf.Configuration configuration) |
static boolean |
getAccessStreamingSegments(org.apache.hadoop.conf.Configuration configuration) |
static String |
getColumnProjection(org.apache.hadoop.conf.Configuration configuration) |
static String |
getDatabaseName(org.apache.hadoop.conf.Configuration configuration) |
protected List<CarbonInputSplit> |
getDataBlocksOfSegment(org.apache.hadoop.mapreduce.JobContext job,
org.apache.carbondata.core.metadata.schema.table.CarbonTable carbonTable,
org.apache.carbondata.core.scan.expression.Expression expression,
BitSet matchedPartitions,
List<org.apache.carbondata.core.datamap.Segment> segmentIds,
org.apache.carbondata.core.metadata.schema.PartitionInfo partitionInfo,
List<Integer> oldPartitionIdList)
get data blocks of given segment
|
static org.apache.carbondata.core.util.DataTypeConverter |
getDataTypeConverter(org.apache.hadoop.conf.Configuration configuration) |
protected org.apache.carbondata.core.scan.expression.Expression |
getFilterPredicates(org.apache.hadoop.conf.Configuration configuration) |
int |
getHitedStreamFiles() |
int |
getNumBlocks() |
int |
getNumSegments() |
int |
getNumStreamFiles() |
int |
getNumStreamSegments() |
abstract org.apache.carbondata.core.metadata.schema.table.CarbonTable |
getOrCreateCarbonTable(org.apache.hadoop.conf.Configuration configuration)
Get the cached CarbonTable or create it by TableInfo in `configuration`
|
static List<org.apache.carbondata.core.indexstore.PartitionSpec> |
getPartitionsToPrune(org.apache.hadoop.conf.Configuration configuration)
get list of partitions to prune
|
static org.apache.carbondata.core.readcommitter.ReadCommittedScope |
getReadCommittedScope(org.apache.hadoop.conf.Configuration configuration) |
CarbonReadSupport<T> |
getReadSupportClass(org.apache.hadoop.conf.Configuration configuration) |
abstract List<org.apache.hadoop.mapreduce.InputSplit> |
getSplits(org.apache.hadoop.mapreduce.JobContext job)
Configurations FileInputFormat.INPUT_DIR
are used to get table path to read.
|
protected static org.apache.carbondata.core.metadata.schema.table.TableInfo |
getTableInfo(org.apache.hadoop.conf.Configuration configuration)
Get TableInfo object from `configuration`
|
static String |
getTableName(org.apache.hadoop.conf.Configuration configuration) |
static boolean |
getValidateSegmentsToAccess(org.apache.hadoop.conf.Configuration configuration)
get list of segment to access
|
static boolean |
isFgDataMapPruningEnable(org.apache.hadoop.conf.Configuration configuration) |
protected boolean |
isSplitable(org.apache.hadoop.mapreduce.JobContext context,
org.apache.hadoop.fs.Path filename) |
String[] |
projectAllColumns(org.apache.carbondata.core.metadata.schema.table.CarbonTable carbonTable)
Project all Columns for carbon reader
|
static void |
setAccessStreamingSegments(org.apache.hadoop.conf.Configuration configuration,
Boolean validate) |
static void |
setCarbonReadSupport(org.apache.hadoop.conf.Configuration configuration,
Class<? extends CarbonReadSupport> readSupportClass) |
static void |
setColumnProjection(org.apache.hadoop.conf.Configuration configuration,
CarbonProjection projection)
Set the column projection column names from CarbonProjection
|
static void |
setColumnProjection(org.apache.hadoop.conf.Configuration configuration,
String[] projectionColumns)
Set the column projection column names
|
static void |
setDatabaseName(org.apache.hadoop.conf.Configuration configuration,
String databaseName) |
static void |
setDataTypeConverter(org.apache.hadoop.conf.Configuration configuration,
Class<? extends org.apache.carbondata.core.util.DataTypeConverter> converterClass)
It is optional, if user does not set then it reads from store
|
static void |
setFgDataMapPruning(org.apache.hadoop.conf.Configuration configuration,
boolean enable) |
static void |
setFilterPredicates(org.apache.hadoop.conf.Configuration configuration,
org.apache.carbondata.core.scan.expression.Expression filterExpression)
It sets unresolved filter expression.
|
static void |
setPartitionIdList(org.apache.hadoop.conf.Configuration configuration,
List<String> partitionIds) |
static void |
setPartitionsToPrune(org.apache.hadoop.conf.Configuration configuration,
List<org.apache.carbondata.core.indexstore.PartitionSpec> partitions)
set list of partitions to prune
|
static void |
setQuerySegment(org.apache.hadoop.conf.Configuration conf,
org.apache.carbondata.core.metadata.AbsoluteTableIdentifier identifier)
Set `CARBON_INPUT_SEGMENTS` from property to configuration
|
static void |
setQuerySegment(org.apache.hadoop.conf.Configuration conf,
String segmentList)
Set `CARBON_INPUT_SEGMENTS` from property to configuration
|
static void |
setReadCommittedScope(org.apache.hadoop.conf.Configuration configuration,
org.apache.carbondata.core.readcommitter.ReadCommittedScope committedScope) |
static void |
setSegmentsToAccess(org.apache.hadoop.conf.Configuration configuration,
List<org.apache.carbondata.core.datamap.Segment> validSegments)
Set list of segments to access
|
static void |
setTableInfo(org.apache.hadoop.conf.Configuration configuration,
org.apache.carbondata.core.metadata.schema.table.TableInfo tableInfo)
Set the `tableInfo` in `configuration`
|
static void |
setTableName(org.apache.hadoop.conf.Configuration configuration,
String tableName) |
static void |
setTablePath(org.apache.hadoop.conf.Configuration configuration,
String tablePath) |
static void |
setTransactionalTable(org.apache.hadoop.conf.Configuration configuration,
boolean isTransactionalTable) |
static void |
setValidateSegmentsToAccess(org.apache.hadoop.conf.Configuration configuration,
Boolean validate)
set list of segment to access
|
addInputPath, addInputPathRecursively, addInputPaths, computeSplitSize, getBlockIndex, getFormatMinSplitSize, getInputDirRecursive, getInputPathFilter, getInputPaths, getMaxSplitSize, getMinSplitSize, listStatus, makeSplit, makeSplit, setInputDirRecursive, setInputPathFilter, setInputPaths, setInputPaths, setMaxInputSplitSize, setMinInputSplitSizepublic static final String INPUT_SEGMENT_NUMBERS
public static final String DATABASE_NAME
public static final String TABLE_NAME
protected int numSegments
protected int numStreamSegments
protected int numStreamFiles
protected int hitedStreamFiles
protected int numBlocks
public int getNumSegments()
public int getNumStreamSegments()
public int getNumStreamFiles()
public int getHitedStreamFiles()
public int getNumBlocks()
public static void setTableInfo(org.apache.hadoop.conf.Configuration configuration,
org.apache.carbondata.core.metadata.schema.table.TableInfo tableInfo)
throws IOException
IOExceptionprotected static org.apache.carbondata.core.metadata.schema.table.TableInfo getTableInfo(org.apache.hadoop.conf.Configuration configuration)
throws IOException
IOExceptionpublic abstract org.apache.carbondata.core.metadata.schema.table.CarbonTable getOrCreateCarbonTable(org.apache.hadoop.conf.Configuration configuration)
throws IOException
IOExceptionpublic static void setTablePath(org.apache.hadoop.conf.Configuration configuration,
String tablePath)
public static void setTransactionalTable(org.apache.hadoop.conf.Configuration configuration,
boolean isTransactionalTable)
public static void setPartitionIdList(org.apache.hadoop.conf.Configuration configuration,
List<String> partitionIds)
public static void setFilterPredicates(org.apache.hadoop.conf.Configuration configuration,
org.apache.carbondata.core.scan.expression.Expression filterExpression)
configuration - public static void setColumnProjection(org.apache.hadoop.conf.Configuration configuration,
String[] projectionColumns)
configuration - Configuration infoprojectionColumns - projection columns namepublic static void setColumnProjection(org.apache.hadoop.conf.Configuration configuration,
CarbonProjection projection)
configuration - Configuration infoprojection - CarbonProjection object that includes unique projection column namepublic static String getColumnProjection(org.apache.hadoop.conf.Configuration configuration)
public static void setFgDataMapPruning(org.apache.hadoop.conf.Configuration configuration,
boolean enable)
public static boolean isFgDataMapPruningEnable(org.apache.hadoop.conf.Configuration configuration)
public static void setSegmentsToAccess(org.apache.hadoop.conf.Configuration configuration,
List<org.apache.carbondata.core.datamap.Segment> validSegments)
public static void setQuerySegment(org.apache.hadoop.conf.Configuration conf,
org.apache.carbondata.core.metadata.AbsoluteTableIdentifier identifier)
public static void setQuerySegment(org.apache.hadoop.conf.Configuration conf,
String segmentList)
public static void setValidateSegmentsToAccess(org.apache.hadoop.conf.Configuration configuration,
Boolean validate)
public static boolean getValidateSegmentsToAccess(org.apache.hadoop.conf.Configuration configuration)
public static void setPartitionsToPrune(org.apache.hadoop.conf.Configuration configuration,
List<org.apache.carbondata.core.indexstore.PartitionSpec> partitions)
public static List<org.apache.carbondata.core.indexstore.PartitionSpec> getPartitionsToPrune(org.apache.hadoop.conf.Configuration configuration) throws IOException
IOExceptionpublic org.apache.carbondata.core.metadata.AbsoluteTableIdentifier getAbsoluteTableIdentifier(org.apache.hadoop.conf.Configuration configuration)
throws IOException
IOExceptionpublic static void setReadCommittedScope(org.apache.hadoop.conf.Configuration configuration,
org.apache.carbondata.core.readcommitter.ReadCommittedScope committedScope)
public static org.apache.carbondata.core.readcommitter.ReadCommittedScope getReadCommittedScope(org.apache.hadoop.conf.Configuration configuration)
throws IOException
IOExceptionpublic abstract List<org.apache.hadoop.mapreduce.InputSplit> getSplits(org.apache.hadoop.mapreduce.JobContext job) throws IOException
getSplits in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat<Void,T>job - IOExceptionprotected org.apache.carbondata.core.scan.expression.Expression getFilterPredicates(org.apache.hadoop.conf.Configuration configuration)
protected List<CarbonInputSplit> getDataBlocksOfSegment(org.apache.hadoop.mapreduce.JobContext job, org.apache.carbondata.core.metadata.schema.table.CarbonTable carbonTable, org.apache.carbondata.core.scan.expression.Expression expression, BitSet matchedPartitions, List<org.apache.carbondata.core.datamap.Segment> segmentIds, org.apache.carbondata.core.metadata.schema.PartitionInfo partitionInfo, List<Integer> oldPartitionIdList) throws IOException
IOExceptionpublic org.apache.hadoop.mapreduce.RecordReader<Void,T> createRecordReader(org.apache.hadoop.mapreduce.InputSplit inputSplit, org.apache.hadoop.mapreduce.TaskAttemptContext taskAttemptContext) throws IOException, InterruptedException
createRecordReader in class org.apache.hadoop.mapreduce.InputFormat<Void,T>IOExceptionInterruptedExceptionpublic org.apache.carbondata.core.scan.model.QueryModel createQueryModel(org.apache.hadoop.mapreduce.InputSplit inputSplit,
org.apache.hadoop.mapreduce.TaskAttemptContext taskAttemptContext)
throws IOException
IOExceptionpublic CarbonReadSupport<T> getReadSupportClass(org.apache.hadoop.conf.Configuration configuration)
protected boolean isSplitable(org.apache.hadoop.mapreduce.JobContext context,
org.apache.hadoop.fs.Path filename)
public static void setCarbonReadSupport(org.apache.hadoop.conf.Configuration configuration,
Class<? extends CarbonReadSupport> readSupportClass)
public static void setDataTypeConverter(org.apache.hadoop.conf.Configuration configuration,
Class<? extends org.apache.carbondata.core.util.DataTypeConverter> converterClass)
configuration - converterClass - is the Data type converter for different computing enginepublic static org.apache.carbondata.core.util.DataTypeConverter getDataTypeConverter(org.apache.hadoop.conf.Configuration configuration)
throws IOException
IOExceptionpublic static void setDatabaseName(org.apache.hadoop.conf.Configuration configuration,
String databaseName)
public static String getDatabaseName(org.apache.hadoop.conf.Configuration configuration) throws org.apache.carbondata.core.exception.InvalidConfigurationException
org.apache.carbondata.core.exception.InvalidConfigurationExceptionpublic static void setTableName(org.apache.hadoop.conf.Configuration configuration,
String tableName)
public static String getTableName(org.apache.hadoop.conf.Configuration configuration) throws org.apache.carbondata.core.exception.InvalidConfigurationException
org.apache.carbondata.core.exception.InvalidConfigurationExceptionpublic static void setAccessStreamingSegments(org.apache.hadoop.conf.Configuration configuration,
Boolean validate)
throws org.apache.carbondata.core.exception.InvalidConfigurationException
org.apache.carbondata.core.exception.InvalidConfigurationExceptionpublic static boolean getAccessStreamingSegments(org.apache.hadoop.conf.Configuration configuration)
public String[] projectAllColumns(org.apache.carbondata.core.metadata.schema.table.CarbonTable carbonTable)
carbonTable - Copyright © 2016–2018 The Apache Software Foundation. All rights reserved.