| Modifier and Type | Method and Description |
|---|---|
static HiveColumnHandle |
HiveColumnHandle.bucketColumnHandle(String connectorId) |
HiveColumnHandle |
HivePageSourceProvider.ColumnMapping.getHiveColumnHandle() |
static HiveColumnHandle |
HiveColumnHandle.pathColumnHandle(String connectorId) |
static HiveColumnHandle |
HiveColumnHandle.updateRowIdHandle(String connectorId) |
| Modifier and Type | Method and Description |
|---|---|
static List<HiveColumnHandle> |
HivePageSourceProvider.ColumnMapping.extractRegularColumnHandles(List<HivePageSourceProvider.ColumnMapping> regularColumnMappings,
boolean doCoercion) |
List<HiveColumnHandle> |
HiveBucketHandle.getColumns() |
TupleDomain<HiveColumnHandle> |
HiveSplit.getEffectivePredicate() |
List<HiveColumnHandle> |
HiveWritableTableHandle.getInputColumns() |
List<HiveColumnHandle> |
HivePartitionResult.getPartitionColumns() |
static List<HiveColumnHandle> |
HiveUtil.getPartitionKeyColumnHandles(String connectorId,
Table table) |
static List<HiveColumnHandle> |
HiveUtil.getRegularColumnHandles(String connectorId,
Table table) |
static List<HiveColumnHandle> |
HiveUtil.hiveColumnHandles(String connectorId,
Table table) |
| Modifier and Type | Method and Description |
|---|---|
static String |
HiveUtil.getPrefilledColumnValue(HiveColumnHandle columnHandle,
HivePartitionKey partitionKey,
org.apache.hadoop.fs.Path path,
OptionalInt bucketNumber) |
static boolean |
HiveColumnHandle.isBucketColumnHandle(HiveColumnHandle column) |
static boolean |
HiveColumnHandle.isPathColumnHandle(HiveColumnHandle column) |
| Modifier and Type | Method and Description |
|---|---|
static List<HivePageSourceProvider.ColumnMapping> |
HivePageSourceProvider.ColumnMapping.buildColumnMappings(List<HivePartitionKey> partitionKeys,
List<HiveColumnHandle> columns,
Map<Integer,HiveType> columnCoercions,
org.apache.hadoop.fs.Path path,
OptionalInt bucketNumber) |
static Optional<ConnectorPageSource> |
HivePageSourceProvider.createHivePageSource(Set<HiveRecordCursorProvider> cursorProviders,
Set<HivePageSourceFactory> pageSourceFactories,
String clientId,
org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
OptionalInt bucketNumber,
long start,
long length,
long fileSize,
Properties schema,
TupleDomain<HiveColumnHandle> effectivePredicate,
List<HiveColumnHandle> hiveColumns,
List<HivePartitionKey> partitionKeys,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager,
Map<Integer,HiveType> columnCoercions) |
static Optional<ConnectorPageSource> |
HivePageSourceProvider.createHivePageSource(Set<HiveRecordCursorProvider> cursorProviders,
Set<HivePageSourceFactory> pageSourceFactories,
String clientId,
org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
OptionalInt bucketNumber,
long start,
long length,
long fileSize,
Properties schema,
TupleDomain<HiveColumnHandle> effectivePredicate,
List<HiveColumnHandle> hiveColumns,
List<HivePartitionKey> partitionKeys,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager,
Map<Integer,HiveType> columnCoercions) |
Optional<? extends ConnectorPageSource> |
HivePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone) |
Optional<? extends ConnectorPageSource> |
HivePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone) |
Optional<RecordCursor> |
HiveRecordCursorProvider.createRecordCursor(String clientId,
org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager) |
Optional<RecordCursor> |
HiveRecordCursorProvider.createRecordCursor(String clientId,
org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager) |
Optional<RecordCursor> |
GenericHiveRecordCursorProvider.createRecordCursor(String clientId,
org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager) |
Optional<RecordCursor> |
GenericHiveRecordCursorProvider.createRecordCursor(String clientId,
org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager) |
static org.apache.hadoop.mapred.RecordReader<?,?> |
HiveUtil.createRecordReader(org.apache.hadoop.conf.Configuration configuration,
org.apache.hadoop.fs.Path path,
long start,
long length,
Properties schema,
List<HiveColumnHandle> columns) |
void |
HiveStorageFormat.validateColumns(List<HiveColumnHandle> handles) |
| Constructor and Description |
|---|
HiveBucketHandle(List<HiveColumnHandle> columns,
int bucketCount) |
HiveInsertTableHandle(String clientId,
String schemaName,
String tableName,
List<HiveColumnHandle> inputColumns,
String filePrefix,
HivePageSinkMetadata pageSinkMetadata,
LocationHandle locationHandle,
Optional<HiveBucketProperty> bucketProperty,
HiveStorageFormat tableStorageFormat,
HiveStorageFormat partitionStorageFormat) |
HiveOutputTableHandle(String clientId,
String schemaName,
String tableName,
List<HiveColumnHandle> inputColumns,
String filePrefix,
HivePageSinkMetadata pageSinkMetadata,
LocationHandle locationHandle,
HiveStorageFormat tableStorageFormat,
HiveStorageFormat partitionStorageFormat,
List<String> partitionedBy,
Optional<HiveBucketProperty> bucketProperty,
String tableOwner,
Map<String,String> additionalTableParameters) |
HivePageSink(HiveWriterFactory writerFactory,
List<HiveColumnHandle> inputColumns,
Optional<HiveBucketProperty> bucketProperty,
PageIndexerFactory pageIndexerFactory,
TypeManager typeManager,
HdfsEnvironment hdfsEnvironment,
int maxOpenWriters,
com.google.common.util.concurrent.ListeningExecutorService writeVerificationExecutor,
io.airlift.json.JsonCodec<PartitionUpdate> partitionUpdateCodec,
ConnectorSession session) |
HivePartitionResult(List<HiveColumnHandle> partitionColumns,
List<HivePartition> partitions,
TupleDomain<? extends ColumnHandle> compactEffectivePredicate,
TupleDomain<ColumnHandle> unenforcedConstraint,
TupleDomain<ColumnHandle> enforcedConstraint,
Optional<HiveBucketHandle> bucketHandle) |
HiveSplit(String clientId,
String database,
String table,
String partitionName,
String path,
long start,
long length,
long fileSize,
Properties schema,
List<HivePartitionKey> partitionKeys,
List<HostAddress> addresses,
OptionalInt bucketNumber,
boolean forceLocalScheduling,
TupleDomain<HiveColumnHandle> effectivePredicate,
Map<Integer,HiveType> columnCoercions) |
HiveWritableTableHandle(String clientId,
String schemaName,
String tableName,
List<HiveColumnHandle> inputColumns,
String filePrefix,
HivePageSinkMetadata pageSinkMetadata,
LocationHandle locationHandle,
Optional<HiveBucketProperty> bucketProperty,
HiveStorageFormat tableStorageFormat,
HiveStorageFormat partitionStorageFormat) |
HiveWriterFactory(Set<HiveFileWriterFactory> fileWriterFactories,
String schemaName,
String tableName,
boolean isCreateTable,
List<HiveColumnHandle> inputColumns,
HiveStorageFormat tableStorageFormat,
HiveStorageFormat partitionStorageFormat,
OptionalInt bucketCount,
LocationHandle locationHandle,
LocationService locationService,
String filePrefix,
HivePageSinkMetadataProvider pageSinkMetadataProvider,
TypeManager typeManager,
HdfsEnvironment hdfsEnvironment,
boolean immutablePartitions,
ConnectorSession session,
NodeManager nodeManager,
io.airlift.event.client.EventClient eventClient,
HiveSessionProperties hiveSessionProperties) |
| Modifier and Type | Method and Description |
|---|---|
static OrcPageSource |
OrcPageSourceFactory.createOrcPageSource(MetadataReader metadataReader,
HdfsEnvironment hdfsEnvironment,
String sessionUser,
org.apache.hadoop.conf.Configuration configuration,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
List<HiveColumnHandle> columns,
boolean useOrcColumnNames,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager,
io.airlift.units.DataSize maxMergeDistance,
io.airlift.units.DataSize maxBufferSize,
io.airlift.units.DataSize streamBufferSize,
io.airlift.units.DataSize maxReadBlockSize,
boolean lazyReadSmallRanges,
boolean orcBloomFiltersEnabled,
FileFormatDataSourceStats stats) |
static OrcPageSource |
OrcPageSourceFactory.createOrcPageSource(MetadataReader metadataReader,
HdfsEnvironment hdfsEnvironment,
String sessionUser,
org.apache.hadoop.conf.Configuration configuration,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
List<HiveColumnHandle> columns,
boolean useOrcColumnNames,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager,
io.airlift.units.DataSize maxMergeDistance,
io.airlift.units.DataSize maxBufferSize,
io.airlift.units.DataSize streamBufferSize,
io.airlift.units.DataSize maxReadBlockSize,
boolean lazyReadSmallRanges,
boolean orcBloomFiltersEnabled,
FileFormatDataSourceStats stats) |
Optional<? extends ConnectorPageSource> |
OrcPageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone) |
Optional<? extends ConnectorPageSource> |
OrcPageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone) |
Optional<? extends ConnectorPageSource> |
DwrfPageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone) |
Optional<? extends ConnectorPageSource> |
DwrfPageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone) |
| Constructor and Description |
|---|
OrcPageSource(OrcRecordReader recordReader,
OrcDataSource orcDataSource,
List<HiveColumnHandle> columns,
TypeManager typeManager,
AggregatedMemoryContext systemMemoryContext,
FileFormatDataSourceStats stats) |
| Modifier and Type | Method and Description |
|---|---|
static parquet.schema.Type |
ParquetTypeUtils.getParquetType(HiveColumnHandle column,
parquet.schema.MessageType messageType,
boolean useParquetColumnNames) |
| Modifier and Type | Method and Description |
|---|---|
Optional<? extends ConnectorPageSource> |
ParquetPageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone) |
Optional<? extends ConnectorPageSource> |
ParquetPageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone) |
static ParquetPageSource |
ParquetPageSourceFactory.createParquetPageSource(HdfsEnvironment hdfsEnvironment,
String user,
org.apache.hadoop.conf.Configuration configuration,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
boolean useParquetColumnNames,
TypeManager typeManager,
boolean predicatePushdownEnabled,
TupleDomain<HiveColumnHandle> effectivePredicate) |
static ParquetPageSource |
ParquetPageSourceFactory.createParquetPageSource(HdfsEnvironment hdfsEnvironment,
String user,
org.apache.hadoop.conf.Configuration configuration,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
boolean useParquetColumnNames,
TypeManager typeManager,
boolean predicatePushdownEnabled,
TupleDomain<HiveColumnHandle> effectivePredicate) |
Optional<RecordCursor> |
ParquetRecordCursorProvider.createRecordCursor(String clientId,
org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager) |
Optional<RecordCursor> |
ParquetRecordCursorProvider.createRecordCursor(String clientId,
org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager) |
| Constructor and Description |
|---|
ParquetHiveRecordCursor(HdfsEnvironment hdfsEnvironment,
String sessionUser,
org.apache.hadoop.conf.Configuration configuration,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties splitSchema,
List<HiveColumnHandle> columns,
boolean useParquetColumnNames,
TypeManager typeManager,
boolean predicatePushdownEnabled,
TupleDomain<HiveColumnHandle> effectivePredicate) |
ParquetHiveRecordCursor(HdfsEnvironment hdfsEnvironment,
String sessionUser,
org.apache.hadoop.conf.Configuration configuration,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties splitSchema,
List<HiveColumnHandle> columns,
boolean useParquetColumnNames,
TypeManager typeManager,
boolean predicatePushdownEnabled,
TupleDomain<HiveColumnHandle> effectivePredicate) |
ParquetPageSource(ParquetReader parquetReader,
ParquetDataSource dataSource,
parquet.schema.MessageType fileSchema,
parquet.schema.MessageType requestedSchema,
Properties splitSchema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
TypeManager typeManager,
boolean useParquetColumnNames,
AggregatedMemoryContext systemMemoryContext) |
ParquetPageSource(ParquetReader parquetReader,
ParquetDataSource dataSource,
parquet.schema.MessageType fileSchema,
parquet.schema.MessageType requestedSchema,
Properties splitSchema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
TypeManager typeManager,
boolean useParquetColumnNames,
AggregatedMemoryContext systemMemoryContext) |
PrestoReadSupport(boolean useParquetColumnNames,
List<HiveColumnHandle> columns,
parquet.schema.MessageType messageType) |
| Modifier and Type | Method and Description |
|---|---|
static TupleDomain<parquet.column.ColumnDescriptor> |
ParquetPredicateUtils.getParquetTupleDomain(parquet.schema.MessageType fileSchema,
parquet.schema.MessageType requestedSchema,
TupleDomain<HiveColumnHandle> effectivePredicate) |
| Modifier and Type | Method and Description |
|---|---|
Optional<? extends ConnectorPageSource> |
RcFilePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone) |
Optional<? extends ConnectorPageSource> |
RcFilePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
org.apache.hadoop.fs.Path path,
long start,
long length,
long fileSize,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone) |
| Constructor and Description |
|---|
RcFilePageSource(RcFileReader rcFileReader,
List<HiveColumnHandle> columns,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager) |
| Modifier and Type | Method and Description |
|---|---|
TupleDomain<HiveColumnHandle> |
HiveFileIterator.getEffectivePredicate() |
| Constructor and Description |
|---|
HiveFileIterator(org.apache.hadoop.fs.Path path,
org.apache.hadoop.fs.FileSystem fileSystem,
DirectoryLister directoryLister,
NamenodeStats namenodeStats,
String partitionName,
org.apache.hadoop.mapred.InputFormat<?,?> inputFormat,
Properties schema,
List<HivePartitionKey> partitionKeys,
TupleDomain<HiveColumnHandle> effectivePredicate,
Map<Integer,HiveTypeName> columnCoercions) |
Copyright © 2012-2017. All Rights Reserved.