| Modifier and Type | Method and Description |
|---|---|
HiveFileSplit |
HiveSplit.getFileSplit() |
| Modifier and Type | Method and Description |
|---|---|
static List<HivePageSourceProvider.ColumnMapping> |
HivePageSourceProvider.ColumnMapping.buildColumnMappings(List<HivePartitionKey> partitionKeys,
List<HiveColumnHandle> columns,
List<HiveColumnHandle> requiredInterimColumns,
TableToPartitionMapping tableToPartitionMapping,
HiveFileSplit fileSplit,
OptionalInt bucketNumber) |
static Optional<ConnectorPageSource> |
HivePageSourceProvider.createHivePageSource(Set<HiveRecordCursorProvider> cursorProviders,
Set<HiveBatchPageSourceFactory> pageSourceFactories,
org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
OptionalInt tableBucketNumber,
Storage storage,
TupleDomain<HiveColumnHandle> effectivePredicate,
List<HiveColumnHandle> hiveColumns,
Map<String,HiveColumnHandle> predicateColumns,
List<HivePartitionKey> partitionKeys,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager,
SchemaTableName tableName,
List<HiveColumnHandle> partitionKeyColumnHandles,
List<Column> tableDataColumns,
Map<String,String> tableParameters,
int partitionDataColumnCount,
TableToPartitionMapping tableToPartitionMapping,
Optional<HiveSplit.BucketConversion> bucketConversion,
boolean s3SelectPushdownEnabled,
HiveFileContext hiveFileContext,
RowExpression remainingPredicate,
boolean isPushdownFilterEnabled,
RowExpressionService rowExpressionService,
Optional<EncryptionInformation> encryptionInformation) |
Optional<? extends ConnectorPageSource> |
HiveSelectivePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
List<HiveColumnHandle> columns,
Map<Integer,String> prefilledValues,
Map<Integer,HiveCoercer> coercers,
Optional<BucketAdaptation> bucketAdaptation,
List<Integer> outputColumns,
TupleDomain<Subfield> domainPredicate,
RowExpression remainingPredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation,
boolean appendRowNumberEnabled,
boolean footerStatsUnreliable) |
Optional<? extends ConnectorPageSource> |
HiveBatchPageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
SchemaTableName tableName,
Map<String,String> tableParameters,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation) |
Optional<RecordCursor> |
HiveRecordCursorProvider.createRecordCursor(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager,
boolean s3SelectPushdownEnabled) |
Optional<RecordCursor> |
GenericHiveRecordCursorProvider.createRecordCursor(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager,
boolean s3SelectPushdownEnabled) |
static Optional<String> |
HiveUtil.getPrefilledColumnValue(HiveColumnHandle columnHandle,
HivePartitionKey partitionKey,
HiveFileSplit fileSplit,
OptionalInt bucketNumber) |
| Constructor and Description |
|---|
HiveSplit(HiveFileSplit fileSplit,
String database,
String table,
String partitionName,
Storage storage,
List<HivePartitionKey> partitionKeys,
List<HostAddress> addresses,
OptionalInt readBucketNumber,
OptionalInt tableBucketNumber,
NodeSelectionStrategy nodeSelectionStrategy,
int partitionDataColumnCount,
TableToPartitionMapping tableToPartitionMapping,
Optional<HiveSplit.BucketConversion> bucketConversion,
boolean s3SelectPushdownEnabled,
CacheQuotaRequirement cacheQuotaRequirement,
Optional<EncryptionInformation> encryptionInformation,
Set<ColumnHandle> redundantColumnDomains,
SplitWeight splitWeight) |
| Modifier and Type | Method and Description |
|---|---|
static ConnectorPageSource |
OrcSelectivePageSourceFactory.createOrcPageSource(ConnectorSession session,
OrcEncoding orcEncoding,
HdfsEnvironment hdfsEnvironment,
org.apache.hadoop.conf.Configuration configuration,
HiveFileSplit fileSplit,
List<HiveColumnHandle> columns,
Map<Integer,String> prefilledValues,
Map<Integer,HiveCoercer> coercers,
Optional<BucketAdaptation> bucketAdaptation,
List<Integer> outputColumns,
TupleDomain<Subfield> domainPredicate,
RowExpression remainingPredicate,
boolean useOrcColumnNames,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager,
StandardFunctionResolution functionResolution,
RowExpressionService rowExpressionService,
boolean orcBloomFiltersEnabled,
FileFormatDataSourceStats stats,
int domainCompactionThreshold,
OrcFileTailSource orcFileTailSource,
StripeMetadataSourceFactory stripeMetadataSourceFactory,
HiveFileContext hiveFileContext,
TupleDomainFilterCache tupleDomainFilterCache,
Optional<EncryptionInformation> encryptionInformation,
DwrfEncryptionProvider dwrfEncryptionProvider,
boolean appendRowNumberEnabled,
boolean footerStatsUnreliable) |
static ConnectorPageSource |
OrcBatchPageSourceFactory.createOrcPageSource(OrcEncoding orcEncoding,
HdfsEnvironment hdfsEnvironment,
String sessionUser,
org.apache.hadoop.conf.Configuration configuration,
HiveFileSplit fileSplit,
List<HiveColumnHandle> columns,
boolean useOrcColumnNames,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager,
StandardFunctionResolution functionResolution,
io.airlift.units.DataSize maxBufferSize,
io.airlift.units.DataSize streamBufferSize,
boolean lazyReadSmallRanges,
boolean orcBloomFiltersEnabled,
FileFormatDataSourceStats stats,
int domainCompactionThreshold,
OrcFileTailSource orcFileTailSource,
StripeMetadataSourceFactory stripeMetadataSourceFactory,
HiveFileContext hiveFileContext,
OrcReaderOptions orcReaderOptions,
Optional<EncryptionInformation> encryptionInformation,
DwrfEncryptionProvider dwrfEncryptionProvider) |
Optional<? extends ConnectorPageSource> |
OrcSelectivePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
List<HiveColumnHandle> columns,
Map<Integer,String> prefilledValues,
Map<Integer,HiveCoercer> coercers,
Optional<BucketAdaptation> bucketAdaptation,
List<Integer> outputColumns,
TupleDomain<Subfield> domainPredicate,
RowExpression remainingPredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation,
boolean appendRowNumberEnabled,
boolean footerStatsUnreliable) |
Optional<? extends ConnectorPageSource> |
DwrfSelectivePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
List<HiveColumnHandle> columns,
Map<Integer,String> prefilledValues,
Map<Integer,HiveCoercer> coercers,
Optional<BucketAdaptation> bucketAdaptation,
List<Integer> outputColumns,
TupleDomain<Subfield> domainPredicate,
RowExpression remainingPredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation,
boolean appendRowNumberEnabled,
boolean footerStatsUnreliable) |
Optional<? extends ConnectorPageSource> |
OrcBatchPageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
SchemaTableName tableName,
Map<String,String> tableParameters,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation) |
Optional<? extends ConnectorPageSource> |
DwrfBatchPageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
SchemaTableName tableName,
Map<String,String> tableParameters,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation) |
| Modifier and Type | Method and Description |
|---|---|
Optional<? extends ConnectorPageSource> |
PageFilePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
SchemaTableName tableName,
Map<String,String> tableParameters,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation) |
| Modifier and Type | Method and Description |
|---|---|
Optional<? extends ConnectorPageSource> |
ParquetSelectivePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
List<HiveColumnHandle> columns,
Map<Integer,String> prefilledValues,
Map<Integer,HiveCoercer> coercers,
Optional<BucketAdaptation> bucketAdaptation,
List<Integer> outputColumns,
TupleDomain<Subfield> domainPredicate,
RowExpression remainingPredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation,
boolean appendRowNumberEnabled,
boolean footerStatsUnreliable) |
Optional<? extends ConnectorPageSource> |
ParquetPageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
SchemaTableName tableName,
Map<String,String> tableParameters,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation) |
static ConnectorPageSource |
ParquetPageSourceFactory.createParquetPageSource(HdfsEnvironment hdfsEnvironment,
ConnectorSession session,
org.apache.hadoop.conf.Configuration configuration,
HiveFileSplit fileSplit,
List<HiveColumnHandle> columns,
SchemaTableName tableName,
TypeManager typeManager,
StandardFunctionResolution functionResolution,
TupleDomain<HiveColumnHandle> effectivePredicate,
FileFormatDataSourceStats stats,
HiveFileContext hiveFileContext,
ParquetMetadataSource parquetMetadataSource) |
| Modifier and Type | Method and Description |
|---|---|
Optional<? extends ConnectorPageSource> |
RcFilePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
SchemaTableName tableName,
Map<String,String> tableParameters,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation) |
| Modifier and Type | Method and Description |
|---|---|
Optional<RecordCursor> |
S3SelectRecordCursorProvider.createRecordCursor(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Properties schema,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager,
boolean s3SelectPushdownEnabled) |
Copyright © 2012–2023. All rights reserved.