| Modifier and Type | Method and Description |
|---|---|
static EncryptionInformation |
EncryptionInformation.fromEncryptionMetadata(EncryptionMetadata encryptionMetadata) |
protected abstract EncryptionInformation |
AbstractDwrfEncryptionInformationSource.getReadEncryptionInformationInternal(ConnectorSession session,
Table table,
Optional<Set<HiveColumnHandle>> requestedColumns,
Map<String,String> fieldToKeyReference,
DwrfTableEncryptionProperties encryptionProperties) |
protected abstract EncryptionInformation |
AbstractDwrfEncryptionInformationSource.getWriteEncryptionInformationInternal(ConnectorSession session,
DwrfTableEncryptionProperties tableEncryptionProperties,
String dbName,
String tableName) |
| Modifier and Type | Method and Description |
|---|---|
Optional<HiveFileWriter> |
RcFileFileWriterFactory.createFileWriter(Path path,
List<String> inputColumnNames,
StorageFormat storageFormat,
Properties schema,
org.apache.hadoop.mapred.JobConf configuration,
ConnectorSession session,
Optional<EncryptionInformation> encryptionInformation) |
Optional<HiveFileWriter> |
OrcFileWriterFactory.createFileWriter(Path path,
List<String> inputColumnNames,
StorageFormat storageFormat,
Properties schema,
org.apache.hadoop.mapred.JobConf configuration,
ConnectorSession session,
Optional<EncryptionInformation> encryptionInformation) |
Optional<HiveFileWriter> |
HiveFileWriterFactory.createFileWriter(Path path,
List<String> inputColumnNames,
StorageFormat storageFormat,
Properties schema,
org.apache.hadoop.mapred.JobConf conf,
ConnectorSession session,
Optional<EncryptionInformation> encryptionInformation) |
static Optional<ConnectorPageSource> |
HivePageSourceProvider.createHivePageSource(Set<HiveRecordCursorProvider> cursorProviders,
Set<HiveBatchPageSourceFactory> pageSourceFactories,
org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
OptionalInt tableBucketNumber,
Storage storage,
TupleDomain<HiveColumnHandle> effectivePredicate,
List<HiveColumnHandle> hiveColumns,
Map<String,HiveColumnHandle> predicateColumns,
List<HivePartitionKey> partitionKeys,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager,
SchemaTableName tableName,
List<HiveColumnHandle> partitionKeyColumnHandles,
List<Column> tableDataColumns,
Map<String,String> tableParameters,
int partitionDataColumnCount,
TableToPartitionMapping tableToPartitionMapping,
Optional<HiveSplit.BucketConversion> bucketConversion,
boolean s3SelectPushdownEnabled,
HiveFileContext hiveFileContext,
RowExpression remainingPredicate,
boolean isPushdownFilterEnabled,
RowExpressionService rowExpressionService,
Optional<EncryptionInformation> encryptionInformation) |
Optional<? extends ConnectorPageSource> |
HiveSelectivePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
List<HiveColumnHandle> columns,
Map<Integer,String> prefilledValues,
Map<Integer,HiveCoercer> coercers,
Optional<BucketAdaptation> bucketAdaptation,
List<Integer> outputColumns,
TupleDomain<Subfield> domainPredicate,
RowExpression remainingPredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation,
boolean appendRowNumberEnabled,
boolean footerStatsUnreliable) |
Optional<? extends ConnectorPageSource> |
HiveBatchPageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
SchemaTableName tableName,
Map<String,String> tableParameters,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation) |
| Constructor and Description |
|---|
HiveInsertTableHandle(String schemaName,
String tableName,
List<HiveColumnHandle> inputColumns,
HivePageSinkMetadata pageSinkMetadata,
LocationHandle locationHandle,
Optional<HiveBucketProperty> bucketProperty,
List<SortingColumn> preferredOrderingColumns,
HiveStorageFormat tableStorageFormat,
HiveStorageFormat partitionStorageFormat,
HiveStorageFormat actualStorageFormat,
HiveCompressionCodec compressionCodec,
Optional<EncryptionInformation> encryptionInformation) |
HiveOutputTableHandle(String schemaName,
String tableName,
List<HiveColumnHandle> inputColumns,
HivePageSinkMetadata pageSinkMetadata,
LocationHandle locationHandle,
HiveStorageFormat tableStorageFormat,
HiveStorageFormat partitionStorageFormat,
HiveStorageFormat actualStorageFormat,
HiveCompressionCodec compressionCodec,
List<String> partitionedBy,
Optional<HiveBucketProperty> bucketProperty,
List<SortingColumn> preferredOrderingColumns,
String tableOwner,
Map<String,String> additionalTableParameters,
Optional<EncryptionInformation> encryptionInformation) |
HiveSplit(HiveFileSplit fileSplit,
String database,
String table,
String partitionName,
Storage storage,
List<HivePartitionKey> partitionKeys,
List<HostAddress> addresses,
OptionalInt readBucketNumber,
OptionalInt tableBucketNumber,
NodeSelectionStrategy nodeSelectionStrategy,
int partitionDataColumnCount,
TableToPartitionMapping tableToPartitionMapping,
Optional<HiveSplit.BucketConversion> bucketConversion,
boolean s3SelectPushdownEnabled,
CacheQuotaRequirement cacheQuotaRequirement,
Optional<EncryptionInformation> encryptionInformation,
Set<ColumnHandle> redundantColumnDomains,
SplitWeight splitWeight) |
HiveWritableTableHandle(String schemaName,
String tableName,
List<HiveColumnHandle> inputColumns,
HivePageSinkMetadata pageSinkMetadata,
LocationHandle locationHandle,
Optional<HiveBucketProperty> bucketProperty,
List<SortingColumn> preferredOrderingColumns,
HiveStorageFormat tableStorageFormat,
HiveStorageFormat partitionStorageFormat,
HiveStorageFormat actualStorageFormat,
HiveCompressionCodec compressionCodec,
Optional<EncryptionInformation> encryptionInformation) |
HiveWriterFactory(Set<HiveFileWriterFactory> fileWriterFactories,
String schemaName,
String tableName,
boolean isCreateTable,
List<HiveColumnHandle> inputColumns,
HiveStorageFormat tableStorageFormat,
HiveStorageFormat partitionStorageFormat,
HiveCompressionCodec compressionCodec,
Map<String,String> additionalTableParameters,
OptionalInt bucketCount,
List<SortingColumn> sortedBy,
LocationHandle locationHandle,
LocationService locationService,
String queryId,
HivePageSinkMetadataProvider pageSinkMetadataProvider,
TypeManager typeManager,
HdfsEnvironment hdfsEnvironment,
PageSorter pageSorter,
io.airlift.units.DataSize sortBufferSize,
int maxOpenSortFiles,
boolean immutablePartitions,
ConnectorSession session,
NodeManager nodeManager,
com.facebook.airlift.event.client.EventClient eventClient,
HiveSessionProperties hiveSessionProperties,
HiveWriterStats hiveWriterStats,
OrcFileWriterFactory orcFileWriterFactory,
boolean commitRequired,
Optional<EncryptionInformation> encryptionInformation) |
InternalHiveSplit(String relativeUri,
long start,
long end,
long fileSize,
long fileModifiedTime,
List<InternalHiveSplit.InternalHiveBlock> blocks,
OptionalInt readBucketNumber,
OptionalInt tableBucketNumber,
boolean splittable,
NodeSelectionStrategy nodeSelectionStrategy,
boolean s3SelectPushdownEnabled,
HiveSplitPartitionInfo partitionInfo,
Optional<byte[]> extraFileInfo,
Optional<EncryptionInformation> encryptionInformation,
Map<String,String> customSplitInfo) |
| Modifier and Type | Method and Description |
|---|---|
static ConnectorPageSource |
OrcSelectivePageSourceFactory.createOrcPageSource(ConnectorSession session,
OrcEncoding orcEncoding,
HdfsEnvironment hdfsEnvironment,
org.apache.hadoop.conf.Configuration configuration,
HiveFileSplit fileSplit,
List<HiveColumnHandle> columns,
Map<Integer,String> prefilledValues,
Map<Integer,HiveCoercer> coercers,
Optional<BucketAdaptation> bucketAdaptation,
List<Integer> outputColumns,
TupleDomain<Subfield> domainPredicate,
RowExpression remainingPredicate,
boolean useOrcColumnNames,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager,
StandardFunctionResolution functionResolution,
RowExpressionService rowExpressionService,
boolean orcBloomFiltersEnabled,
FileFormatDataSourceStats stats,
int domainCompactionThreshold,
OrcFileTailSource orcFileTailSource,
StripeMetadataSourceFactory stripeMetadataSourceFactory,
HiveFileContext hiveFileContext,
TupleDomainFilterCache tupleDomainFilterCache,
Optional<EncryptionInformation> encryptionInformation,
DwrfEncryptionProvider dwrfEncryptionProvider,
boolean appendRowNumberEnabled,
boolean footerStatsUnreliable) |
static ConnectorPageSource |
OrcBatchPageSourceFactory.createOrcPageSource(OrcEncoding orcEncoding,
HdfsEnvironment hdfsEnvironment,
String sessionUser,
org.apache.hadoop.conf.Configuration configuration,
HiveFileSplit fileSplit,
List<HiveColumnHandle> columns,
boolean useOrcColumnNames,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
TypeManager typeManager,
StandardFunctionResolution functionResolution,
io.airlift.units.DataSize maxBufferSize,
io.airlift.units.DataSize streamBufferSize,
boolean lazyReadSmallRanges,
boolean orcBloomFiltersEnabled,
FileFormatDataSourceStats stats,
int domainCompactionThreshold,
OrcFileTailSource orcFileTailSource,
StripeMetadataSourceFactory stripeMetadataSourceFactory,
HiveFileContext hiveFileContext,
OrcReaderOptions orcReaderOptions,
Optional<EncryptionInformation> encryptionInformation,
DwrfEncryptionProvider dwrfEncryptionProvider) |
Optional<? extends ConnectorPageSource> |
OrcSelectivePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
List<HiveColumnHandle> columns,
Map<Integer,String> prefilledValues,
Map<Integer,HiveCoercer> coercers,
Optional<BucketAdaptation> bucketAdaptation,
List<Integer> outputColumns,
TupleDomain<Subfield> domainPredicate,
RowExpression remainingPredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation,
boolean appendRowNumberEnabled,
boolean footerStatsUnreliable) |
Optional<? extends ConnectorPageSource> |
DwrfSelectivePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
List<HiveColumnHandle> columns,
Map<Integer,String> prefilledValues,
Map<Integer,HiveCoercer> coercers,
Optional<BucketAdaptation> bucketAdaptation,
List<Integer> outputColumns,
TupleDomain<Subfield> domainPredicate,
RowExpression remainingPredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation,
boolean appendRowNumberEnabled,
boolean footerStatsUnreliable) |
Optional<? extends ConnectorPageSource> |
OrcBatchPageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
SchemaTableName tableName,
Map<String,String> tableParameters,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation) |
Optional<? extends ConnectorPageSource> |
DwrfBatchPageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
SchemaTableName tableName,
Map<String,String> tableParameters,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation) |
| Constructor and Description |
|---|
ProjectionBasedDwrfKeyProvider(Optional<EncryptionInformation> encryptionInformation,
List<HiveColumnHandle> columns,
boolean useOrcColumnNames,
Path path) |
| Modifier and Type | Method and Description |
|---|---|
Optional<HiveFileWriter> |
PageFileWriterFactory.createFileWriter(Path path,
List<String> inputColumnNames,
StorageFormat storageFormat,
Properties schema,
org.apache.hadoop.mapred.JobConf configuration,
ConnectorSession session,
Optional<EncryptionInformation> encryptionInformation) |
Optional<? extends ConnectorPageSource> |
PageFilePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
SchemaTableName tableName,
Map<String,String> tableParameters,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation) |
| Modifier and Type | Method and Description |
|---|---|
Optional<HiveFileWriter> |
ParquetFileWriterFactory.createFileWriter(Path path,
List<String> inputColumnNames,
StorageFormat storageFormat,
Properties schema,
org.apache.hadoop.mapred.JobConf conf,
ConnectorSession session,
Optional<EncryptionInformation> encryptionInformation) |
Optional<? extends ConnectorPageSource> |
ParquetSelectivePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
List<HiveColumnHandle> columns,
Map<Integer,String> prefilledValues,
Map<Integer,HiveCoercer> coercers,
Optional<BucketAdaptation> bucketAdaptation,
List<Integer> outputColumns,
TupleDomain<Subfield> domainPredicate,
RowExpression remainingPredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation,
boolean appendRowNumberEnabled,
boolean footerStatsUnreliable) |
Optional<? extends ConnectorPageSource> |
ParquetPageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
SchemaTableName tableName,
Map<String,String> tableParameters,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation) |
| Modifier and Type | Method and Description |
|---|---|
Optional<? extends ConnectorPageSource> |
RcFilePageSourceFactory.createPageSource(org.apache.hadoop.conf.Configuration configuration,
ConnectorSession session,
HiveFileSplit fileSplit,
Storage storage,
SchemaTableName tableName,
Map<String,String> tableParameters,
List<HiveColumnHandle> columns,
TupleDomain<HiveColumnHandle> effectivePredicate,
org.joda.time.DateTimeZone hiveStorageTimeZone,
HiveFileContext hiveFileContext,
Optional<EncryptionInformation> encryptionInformation) |
| Constructor and Description |
|---|
InternalHiveSplitFactory(FileSystem fileSystem,
org.apache.hadoop.mapred.InputFormat<?,?> inputFormat,
Optional<Domain> pathDomain,
NodeSelectionStrategy nodeSelectionStrategy,
io.airlift.units.DataSize minimumTargetSplitSize,
boolean s3SelectPushdownEnabled,
HiveSplitPartitionInfo partitionInfo,
boolean schedulerUsesHostAddresses,
Optional<EncryptionInformation> encryptionInformation) |
Copyright © 2012–2023. All rights reserved.