Uses of Class
io.trino.plugin.hive.HdfsEnvironment
-
-
Uses of HdfsEnvironment in io.trino.plugin.hive
Methods in io.trino.plugin.hive with parameters of type HdfsEnvironment Modifier and Type Method Description protected TransactionalMetadataHiveMetadataFactory. create(CatalogName catalogName, SemiTransactionalHiveMetastore metastore, HdfsEnvironment hdfsEnvironment, HivePartitionManager partitionManager, boolean writesToNonManagedTablesEnabled, boolean createsOfNonManagedTablesEnabled, boolean translateHiveViews, boolean hideDeltaLakeTables, TypeManager typeManager, LocationService locationService, io.airlift.json.JsonCodec<PartitionUpdate> partitionUpdateCodec, String trinoVersion, HiveStatisticsProvider hiveStatisticsProvider, HiveRedirectionsProvider hiveRedirectionsProvider, Set<SystemTableProvider> systemTableProviders, HiveMaterializedViewMetadata hiveMaterializedViewMetadata, AccessControlMetadata accessControlMetadata)Constructors in io.trino.plugin.hive with parameters of type HdfsEnvironment Constructor Description BackgroundHiveSplitLoader(Table table, AcidTransaction transaction, Iterable<HivePartitionMetadata> partitions, TupleDomain<? extends ColumnHandle> compactEffectivePredicate, DynamicFilter dynamicFilter, io.airlift.units.Duration dynamicFilteringProbeBlockingTimeout, TypeManager typeManager, Optional<BackgroundHiveSplitLoader.BucketSplitInfo> tableBucketInfo, ConnectorSession session, HdfsEnvironment hdfsEnvironment, NamenodeStats namenodeStats, DirectoryLister directoryLister, Executor executor, int loaderConcurrency, boolean recursiveDirWalkerEnabled, boolean ignoreAbsentPartitions, boolean optimizeSymlinkListing, Optional<org.apache.hadoop.hive.common.ValidWriteIdList> validWriteIds)GenericHiveRecordCursorProvider(HdfsEnvironment hdfsEnvironment, io.airlift.units.DataSize textMaxLineLength)GenericHiveRecordCursorProvider(HdfsEnvironment hdfsEnvironment, HiveConfig config)HiveLocationService(HdfsEnvironment hdfsEnvironment)HiveMetadata(CatalogName catalogName, SemiTransactionalHiveMetastore metastore, HdfsEnvironment hdfsEnvironment, HivePartitionManager partitionManager, boolean writesToNonManagedTablesEnabled, boolean createsOfNonManagedTablesEnabled, boolean translateHiveViews, boolean hideDeltaLakeTables, TypeManager typeManager, LocationService locationService, io.airlift.json.JsonCodec<PartitionUpdate> partitionUpdateCodec, String trinoVersion, HiveStatisticsProvider hiveStatisticsProvider, HiveRedirectionsProvider hiveRedirectionsProvider, Set<SystemTableProvider> systemTableProviders, HiveMaterializedViewMetadata hiveMaterializedViewMetadata, AccessControlMetadata accessControlMetadata)HiveMetadataFactory(CatalogName catalogName, HiveConfig hiveConfig, MetastoreConfig metastoreConfig, HiveMetastore metastore, HdfsEnvironment hdfsEnvironment, HivePartitionManager partitionManager, ExecutorService executorService, ScheduledExecutorService heartbeatService, TypeManager typeManager, LocationService locationService, io.airlift.json.JsonCodec<PartitionUpdate> partitionUpdateCodec, NodeVersion nodeVersion, HiveRedirectionsProvider hiveRedirectionsProvider, Set<SystemTableProvider> systemTableProviders, HiveMaterializedViewMetadataFactory hiveMaterializedViewMetadataFactory, AccessControlMetadataFactory accessControlMetadataFactory)HiveMetadataFactory(CatalogName catalogName, HiveMetastore metastore, HdfsEnvironment hdfsEnvironment, HivePartitionManager partitionManager, int maxConcurrentFileRenames, int maxConcurrentMetastoreDrops, int maxConcurrentMetastoreUpdates, boolean skipDeletionForAlter, boolean skipTargetCleanupOnRollback, boolean writesToNonManagedTablesEnabled, boolean createsOfNonManagedTablesEnabled, boolean translateHiveViews, long perTransactionCacheMaximumSize, Optional<io.airlift.units.Duration> hiveTransactionHeartbeatInterval, boolean hideDeltaLakeTables, TypeManager typeManager, LocationService locationService, io.airlift.json.JsonCodec<PartitionUpdate> partitionUpdateCodec, ExecutorService executorService, ScheduledExecutorService heartbeatService, String trinoVersion, HiveRedirectionsProvider hiveRedirectionsProvider, Set<SystemTableProvider> systemTableProviders, HiveMaterializedViewMetadataFactory hiveMaterializedViewMetadataFactory, AccessControlMetadataFactory accessControlMetadataFactory)HivePageSink(HiveWriterFactory writerFactory, List<HiveColumnHandle> inputColumns, Optional<HiveBucketProperty> bucketProperty, PageIndexerFactory pageIndexerFactory, HdfsEnvironment hdfsEnvironment, int maxOpenWriters, com.google.common.util.concurrent.ListeningExecutorService writeVerificationExecutor, io.airlift.json.JsonCodec<PartitionUpdate> partitionUpdateCodec, ConnectorSession session)HivePageSinkProvider(Set<HiveFileWriterFactory> fileWriterFactories, HdfsEnvironment hdfsEnvironment, PageSorter pageSorter, HiveMetastore metastore, PageIndexerFactory pageIndexerFactory, TypeManager typeManager, HiveConfig config, LocationService locationService, io.airlift.json.JsonCodec<PartitionUpdate> partitionUpdateCodec, NodeManager nodeManager, io.airlift.event.client.EventClient eventClient, HiveSessionProperties hiveSessionProperties, HiveWriterStats hiveWriterStats)HivePageSourceProvider(TypeManager typeManager, HdfsEnvironment hdfsEnvironment, HiveConfig hiveConfig, Set<HivePageSourceFactory> pageSourceFactories, Set<HiveRecordCursorProvider> cursorProviders, GenericHiveRecordCursorProvider genericCursorProvider, OrcFileWriterFactory orcFileWriterFactory)HivePageSourceProvider(TypeManager typeManager, HdfsEnvironment hdfsEnvironment, HiveConfig hiveConfig, Set<HivePageSourceFactory> pageSourceFactories, Set<HiveRecordCursorProvider> cursorProviders, GenericHiveRecordCursorProvider genericCursorProvider, Optional<OrcFileWriterFactory> orcFileWriterFactory)HiveSplitManager(HiveConfig hiveConfig, Function<HiveTransactionHandle,SemiTransactionalHiveMetastore> metastoreProvider, HivePartitionManager partitionManager, NamenodeStats namenodeStats, HdfsEnvironment hdfsEnvironment, DirectoryLister directoryLister, ExecutorService executorService, VersionEmbedder versionEmbedder, TypeManager typeManager)HiveSplitManager(Function<HiveTransactionHandle,SemiTransactionalHiveMetastore> metastoreProvider, HivePartitionManager partitionManager, NamenodeStats namenodeStats, HdfsEnvironment hdfsEnvironment, DirectoryLister directoryLister, Executor executor, io.airlift.stats.CounterStat highMemorySplitSourceCounter, int maxOutstandingSplits, io.airlift.units.DataSize maxOutstandingSplitsSize, int minPartitionBatchSize, int maxPartitionBatchSize, int maxInitialSplits, int splitLoaderConcurrency, Integer maxSplitsPerSecond, boolean recursiveDfsWalkerEnabled, TypeManager typeManager)HiveWriterFactory(Set<HiveFileWriterFactory> fileWriterFactories, String schemaName, String tableName, boolean isCreateTable, AcidTransaction transaction, List<HiveColumnHandle> inputColumns, HiveStorageFormat tableStorageFormat, HiveStorageFormat partitionStorageFormat, Map<String,String> additionalTableParameters, OptionalInt bucketCount, List<SortingColumn> sortedBy, LocationHandle locationHandle, LocationService locationService, String queryId, HivePageSinkMetadataProvider pageSinkMetadataProvider, TypeManager typeManager, HdfsEnvironment hdfsEnvironment, PageSorter pageSorter, io.airlift.units.DataSize sortBufferSize, int maxOpenSortFiles, org.joda.time.DateTimeZone parquetTimeZone, ConnectorSession session, NodeManager nodeManager, io.airlift.event.client.EventClient eventClient, HiveSessionProperties hiveSessionProperties, HiveWriterStats hiveWriterStats)RcFileFileWriterFactory(HdfsEnvironment hdfsEnvironment, TypeManager typeManager, NodeVersion nodeVersion, HiveConfig hiveConfig, FileFormatDataSourceStats stats)RcFileFileWriterFactory(HdfsEnvironment hdfsEnvironment, TypeManager typeManager, NodeVersion nodeVersion, org.joda.time.DateTimeZone timeZone, FileFormatDataSourceStats stats) -
Uses of HdfsEnvironment in io.trino.plugin.hive.metastore
Constructors in io.trino.plugin.hive.metastore with parameters of type HdfsEnvironment Constructor Description SemiTransactionalHiveMetastore(HdfsEnvironment hdfsEnvironment, HiveMetastoreClosure delegate, Executor renameExecutor, Executor dropExecutor, Executor updateExecutor, boolean skipDeletionForAlter, boolean skipTargetCleanupOnRollback, Optional<io.airlift.units.Duration> hiveTransactionHeartbeatInterval, ScheduledExecutorService heartbeatService) -
Uses of HdfsEnvironment in io.trino.plugin.hive.metastore.file
Constructors in io.trino.plugin.hive.metastore.file with parameters of type HdfsEnvironment Constructor Description FileHiveMetastore(NodeVersion nodeVersion, HdfsEnvironment hdfsEnvironment, MetastoreConfig metastoreConfig, FileHiveMetastoreConfig config) -
Uses of HdfsEnvironment in io.trino.plugin.hive.metastore.glue
Constructors in io.trino.plugin.hive.metastore.glue with parameters of type HdfsEnvironment Constructor Description GlueHiveMetastore(HdfsEnvironment hdfsEnvironment, GlueHiveMetastoreConfig glueConfig, HiveConfig hiveConfig, Executor partitionsReadExecutor, GlueColumnStatisticsProviderFactory columnStatisticsProviderFactory, Optional<com.amazonaws.handlers.RequestHandler2> requestHandler, Predicate<com.amazonaws.services.glue.model.Table> tableFilter) -
Uses of HdfsEnvironment in io.trino.plugin.hive.metastore.thrift
Constructors in io.trino.plugin.hive.metastore.thrift with parameters of type HdfsEnvironment Constructor Description ThriftHiveMetastore(MetastoreLocator metastoreLocator, HiveConfig hiveConfig, MetastoreConfig metastoreConfig, ThriftMetastoreConfig thriftConfig, HdfsEnvironment hdfsEnvironment, boolean authenticationEnabled)ThriftHiveMetastore(MetastoreLocator metastoreLocator, HiveConfig hiveConfig, MetastoreConfig metastoreConfig, ThriftMetastoreConfig thriftConfig, ThriftMetastoreAuthenticationConfig authenticationConfig, HdfsEnvironment hdfsEnvironment) -
Uses of HdfsEnvironment in io.trino.plugin.hive.orc
Methods in io.trino.plugin.hive.orc with parameters of type HdfsEnvironment Modifier and Type Method Description static Optional<ConnectorPageSource>OrcDeleteDeltaPageSource. createOrcDeleteDeltaPageSource(org.apache.hadoop.fs.Path path, long fileSize, OrcReaderOptions options, ConnectorIdentity identity, org.apache.hadoop.conf.Configuration configuration, HdfsEnvironment hdfsEnvironment, FileFormatDataSourceStats stats)static longOriginalFilesUtils. getPrecedingRowCount(Collection<AcidInfo.OriginalFileInfo> originalFileInfos, org.apache.hadoop.fs.Path splitPath, HdfsEnvironment hdfsEnvironment, ConnectorIdentity identity, OrcReaderOptions options, org.apache.hadoop.conf.Configuration configuration, FileFormatDataSourceStats stats)Returns total number of rows present before the given original file in the same bucket. -
Uses of HdfsEnvironment in io.trino.plugin.hive.parquet
Methods in io.trino.plugin.hive.parquet with parameters of type HdfsEnvironment Modifier and Type Method Description static ReaderPageSourceParquetPageSourceFactory. createPageSource(org.apache.hadoop.fs.Path path, long start, long length, long estimatedFileSize, List<HiveColumnHandle> columns, TupleDomain<HiveColumnHandle> effectivePredicate, boolean useColumnNames, HdfsEnvironment hdfsEnvironment, org.apache.hadoop.conf.Configuration configuration, ConnectorIdentity identity, org.joda.time.DateTimeZone timeZone, FileFormatDataSourceStats stats, ParquetReaderOptions options)This method is available for other callers to use directly.Constructors in io.trino.plugin.hive.parquet with parameters of type HdfsEnvironment Constructor Description ParquetFileWriterFactory(HdfsEnvironment hdfsEnvironment, TypeManager typeManager)ParquetPageSourceFactory(HdfsEnvironment hdfsEnvironment, FileFormatDataSourceStats stats, ParquetReaderConfig config, HiveConfig hiveConfig) -
Uses of HdfsEnvironment in io.trino.plugin.hive.procedure
Constructors in io.trino.plugin.hive.procedure with parameters of type HdfsEnvironment Constructor Description RegisterPartitionProcedure(HiveConfig hiveConfig, TransactionalMetadataFactory hiveMetadataFactory, HiveMetastore metastore, HdfsEnvironment hdfsEnvironment)SyncPartitionMetadataProcedure(TransactionalMetadataFactory hiveMetadataFactory, HdfsEnvironment hdfsEnvironment) -
Uses of HdfsEnvironment in io.trino.plugin.hive.rcfile
Constructors in io.trino.plugin.hive.rcfile with parameters of type HdfsEnvironment Constructor Description RcFilePageSourceFactory(TypeManager typeManager, HdfsEnvironment hdfsEnvironment, FileFormatDataSourceStats stats, HiveConfig hiveConfig) -
Uses of HdfsEnvironment in io.trino.plugin.hive.s3select
Constructors in io.trino.plugin.hive.s3select with parameters of type HdfsEnvironment Constructor Description S3SelectRecordCursorProvider(HdfsEnvironment hdfsEnvironment, TrinoS3ClientFactory s3ClientFactory) -
Uses of HdfsEnvironment in io.trino.plugin.hive.util
Methods in io.trino.plugin.hive.util with parameters of type HdfsEnvironment Modifier and Type Method Description static voidHiveWriteUtils. createDirectory(HdfsEnvironment.HdfsContext context, HdfsEnvironment hdfsEnvironment, org.apache.hadoop.fs.Path path)static org.apache.hadoop.fs.PathHiveWriteUtils. createTemporaryPath(ConnectorSession session, HdfsEnvironment.HdfsContext context, HdfsEnvironment hdfsEnvironment, org.apache.hadoop.fs.Path targetPath)static org.apache.hadoop.fs.PathHiveWriteUtils. getTableDefaultLocation(HdfsEnvironment.HdfsContext context, SemiTransactionalHiveMetastore metastore, HdfsEnvironment hdfsEnvironment, String schemaName, String tableName)static org.apache.hadoop.fs.PathHiveWriteUtils. getTableDefaultLocation(Database database, HdfsEnvironment.HdfsContext context, HdfsEnvironment hdfsEnvironment, String schemaName, String tableName)static booleanHiveWriteUtils. isHdfsEncrypted(HdfsEnvironment.HdfsContext context, HdfsEnvironment hdfsEnvironment, org.apache.hadoop.fs.Path path)static booleanHiveWriteUtils. isS3FileSystem(HdfsEnvironment.HdfsContext context, HdfsEnvironment hdfsEnvironment, org.apache.hadoop.fs.Path path)static booleanHiveWriteUtils. isViewFileSystem(HdfsEnvironment.HdfsContext context, HdfsEnvironment hdfsEnvironment, org.apache.hadoop.fs.Path path)static booleanHiveWriteUtils. pathExists(HdfsEnvironment.HdfsContext context, HdfsEnvironment hdfsEnvironment, org.apache.hadoop.fs.Path path)
-