BaseFileOnlyRelation.HoodieBaseFileSplit$ HoodieBaseFileSplit$module
scala.collection.Seq<A> mandatoryFields
scala.Option<A> org$apache$hudi$BaseFileOnlyRelation$$userSchema
scala.collection.Seq<A> org$apache$hudi$BaseFileOnlyRelation$$globPaths
scala.Option<A> org$apache$hudi$BaseFileOnlyRelation$$prunedDataSchema
boolean shouldExtractPartitionValuesFromPartitionPath
String BASE_PATH_PARAM
org.apache.spark.sql.hudi.SparkAdapter sparkAdapter
byte bitmap$0
org.apache.spark.sql.execution.datasources.FilePartition filePartition
private Object readResolve()
org.slf4j.Logger log
org.apache.spark.sql.hudi.SparkAdapter sparkAdapter
boolean bitmap$0
private Object readResolve()
org.slf4j.Logger log
scala.Function1<T1,R> read
org.apache.spark.sql.types.StructType schema
private Object readResolve()
HoodieBaseRelation.BaseFileReader regularFileReader
HoodieTableSchema tableSchema
HoodieTableSchema requiredSchema
HoodieTableState tableState
long maxCompactionMemoryInBytes
org.apache.spark.broadcast.Broadcast<T> hadoopConfBroadcast
scala.collection.Seq<A> mandatoryFieldsForMerging
scala.collection.Seq<A> mandatoryFields
org.apache.spark.sql.SQLContext sqlContext
scala.Option<A> org$apache$hudi$HoodieBootstrapMORRelation$$userSchema
scala.collection.Seq<A> org$apache$hudi$HoodieBootstrapMORRelation$$globPaths
HoodieTableMetaClient metaClient
scala.collection.immutable.Map<K,V> optParams
scala.Option<A> org$apache$hudi$HoodieBootstrapMORRelation$$prunedDataSchema
byte bitmap$0
private Object readResolve()
org.apache.spark.sql.execution.datasources.PartitionedFile dataFile
scala.Option<A> skeletonFile
scala.collection.immutable.List<A> logFiles
private Object readResolve()
int index
BaseHoodieBootstrapSplit split
private Object readResolve()
HoodieBaseRelation.BaseFileReader bootstrapDataFileReader
HoodieBaseRelation.BaseFileReader bootstrapSkeletonFileReader
HoodieBaseRelation.BaseFileReader regularFileReader
HoodieTableSchema requiredSchema
org.apache.spark.sql.SQLContext sqlContext
scala.Option<A> org$apache$hudi$HoodieBootstrapRelation$$userSchema
scala.collection.Seq<A> org$apache$hudi$HoodieBootstrapRelation$$globPaths
HoodieTableMetaClient metaClient
scala.collection.immutable.Map<K,V> optParams
scala.Option<A> org$apache$hudi$HoodieBootstrapRelation$$prunedDataSchema
private Object readResolve()
org.apache.spark.sql.hudi.SparkAdapter sparkAdapter
boolean bitmap$0
org.apache.spark.sql.execution.datasources.PartitionedFile dataFile
scala.Option<A> skeletonFile
private Object readResolve()
org.apache.spark.sql.Dataset<T> df
HoodieWriteConfig config
scala.collection.immutable.Map<K,V> parameters
String recordName
String recordNameSpace
org.apache.avro.Schema writerSchema
org.apache.avro.Schema dataFileSchema
WriteOperationType operation
String instantTime
boolean preppedSparkSqlWrites
boolean preppedSparkSqlMergeInto
boolean preppedWriteOperation
private Object readResolve()
org.apache.spark.sql.SparkSession spark
HoodieTableMetaClient metaClient
scala.Option<A> schemaSpec
scala.collection.immutable.Map<K,V> options
boolean includeLogFiles
boolean shouldEmbedFileSlices
String configName
HoodieFileIndex$DataSkippingFailureMode$Val Fallback
HoodieFileIndex$DataSkippingFailureMode$Val Strict
private Object readResolve()
String value
private Object readResolve()
HoodieBaseRelation.BaseFileReader fullSchemaReader
HoodieBaseRelation.BaseFileReader requiredSchemaReader
HoodieBaseRelation.BaseFileReader requiredSchemaReaderSkipMerging
private Object readResolve()
scala.Option<A> dataFile
scala.collection.immutable.List<A> logFiles
private Object readResolve()
int index
HoodieMergeOnReadFileSplit split
private Object readResolve()
HoodieMergeOnReadBaseFileReaders fileReaders
HoodieTableSchema tableSchema
HoodieTableSchema requiredSchema
HoodieTableState tableState
String mergeType
boolean includeStartTime
String org$apache$hudi$HoodieMergeOnReadRDD$$startTimestamp
String org$apache$hudi$HoodieMergeOnReadRDD$$endTimestamp
long maxCompactionMemoryInBytes
org.apache.spark.broadcast.Broadcast<T> hadoopConfBroadcast
scala.Option<A> hoodieTableConfigOpt
scala.Option<A> asyncCompactionTriggerFn
scala.Option<A> asyncClusteringTriggerFn
scala.Option<A> extraPreCommitFn
private Object readResolve()
org.apache.hadoop.fs.Path basePath
String instantTime
String commitActionType
WriteOperationType operation
org.apache.spark.sql.SQLContext sqlContext
scala.collection.immutable.Map<K,V> org$apache$hudi$HoodieStreamingSink$$options
long latestCommittedBatchId
org.slf4j.Logger org$apache$hudi$HoodieStreamingSink$$log
scala.Option<A> tablePath
scala.Option<A> metaClient
int retryCnt
long retryIntervalMs
boolean ignoreFailedBatch
boolean disableCompaction
boolean org$apache$hudi$HoodieStreamingSink$$isAsyncCompactorServiceShutdownAbnormally
boolean org$apache$hudi$HoodieStreamingSink$$isAsyncClusteringServiceShutdownAbnormally
org.apache.spark.sql.SaveMode mode
AsyncCompactService asyncCompactorService
AsyncClusteringService asyncClusteringService
scala.Option<A> writeClient
scala.Option<A> hoodieTableConfig
org.apache.spark.sql.types.StructType structTypeSchema
String avroSchemaStr
scala.Option<A> internalSchema
private Object readResolve()
String tablePath
scala.Option<A> latestCommitTimestamp
String recordKeyField
scala.Option<A> preCombineFieldOpt
boolean usesVirtualKeys
String recordPayloadClassName
HoodieMetadataConfig metadataConfig
scala.collection.immutable.List<A> recordMergerImpls
String recordMergerStrategy
private Object readResolve()
scala.Option<A> org$apache$hudi$MergeOnReadIncrementalRelation$$userSchema
scala.Option<A> org$apache$hudi$MergeOnReadIncrementalRelation$$prunedDataSchema
TimelineUtils.HollowCommitHandling hollowCommitHandling
boolean fullTableScan
scala.collection.immutable.Seq<A> includedCommits
List<E> commitsMetadata
org.apache.hadoop.fs.FileStatus[] affectedFilesInCommits
scala.Tuple2<T1,T2> org$apache$hudi$HoodieIncrementalRelationTrait$$x$4
boolean includeStartTime
String startTs
String endTs
scala.collection.Seq<A> incrementalSpanRecordFilters
scala.collection.Seq<A> mandatoryFields
int bitmap$0
private Object readResolve()
scala.collection.Seq<A> org$apache$hudi$MergeOnReadSnapshotRelation$$globPaths
scala.Option<A> org$apache$hudi$MergeOnReadSnapshotRelation$$userSchema
scala.Option<A> org$apache$hudi$MergeOnReadSnapshotRelation$$prunedDataSchema
private Object readResolve()
scala.collection.immutable.Set<A> projectionCompatiblePayloadClasses
org.apache.spark.sql.hudi.SparkAdapter sparkAdapter
boolean bitmap$0
org.apache.spark.sql.catalyst.InternalRow internalRow
scala.collection.immutable.Map<K,V> slices
org.apache.spark.sql.catalyst.encoders.ExpressionEncoder<T> encoder
org.apache.spark.sql.catalyst.encoders.ExpressionEncoder.Deserializer<T> deserializer
org.apache.spark.sql.catalyst.encoders.ExpressionEncoder.Serializer<T> serializer
BootstrapExecutorUtils.Config cfg
TypedProperties props
org.apache.hadoop.conf.Configuration configuration
HoodieWriteConfig bootstrapConfig
String bootstrapBasePath
String command
String srcPath
String targetPath
String tableName
String tableType
String rowKey
String partitionKey
int parallelism
String schemaFile
int retry
String propsFilePath
List<E> configs
TypedProperties props
TypedProperties config
org.apache.spark.api.java.JavaSparkContext jssc
Copyright © 2024 The Apache Software Foundation. All rights reserved.