BaseFileOnlyRelation.HoodieBaseFileSplit$ HoodieBaseFileSplit$module
scala.collection.Seq<A> mandatoryFields
scala.Option<A> org$apache$hudi$BaseFileOnlyRelation$$userSchema
scala.collection.Seq<A> org$apache$hudi$BaseFileOnlyRelation$$globPaths
scala.Option<A> org$apache$hudi$BaseFileOnlyRelation$$prunedDataSchema
boolean shouldExtractPartitionValuesFromPartitionPath
boolean bitmap$0
org.apache.spark.sql.execution.datasources.FilePartition filePartition
private Object readResolve()
org.apache.log4j.Logger log
org.apache.spark.sql.hudi.SparkAdapter sparkAdapter
boolean bitmap$0
private Object readResolve()
org.apache.log4j.Logger log
scala.Function1<T1,R> read
org.apache.spark.sql.types.StructType schema
private Object readResolve()
int index
HoodieBootstrapSplit split
private Object readResolve()
HoodieBaseRelation.BaseFileReader bootstrapDataFileReader
HoodieBaseRelation.BaseFileReader bootstrapSkeletonFileReader
HoodieBaseRelation.BaseFileReader regularFileReader
HoodieTableSchema requiredSchema
org.apache.spark.sql.types.StructType skeletonSchema
scala.Option<A> org$apache$hudi$HoodieBootstrapRelation$$userSchema
scala.collection.Seq<A> org$apache$hudi$HoodieBootstrapRelation$$globPaths
scala.Option<A> org$apache$hudi$HoodieBootstrapRelation$$prunedDataSchema
scala.collection.Seq<A> mandatoryFields
boolean bitmap$0
private Object readResolve()
org.apache.spark.sql.execution.datasources.PartitionedFile dataFile
scala.Option<A> skeletonFile
private Object readResolve()
org.apache.spark.sql.SparkSession spark
HoodieTableMetaClient metaClient
scala.Option<A> schemaSpec
scala.collection.immutable.Map<K,V> options
String configName
HoodieFileIndex$DataSkippingFailureMode$Val Fallback
HoodieFileIndex$DataSkippingFailureMode$Val Strict
private Object readResolve()
String value
private Object readResolve()
HoodieBaseRelation.BaseFileReader fullSchemaReader
HoodieBaseRelation.BaseFileReader requiredSchemaReader
HoodieBaseRelation.BaseFileReader requiredSchemaReaderSkipMerging
private Object readResolve()
scala.Option<A> dataFile
scala.collection.immutable.List<A> logFiles
private Object readResolve()
int index
HoodieMergeOnReadFileSplit split
private Object readResolve()
HoodieMergeOnReadBaseFileReaders fileReaders
HoodieTableSchema tableSchema
HoodieTableSchema requiredSchema
HoodieTableState tableState
String mergeType
long maxCompactionMemoryInBytes
org.apache.spark.broadcast.Broadcast<T> hadoopConfBroadcast
org.apache.hadoop.fs.Path basePath
String instantTime
String commitActionType
WriteOperationType operation
private Object readResolve()
org.apache.spark.sql.SQLContext sqlContext
scala.collection.immutable.Map<K,V> org$apache$hudi$HoodieStreamingSink$$options
long latestCommittedBatchId
org.apache.log4j.Logger org$apache$hudi$HoodieStreamingSink$$log
scala.Option<A> tablePath
scala.Option<A> metaClient
int retryCnt
long retryIntervalMs
boolean ignoreFailedBatch
boolean org$apache$hudi$HoodieStreamingSink$$isAsyncCompactorServiceShutdownAbnormally
boolean org$apache$hudi$HoodieStreamingSink$$isAsyncClusteringServiceShutdownAbnormally
org.apache.spark.sql.SaveMode mode
AsyncCompactService asyncCompactorService
AsyncClusteringService asyncClusteringService
scala.Option<A> writeClient
scala.Option<A> hoodieTableConfig
org.apache.spark.sql.types.StructType structTypeSchema
String avroSchemaStr
scala.Option<A> internalSchema
private Object readResolve()
String tablePath
scala.Option<A> latestCommitTimestamp
String recordKeyField
scala.Option<A> preCombineFieldOpt
boolean usesVirtualKeys
String recordPayloadClassName
HoodieMetadataConfig metadataConfig
scala.collection.immutable.List<A> recordMergerImpls
String recordMergerStrategy
private Object readResolve()
scala.Option<A> org$apache$hudi$MergeOnReadIncrementalRelation$$userSchema
scala.Option<A> org$apache$hudi$MergeOnReadIncrementalRelation$$prunedDataSchema
boolean fullTableScan
scala.collection.immutable.Seq<A> includedCommits
List<E> commitsMetadata
org.apache.hadoop.fs.FileStatus[] affectedFilesInCommits
scala.collection.Seq<A> incrementalSpanRecordFilters
scala.collection.Seq<A> mandatoryFields
byte bitmap$0
private Object readResolve()
scala.collection.Seq<A> org$apache$hudi$MergeOnReadSnapshotRelation$$globPaths
scala.Option<A> org$apache$hudi$MergeOnReadSnapshotRelation$$userSchema
scala.Option<A> org$apache$hudi$MergeOnReadSnapshotRelation$$prunedDataSchema
private Object readResolve()
scala.collection.immutable.Set<A> projectionCompatiblePayloadClasses
org.apache.spark.sql.catalyst.encoders.ExpressionEncoder<T> encoder
org.apache.spark.sql.catalyst.encoders.ExpressionEncoder.Deserializer<T> deserializer
org.apache.spark.sql.catalyst.encoders.ExpressionEncoder.Serializer<T> serializer
BootstrapExecutorUtils.Config cfg
TypedProperties props
org.apache.hadoop.conf.Configuration configuration
HoodieWriteConfig bootstrapConfig
String bootstrapBasePath
String command
String srcPath
String targetPath
String tableName
String tableType
String rowKey
String partitionKey
int parallelism
String schemaFile
int retry
String propsFilePath
List<E> configs
TypedProperties props
TypedProperties config
org.apache.spark.api.java.JavaSparkContext jssc
Copyright © 2023 The Apache Software Foundation. All rights reserved.