org.apache.iceberg.FileScanTask[] tasks
org.apache.iceberg.DeleteFile[] deletes
final java.lang.Object writeReplace()
org.apache.iceberg.PartitionSpec spec
org.apache.iceberg.SortOrder sortOrder
org.apache.iceberg.BaseTable table
java.lang.String name
java.util.UUID uuid
org.apache.iceberg.StructLike groupingKey
java.lang.Object[] tasks
java.lang.Object writeReplace()
org.apache.iceberg.TableOperations ops
java.lang.String name
org.apache.iceberg.metrics.MetricsReporter reporter
java.lang.Object writeReplace()
int[] fromProjectionPos
org.apache.iceberg.io.InputFile file
java.lang.String manifestPath
java.lang.Long length
int specId
org.apache.iceberg.ManifestContent content
long sequenceNumber
long minSequenceNumber
java.lang.Long snapshotId
java.lang.Integer addedFilesCount
java.lang.Integer existingFilesCount
java.lang.Integer deletedFilesCount
java.lang.Long addedRowsCount
java.lang.Long existingRowsCount
java.lang.Long deletedRowsCount
org.apache.iceberg.ManifestFile.PartitionFieldSummary[] partitions
byte[] keyMetadata
int[] fromProjectionPos
boolean containsNull
java.lang.Boolean containsNaN
byte[] lowerBound
byte[] upperBound
java.lang.String partition
org.apache.iceberg.UnboundPartitionSpec spec
org.apache.iceberg.Schema schema
int lastColumnId
org.apache.iceberg.Snapshot snapshot
org.apache.iceberg.UnboundSortOrder sortOrder
org.apache.iceberg.view.ViewVersion viewVersion
java.lang.String uuid
long snapshotId
java.util.Set<E> removed
long snapshotId
java.lang.String refName
long snapshotId
int schemaId
int versionId
int specId
int sortOrderId
java.lang.String location
org.apache.iceberg.PartitionStatisticsFile partitionStatisticsFile
java.util.Map<K,V> updated
java.lang.String refName
java.lang.Long snapshotId
org.apache.iceberg.SnapshotRefType type
java.lang.Integer minSnapshotsToKeep
java.lang.Long maxSnapshotAgeMs
java.lang.Long maxRefAgeMs
long snapshotId
org.apache.iceberg.StatisticsFile statisticsFile
int formatVersion
private void readObject(java.io.ObjectInputStream arg0)
throws java.io.IOException,
java.lang.ClassNotFoundException
java.io.IOExceptionjava.lang.ClassNotFoundExceptionprivate void writeObject(java.io.ObjectOutputStream arg0)
throws java.io.IOException
java.io.IOExceptionjava.lang.Long rowCount
java.util.Map<K,V> columnSizes
java.util.Map<K,V> valueCounts
java.util.Map<K,V> nullValueCounts
java.util.Map<K,V> nanValueCounts
java.util.Map<K,V> lowerBounds
java.util.Map<K,V> upperBounds
java.util.Map<K,V> columnModes
org.apache.iceberg.MetricsModes.MetricsMode defaultMode
int length
org.apache.iceberg.types.Types.StructType partitionType
int size
java.lang.Object[] data
java.lang.String stringSchema
int sourceId
int fieldId
java.lang.String name
org.apache.iceberg.transforms.Transform<S,T> transform
org.apache.iceberg.PartitionSpec spec
org.apache.iceberg.Schema inputSchema
org.apache.iceberg.Schema schema
int specId
org.apache.iceberg.PartitionField[] fields
int lastAssignedFieldId
org.apache.iceberg.Schema schema
boolean unpartitionedTable
org.apache.iceberg.Schema schema
int defaultSpecId
java.util.Map<K,V> specs
org.apache.iceberg.types.Types.StructType struct
int schemaId
int[] identifierFieldIds
int highestFieldId
java.lang.String name
java.lang.String location
java.lang.String metadataFileLocation
java.util.Map<K,V> properties
java.lang.String schemaAsJson
int defaultSpecId
java.util.Map<K,V> specAsJsonMap
java.lang.String sortOrderAsJson
org.apache.iceberg.io.FileIO io
org.apache.iceberg.encryption.EncryptionManager encryption
java.util.Map<K,V> refs
java.util.UUID uuid
org.apache.iceberg.MetadataTableType type
java.lang.String baseTableName
long snapshotId
org.apache.iceberg.SnapshotRefType type
java.lang.Integer minSnapshotsToKeep
java.lang.Long maxSnapshotAgeMs
java.lang.Long maxRefAgeMs
org.apache.iceberg.transforms.Transform<S,T> transform
int sourceId
org.apache.iceberg.SortDirection direction
org.apache.iceberg.NullOrder nullOrder
org.apache.iceberg.Schema schema
org.apache.iceberg.SortOrder sortOrder
org.apache.iceberg.Schema schema
int orderId
org.apache.iceberg.SortField[] fields
java.lang.String metadataFileLocation
int formatVersion
java.lang.String uuid
java.lang.String location
long lastSequenceNumber
long lastUpdatedMillis
int lastColumnId
int currentSchemaId
java.util.List<E> schemas
int defaultSpecId
java.util.List<E> specs
int lastAssignedPartitionId
int defaultSortOrderId
java.util.List<E> sortOrders
java.util.Map<K,V> properties
long currentSnapshotId
java.util.Map<K,V> schemasById
java.util.Map<K,V> specsById
java.util.Map<K,V> sortOrdersById
java.util.List<E> snapshotLog
java.util.List<E> previousFiles
java.util.List<E> statisticsFiles
java.util.List<E> partitionStatisticsFiles
java.util.List<E> changes
org.apache.iceberg.util.SerializableSupplier<T> snapshotsSupplier
java.util.List<E> snapshots
java.util.Map<K,V> snapshotsById
java.util.Map<K,V> refs
boolean snapshotsLoaded
long timestampMillis
long snapshotId
java.util.Map<K,V> values
java.lang.String uri
java.lang.String format
java.lang.String path
java.lang.Long length
java.lang.Integer partitionSpecId
java.lang.Long addedSnapshotId
java.lang.Integer content
java.lang.Long sequenceNumber
java.util.Set<E> set
org.apache.spark.sql.types.DataType sqlType
int precision
int scale
org.apache.spark.sql.types.DataType sqlType
org.apache.spark.sql.types.DataType sqlType
int precision
int scale
org.apache.iceberg.DeleteFile[] taskFiles
org.apache.iceberg.DataFile[] dataFiles
org.apache.iceberg.DeleteFile[] deleteFiles
java.lang.CharSequence[] referencedDataFiles
org.apache.iceberg.DataFile[] taskFiles
scala.collection.Seq<A> catalogAndNamespace
int nestedViewDepth
int maxNestedViewDepth
scala.collection.mutable.Map<K,V> relationCache
scala.collection.Seq<A> referredTempViewNames
scala.collection.mutable.Set<A> referredTempFunctionNames
scala.Option<A> outerPlan
private java.lang.Object readResolve()
java.lang.ThreadLocal<T> value
scala.Enumeration.Value BOTH
scala.Enumeration.Value SOURCE
scala.Enumeration.Value TARGET
long timestamp
private java.lang.Object readResolve()
java.lang.String version
private java.lang.Object readResolve()
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression identifierExpr
scala.Function1<T1,R> exprBuilder
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
int ordinal
org.apache.spark.sql.types.DataType dataType
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
java.lang.String viewName
java.lang.String colName
int ordinal
int expectedNumCandidates
scala.Option<A> viewDDL
boolean bitmap$0
private java.lang.Object readResolve()
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression child
scala.collection.Seq<A> names
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
java.lang.String name
boolean resolved
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan child
scala.collection.immutable.Map<K,V> args
private java.lang.Object readResolve()
java.lang.String message
scala.Option<A> cause
private java.lang.Object readResolve()
private java.lang.Object readResolve()
private java.lang.Object readResolve()
private java.lang.Object readResolve()
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.expressions.Expression identifierExpr
scala.Function1<T1,R> planBuilder
scala.collection.Seq<A> nodePatterns
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
int pos
boolean resolved
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan child
org.apache.spark.sql.catalyst.expressions.Expression[] args
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan relation
scala.Option<A> timestamp
scala.Option<A> version
scala.collection.Seq<A> nodePatterns
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
java.lang.Class<T> cls
scala.collection.Seq<A> outputAttrIds
private java.lang.Object readResolve()
private java.lang.Object readResolve()
scala.collection.Seq<A> org$apache$spark$sql$catalyst$analysis$ColumnResolutionHelper$$literalFunctions
scala.collection.Seq<A> path
org.apache.spark.sql.types.StructField field
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.TableChange.ColumnPosition position
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.CatalogPlugin catalog
org.apache.spark.sql.connector.catalog.Identifier identifier
private java.lang.Object readResolve()
scala.collection.Seq<A> rows
scala.collection.Seq<A> output
scala.collection.Seq<A> nodePatterns
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.CatalogPlugin catalog
scala.collection.Seq<A> namespace
private java.lang.Object readResolve()
java.lang.String name
org.apache.spark.sql.connector.catalog.functions.UnboundFunction func
private java.lang.Object readResolve()
scala.collection.Seq<A> names
org.apache.spark.sql.catalyst.InternalRow ident
scala.Option<A> location
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.FunctionCatalog catalog
org.apache.spark.sql.connector.catalog.Identifier identifier
org.apache.spark.sql.connector.catalog.functions.UnboundFunction func
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.CatalogPlugin catalog
org.apache.spark.sql.connector.catalog.Identifier identifier
org.apache.spark.sql.types.StructType viewSchema
private java.lang.Object readResolve()
scala.collection.Seq<A> expressions
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.TableCatalog catalog
org.apache.spark.sql.connector.catalog.Identifier identifier
org.apache.spark.sql.connector.catalog.Table table
scala.collection.Seq<A> outputAttributes
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.Identifier identifier
org.apache.spark.sql.types.StructType viewSchema
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.expressions.Alias alias
int index
private java.lang.Object readResolve()
private java.lang.Object readResolve()
scala.collection.Seq<A> org$apache$spark$sql$catalyst$analysis$ColumnResolutionHelper$$literalFunctions
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression canonicalized
org.apache.spark.sql.catalyst.expressions.Expression child
scala.collection.Seq<A> nameParts
boolean hasTried
scala.collection.Seq<A> nodePatterns
byte bitmap$0
private java.lang.Object readResolve()
private java.lang.Object readResolve()
java.lang.String errorSubClass
scala.collection.immutable.Map<K,V> messageParameters
private java.lang.Object readResolve()
java.lang.String errorSubClass
scala.collection.immutable.Map<K,V> messageParameters
private java.lang.Object readResolve()
java.lang.String message
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression child
scala.Option<A> aliasFunc
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> nameParts
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression deserializer
scala.collection.Seq<A> inputAttributes
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression child
org.apache.spark.sql.catalyst.expressions.Expression extraction
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> name
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.connector.catalog.TableChange.ColumnPosition position
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> nameParts
scala.collection.Seq<A> arguments
boolean isDistinct
scala.Option<A> filter
boolean ignoreNulls
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> multipartIdentifier
java.lang.String commandName
boolean requirePersistent
scala.Option<A> funcTypeMismatchHint
scala.Option<A> possibleQualifiedName
scala.collection.Seq<A> nodePatterns
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.FunctionIdentifier name
scala.collection.Seq<A> children
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.expressions.Expression havingCondition
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan child
scala.collection.Seq<A> nodePatterns
boolean resolved
org.apache.spark.sql.catalyst.expressions.ExpressionSet validConstraints
byte bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> nameParts
boolean allowTemp
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
boolean expressionsResolved
scala.collection.Seq<A> names
scala.collection.Seq<A> rows
boolean resolved
byte bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> multipartIdentifier
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
int ordinal
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.immutable.Map<K,V> spec
scala.Option<A> location
boolean bitmap$0
private java.lang.Object readResolve()
java.lang.String regexPattern
scala.Option<A> table
boolean caseSensitive
private java.lang.Object readResolve()
scala.collection.Seq<A> multipartIdentifier
org.apache.spark.sql.util.CaseInsensitiveStringMap options
boolean isStreaming
scala.collection.Seq<A> nodePatterns
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
private java.lang.Object readResolve()
boolean resolved
boolean bitmap$0
scala.Option<A> target
private java.lang.Object readResolve()
scala.collection.Seq<A> outputColumnNames
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan child
scala.collection.Seq<A> nodePatterns
boolean resolved
org.apache.spark.sql.catalyst.expressions.ExpressionSet validConstraints
byte bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> multipartIdentifier
java.lang.String commandName
scala.Option<A> relationTypeMismatchHint
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> multipartIdentifier
java.lang.String commandName
boolean allowTempView
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> name
scala.collection.Seq<A> functionArgs
scala.collection.Seq<A> nodePatterns
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> name
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan child
scala.collection.Seq<A> outputNames
scala.collection.Seq<A> nodePatterns
boolean resolved
org.apache.spark.sql.catalyst.expressions.ExpressionSet validConstraints
byte bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> multipartIdentifier
java.lang.String commandName
boolean allowTemp
scala.Option<A> relationTypeMismatchHint
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()