java.util.Map<K,V> values
java.lang.String uri
java.lang.String format
java.util.Set<E> set
org.apache.iceberg.Table table
org.apache.spark.sql.SparkSession spark
SparkTableCache tableCache
FileScanTaskSetManager manager
FileRewriteCoordinator rewriteCoordinator
org.apache.iceberg.Table table
org.apache.spark.sql.SparkSession spark
SparkTableCache tableCache
FileScanTaskSetManager manager
FileRewriteCoordinator rewriteCoordinator
double sizeEstimateMultiple
java.util.List<E> zOrderColNames
int maxOutputSize
int varLengthContribution
org.apache.spark.sql.types.DataType sqlType
int precision
int scale
org.apache.spark.sql.types.DataType sqlType
org.apache.spark.sql.types.DataType sqlType
int precision
int scale
org.apache.spark.broadcast.Broadcast<T> tableBroadcast
org.apache.iceberg.PartitionSpec spec
org.apache.iceberg.FileFormat format
boolean caseSensitive
org.apache.iceberg.DataFile[] dataFiles
org.apache.iceberg.DeleteFile[] deleteFiles
java.lang.CharSequence[] referencedDataFiles
org.apache.iceberg.DataFile[] taskFiles
scala.collection.Seq<A> catalogAndNamespace
int nestedViewDepth
int maxNestedViewDepth
scala.collection.mutable.Map<K,V> relationCache
scala.collection.Seq<A> referredTempViewNames
scala.collection.mutable.Set<A> referredTempFunctionNames
scala.Option<A> outerPlan
private java.lang.Object readResolve()
java.lang.ThreadLocal<T> value
org.apache.spark.sql.connector.catalog.CatalogManager catalogManager
org.apache.spark.sql.connector.catalog.LookupCatalog$CatalogAndMultipartIdentifier$ CatalogAndMultipartIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog$SessionCatalogAndIdentifier$ SessionCatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog$NonSessionCatalogAndIdentifier$ NonSessionCatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog$CatalogAndNamespace$ CatalogAndNamespace$module
org.apache.spark.sql.connector.catalog.LookupCatalog$CatalogAndIdentifier$ CatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog$AsTableIdentifier$ AsTableIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog$AsFunctionIdentifier$ AsFunctionIdentifier$module
private java.lang.Object readResolve()
boolean resolved
int ordinal
org.apache.spark.sql.types.DataType dataType
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
java.lang.String viewName
java.lang.String colName
int ordinal
int expectedNumCandidates
scala.Option<A> viewDDL
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression child
scala.collection.Seq<A> names
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
java.lang.String db
private java.lang.Object readResolve()
private java.lang.Object readResolve()
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan
int _hashCode
private java.lang.Object readResolve()
scala.collection.Seq<A> path
org.apache.spark.sql.types.StructField field
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.TableChange.ColumnPosition position
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.Identifier identifier
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.CatalogPlugin catalog
scala.collection.Seq<A> namespace
private java.lang.Object readResolve()
scala.collection.Seq<A> names
org.apache.spark.sql.catalyst.InternalRow ident
scala.Option<A> location
private java.lang.Object readResolve()
scala.collection.Seq<A> expressions
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.TableCatalog catalog
org.apache.spark.sql.connector.catalog.Identifier identifier
org.apache.spark.sql.connector.catalog.Table table
scala.collection.Seq<A> outputAttributes
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.Identifier identifier
boolean isTemp
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.CatalogManager catalogManager
org.apache.spark.sql.connector.catalog.LookupCatalog$CatalogAndMultipartIdentifier$ CatalogAndMultipartIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog$SessionCatalogAndIdentifier$ SessionCatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog$NonSessionCatalogAndIdentifier$ NonSessionCatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog$CatalogAndNamespace$ CatalogAndNamespace$module
org.apache.spark.sql.connector.catalog.LookupCatalog$CatalogAndIdentifier$ CatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog$AsTableIdentifier$ AsTableIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog$AsFunctionIdentifier$ AsFunctionIdentifier$module
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.catalog.SessionCatalog catalog
private java.lang.Object readResolve()
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.expressions.Expression canonicalized
org.apache.spark.sql.catalyst.expressions.Expression child
scala.collection.Seq<A> nameParts
boolean bitmap$0
private java.lang.Object readResolve()
java.lang.String message
private java.lang.Object readResolve()
scala.collection.Seq<A> rules
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression child
scala.Option<A> aliasFunc
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> nameParts
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression deserializer
scala.collection.Seq<A> inputAttributes
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression child
org.apache.spark.sql.catalyst.expressions.Expression extraction
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> name
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.connector.catalog.TableChange.ColumnPosition position
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> multipartIdentifier
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> nameParts
scala.collection.Seq<A> arguments
boolean isDistinct
scala.Option<A> filter
boolean ignoreNulls
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.FunctionIdentifier name
scala.collection.Seq<A> children
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression havingCondition
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan child
org.apache.spark.sql.catalyst.expressions.ExpressionSet validConstraints
byte bitmap$0
private java.lang.Object readResolve()
boolean expressionsResolved
boolean resolved
scala.collection.Seq<A> names
scala.collection.Seq<A> rows
byte bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> multipartIdentifier
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
int ordinal
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.immutable.Map<K,V> spec
scala.Option<A> location
boolean bitmap$0
private java.lang.Object readResolve()
java.lang.String regexPattern
scala.Option<A> table
boolean caseSensitive
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> multipartIdentifier
org.apache.spark.sql.util.CaseInsensitiveStringMap options
boolean isStreaming
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
private java.lang.Object readResolve()
boolean resolved
boolean bitmap$0
scala.Option<A> target
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> outputColumnNames
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan child
scala.collection.Seq<A> nodePatterns
org.apache.spark.sql.catalyst.expressions.ExpressionSet validConstraints
byte bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> multipartIdentifier
java.lang.String commandName
scala.Option<A> relationTypeMismatchHint
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> multipartIdentifier
java.lang.String commandName
boolean allowTempView
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.FunctionIdentifier name
scala.collection.Seq<A> functionArgs
scala.collection.Seq<A> outputNames
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> originalNameParts
org.apache.spark.sql.connector.catalog.TableCatalog catalog
org.apache.spark.sql.connector.catalog.Identifier tableName
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> multipartIdentifier
java.lang.String commandName
boolean allowTemp
scala.Option<A> relationTypeMismatchHint
boolean bitmap$0
private java.lang.Object readResolve()