Packages

case class DeltaParquetFileFormat(protocol: Protocol, metadata: Metadata, nullableRowTrackingFields: Boolean = false, optimizationsEnabled: Boolean = true, tablePath: Option[String] = None, isCDCRead: Boolean = false) extends ParquetFileFormat with LoggingShims with Product with Serializable

A thin wrapper over the Parquet file format to support

  • columns names without restrictions.
  • populated a column from the deletion vector of this file (if exists) to indicate whether the row is deleted or not according to the deletion vector. Consumers of this scan can use the column values to filter out the deleted rows.
Linear Supertypes
Product, Equals, LoggingShims, ParquetFileFormat, Serializable, Serializable, Logging, DataSourceRegister, FileFormat, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. DeltaParquetFileFormat
  2. Product
  3. Equals
  4. LoggingShims
  5. ParquetFileFormat
  6. Serializable
  7. Serializable
  8. Logging
  9. DataSourceRegister
  10. FileFormat
  11. AnyRef
  12. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new DeltaParquetFileFormat(protocol: Protocol, metadata: Metadata, nullableRowTrackingFields: Boolean = false, optimizationsEnabled: Boolean = true, tablePath: Option[String] = None, isCDCRead: Boolean = false)

Type Members

  1. implicit class LogStringContext extends AnyRef
    Definition Classes
    LoggingShims

Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  5. def buildReader(sparkSession: SparkSession, dataSchema: StructType, partitionSchema: StructType, requiredSchema: StructType, filters: Seq[Filter], options: Map[String, String], hadoopConf: Configuration): (PartitionedFile) ⇒ Iterator[InternalRow]
    Attributes
    protected
    Definition Classes
    FileFormat
  6. def buildReaderWithPartitionValues(sparkSession: SparkSession, dataSchema: StructType, partitionSchema: StructType, requiredSchema: StructType, filters: Seq[Filter], options: Map[String, String], hadoopConf: Configuration): (PartitionedFile) ⇒ Iterator[InternalRow]
    Definition Classes
    DeltaParquetFileFormat → ParquetFileFormat → FileFormat
  7. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()
  8. val columnMappingMode: DeltaColumnMappingMode
  9. def copyWithDVInfo(tablePath: String, optimizationsEnabled: Boolean): DeltaParquetFileFormat
  10. def createFileMetadataCol(): AttributeReference
    Definition Classes
    FileFormat
  11. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  12. def equals(other: Any): Boolean

    We sometimes need to replace FileFormat within LogicalPlans, so we have to override equals to ensure file format changes are captured

    We sometimes need to replace FileFormat within LogicalPlans, so we have to override equals to ensure file format changes are captured

    Definition Classes
    DeltaParquetFileFormat → Equals → ParquetFileFormat → AnyRef → Any
  13. def fileConstantMetadataExtractors: Map[String, (PartitionedFile) ⇒ Any]
    Definition Classes
    DeltaParquetFileFormat → FileFormat
  14. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  15. final def getClass(): Class[_]
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  16. def hasTablePath: Boolean
  17. def hashCode(): Int
    Definition Classes
    DeltaParquetFileFormat → ParquetFileFormat → AnyRef → Any
  18. def inferSchema(sparkSession: SparkSession, parameters: Map[String, String], files: Seq[FileStatus]): Option[StructType]
    Definition Classes
    ParquetFileFormat → FileFormat
  19. def initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
    Attributes
    protected
    Definition Classes
    Logging
  20. def initializeLogIfNecessary(isInterpreter: Boolean): Unit
    Attributes
    protected
    Definition Classes
    Logging
  21. val isCDCRead: Boolean
  22. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  23. def isSplitable(sparkSession: SparkSession, options: Map[String, String], path: Path): Boolean
    Definition Classes
    DeltaParquetFileFormat → ParquetFileFormat → FileFormat
  24. def isTraceEnabled(): Boolean
    Attributes
    protected
    Definition Classes
    Logging
  25. def log: Logger
    Attributes
    protected
    Definition Classes
    Logging
  26. def logDebug(entry: LogEntry, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    LoggingShims
  27. def logDebug(entry: LogEntry): Unit
    Attributes
    protected
    Definition Classes
    LoggingShims
  28. def logDebug(msg: ⇒ String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  29. def logDebug(msg: ⇒ String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  30. def logError(entry: LogEntry, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    LoggingShims
  31. def logError(entry: LogEntry): Unit
    Attributes
    protected
    Definition Classes
    LoggingShims
  32. def logError(msg: ⇒ String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  33. def logError(msg: ⇒ String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  34. def logInfo(entry: LogEntry, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    LoggingShims
  35. def logInfo(entry: LogEntry): Unit
    Attributes
    protected
    Definition Classes
    LoggingShims
  36. def logInfo(msg: ⇒ String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  37. def logInfo(msg: ⇒ String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  38. def logName: String
    Attributes
    protected
    Definition Classes
    Logging
  39. def logTrace(entry: LogEntry, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    LoggingShims
  40. def logTrace(entry: LogEntry): Unit
    Attributes
    protected
    Definition Classes
    LoggingShims
  41. def logTrace(msg: ⇒ String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  42. def logTrace(msg: ⇒ String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  43. def logWarning(entry: LogEntry, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    LoggingShims
  44. def logWarning(entry: LogEntry): Unit
    Attributes
    protected
    Definition Classes
    LoggingShims
  45. def logWarning(msg: ⇒ String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  46. def logWarning(msg: ⇒ String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  47. val metadata: Metadata
  48. def metadataSchemaFields: Seq[StructField]
    Definition Classes
    DeltaParquetFileFormat → ParquetFileFormat → FileFormat
  49. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  50. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  51. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  52. val nullableRowTrackingFields: Boolean
  53. val optimizationsEnabled: Boolean
  54. def prepareSchemaForRead(inputSchema: StructType): StructType

    prepareSchemaForRead must only be used for parquet read.

    prepareSchemaForRead must only be used for parquet read. It removes "PARQUET_FIELD_ID_METADATA_KEY" for name mapping mode which address columns by physical name instead of id.

  55. def prepareWrite(sparkSession: SparkSession, job: Job, options: Map[String, String], dataSchema: StructType): OutputWriterFactory
    Definition Classes
    DeltaParquetFileFormat → ParquetFileFormat → FileFormat
  56. val protocol: Protocol
  57. val referenceSchema: StructType
  58. def shortName(): String
    Definition Classes
    ParquetFileFormat → DataSourceRegister
  59. def supportBatch(sparkSession: SparkSession, schema: StructType): Boolean
    Definition Classes
    ParquetFileFormat → FileFormat
  60. def supportDataType(dataType: DataType): Boolean
    Definition Classes
    ParquetFileFormat → FileFormat
  61. def supportFieldName(name: String): Boolean
    Definition Classes
    DeltaParquetFileFormat → FileFormat
  62. final def synchronized[T0](arg0: ⇒ T0): T0
    Definition Classes
    AnyRef
  63. val tablePath: Option[String]
  64. def toString(): String
    Definition Classes
    ParquetFileFormat → AnyRef → Any
  65. def vectorTypes(requiredSchema: StructType, partitionSchema: StructType, sqlConf: SQLConf): Option[Seq[String]]
    Definition Classes
    ParquetFileFormat → FileFormat
  66. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  67. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  68. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()

Inherited from Product

Inherited from Equals

Inherited from LoggingShims

Inherited from ParquetFileFormat

Inherited from Serializable

Inherited from Serializable

Inherited from Logging

Inherited from DataSourceRegister

Inherited from FileFormat

Inherited from AnyRef

Inherited from Any

Ungrouped