class DeltaRecordReaderWrapper extends ParquetRecordReaderWrapper
A record reader that reads data from the underlying Parquet reader and inserts partition values which don't exist in the Parquet files.
As we have verified the Hive schema in metastore is consistent with the Delta schema, the row
returned by the underlying Parquet reader will match the Delta schema except that it leaves all
partition columns as null since they are not in the raw parquet files. Hence, for the missing
partition values, we need to use the partition information in DeltaInputSplit to create the
corresponding Writables, and insert them into the corresponding positions when reading a row.
Linear Supertypes
Ordering
- Alphabetic
- By Inheritance
Inherited
- DeltaRecordReaderWrapper
- ParquetRecordReaderWrapper
- StatsProvidingRecordReader
- RecordReader
- Closeable
- AutoCloseable
- ParquetRecordReaderBase
- AnyRef
- Any
- Hide All
- Show All
Visibility
- Public
- All
Instance Constructors
- new DeltaRecordReaderWrapper(inputFormat: ParquetInputFormat[ArrayWritable], split: DeltaInputSplit, jobConf: JobConf, reporter: Reporter)
Value Members
-
final
def
!=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
##(): Int
- Definition Classes
- AnyRef → Any
-
final
def
==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
asInstanceOf[T0]: T0
- Definition Classes
- Any
-
def
clone(): AnyRef
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()
-
def
close(): Unit
- Definition Classes
- ParquetRecordReaderWrapper → RecordReader → Closeable → AutoCloseable
- Annotations
- @throws( classOf[java.io.IOException] )
-
def
createKey(): NullWritable
- Definition Classes
- ParquetRecordReaderWrapper → RecordReader
-
def
createValue(): ArrayWritable
- Definition Classes
- ParquetRecordReaderWrapper → RecordReader
-
final
def
eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
def
equals(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
def
finalize(): Unit
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( classOf[java.lang.Throwable] )
-
final
def
getClass(): Class[_]
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
def
getFiltedBlocks(): List[BlockMetaData]
- Definition Classes
- ParquetRecordReaderBase
-
def
getPos(): Long
- Definition Classes
- ParquetRecordReaderWrapper → RecordReader
- Annotations
- @throws( classOf[java.io.IOException] )
-
def
getProgress(): Float
- Definition Classes
- ParquetRecordReaderWrapper → RecordReader
- Annotations
- @throws( classOf[java.io.IOException] )
-
def
getSplit(arg0: InputSplit, arg1: JobConf): ParquetInputSplit
- Attributes
- protected[parquet]
- Definition Classes
- ParquetRecordReaderBase
- Annotations
- @throws( classOf[java.io.IOException] )
-
def
getStats(): SerDeStats
- Definition Classes
- ParquetRecordReaderBase
-
def
hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
final
def
isInstanceOf[T0]: Boolean
- Definition Classes
- Any
-
final
def
ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
def
next(key: NullWritable, value: ArrayWritable): Boolean
- Definition Classes
- DeltaRecordReaderWrapper → ParquetRecordReaderWrapper → RecordReader
-
final
def
notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
final
def
notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
def
setFilter(arg0: JobConf, arg1: MessageType): Filter
- Definition Classes
- ParquetRecordReaderBase
-
final
def
synchronized[T0](arg0: ⇒ T0): T0
- Definition Classes
- AnyRef
-
def
toString(): String
- Definition Classes
- AnyRef → Any
-
final
def
wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()