c

ai.chronon.spark

TableUtils

case class TableUtils(sparkSession: SparkSession) extends Product with Serializable

Linear Supertypes
Serializable, Product, Equals, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. TableUtils
  2. Serializable
  3. Product
  4. Equals
  5. AnyRef
  6. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. Protected

Instance Constructors

  1. new TableUtils(sparkSession: SparkSession)

Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##: Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. val aggregationParallelism: Int
  5. def allPartitions(tableName: String, partitionColumnsFilter: Seq[String] = Seq.empty): Seq[Map[String, String]]
  6. def archiveOrDropTableIfExists(tableName: String, timestamp: Option[Instant]): Unit
  7. def archiveTableIfExists(tableName: String, timestamp: Option[Instant]): Unit
  8. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  9. val backfillValidationEnforced: Boolean
  10. val blockingCacheEviction: Boolean
  11. val bloomFilterThreshold: Long
  12. val cacheLevel: Option[StorageLevel]
  13. val cacheLevelString: String
  14. def checkTablePermission(tableName: String, fallbackPartition: String = partitionSpec.before(partitionSpec.at(System.currentTimeMillis()))): Boolean
  15. def chunk(partitions: Set[String]): Seq[PartitionRange]
  16. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.CloneNotSupportedException]) @native()
  17. def columnSizeEstimator(dataType: DataType): Long
  18. def dropPartitionRange(tableName: String, startDate: String, endDate: String, subPartitionFilters: Map[String, String] = Map.empty): Unit
  19. def dropPartitions(tableName: String, partitions: Seq[String], partitionColumn: String = partitionColumn, subPartitionFilters: Map[String, String] = Map.empty): Unit
  20. def dropTableIfExists(tableName: String): Unit
  21. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  22. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.Throwable])
  23. def firstAvailablePartition(tableName: String, subPartitionFilters: Map[String, String] = Map.empty): Option[String]
  24. final def getClass(): Class[_ <: AnyRef]
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  25. def getColumnsFromQuery(query: String): Seq[String]
  26. def getFieldNames(schema: StructType): Seq[String]
  27. def getSchemaFromTable(tableName: String): StructType
  28. def getTableProperties(tableName: String): Option[Map[String, String]]
  29. def ifPartitionExistsInTable(tableName: String, partition: String): Boolean
  30. def insertPartitions(df: DataFrame, tableName: String, tableProperties: Map[String, String] = null, partitionColumns: Seq[String] = Seq(partitionColumn), saveMode: SaveMode = SaveMode.Overwrite, fileFormat: String = "PARQUET", autoExpand: Boolean = false, stats: Option[DfStats] = None): Unit
  31. def insertUnPartitioned(df: DataFrame, tableName: String, tableProperties: Map[String, String] = null, saveMode: SaveMode = SaveMode.Overwrite, fileFormat: String = "PARQUET"): Unit
  32. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  33. def isPartitioned(tableName: String): Boolean
  34. val joinPartParallelism: Int
  35. def lastAvailablePartition(tableName: String, subPartitionFilters: Map[String, String] = Map.empty): Option[String]
  36. def loadEntireTable(tableName: String): DataFrame
  37. lazy val logger: Logger
    Annotations
    @transient()
  38. val maxWait: Int
  39. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  40. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  41. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  42. def parsePartition(pstring: String): Map[String, String]
  43. val partitionColumn: String
  44. val partitionSpec: PartitionSpec
  45. def partitions(tableName: String, subPartitionsFilter: Map[String, String] = Map.empty): Seq[String]
  46. def preAggRepartition(rdd: RDD[Row]): RDD[Row]
  47. def preAggRepartition(df: DataFrame): DataFrame
  48. def productElementNames: Iterator[String]
    Definition Classes
    Product
  49. val sparkSession: SparkSession
  50. def sql(query: String): DataFrame
  51. final def synchronized[T0](arg0: => T0): T0
    Definition Classes
    AnyRef
  52. def tableExists(tableName: String): Boolean
  53. def unfilledRanges(outputTable: String, outputPartitionRange: PartitionRange, inputTables: Option[Seq[String]] = None, inputTableToSubPartitionFiltersMap: Map[String, Map[String, String]] = Map.empty, inputToOutputShift: Int = 0, skipFirstHole: Boolean = true): Option[Seq[PartitionRange]]
  54. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException])
  55. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException])
  56. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException]) @native()
  57. def wrapWithCache[T](opString: String, dataFrame: DataFrame)(func: => T): Try[T]

Deprecated Value Members

  1. def dropPartitionsAfterHole(inputTable: String, outputTable: String, partitionRange: PartitionRange, subPartitionFilters: Map[String, String] = Map.empty): Option[String]
    Annotations
    @deprecated
    Deprecated

Inherited from Serializable

Inherited from Product

Inherited from Equals

Inherited from AnyRef

Inherited from Any

Ungrouped