class MetricsJob extends SparkJob
- Alphabetic
- By Inheritance
- MetricsJob
- SparkJob
- JobBase
- StrictLogging
- AnyRef
- Any
- Hide All
- Show All
- Public
- All
Instance Constructors
-
new
MetricsJob(domain: Domain, schema: Schema, stage: Stage, storageHandler: StorageHandler, schemaHandler: SchemaHandler)(implicit settings: Settings)
- domain
: Domain name
- schema
: Schema
- stage
: stage
- storageHandler
: Storage Handler
Value Members
-
final
def
!=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
##(): Int
- Definition Classes
- AnyRef → Any
-
final
def
==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
def
analyze(fullTableName: String): Any
- Attributes
- protected
- Definition Classes
- SparkJob
-
final
def
asInstanceOf[T0]: T0
- Definition Classes
- Any
-
def
clone(): AnyRef
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()
-
def
createSparkViews(views: Views, sqlParameters: Map[String, String]): Unit
- Attributes
- protected
- Definition Classes
- SparkJob
-
final
def
eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
def
equals(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
def
finalize(): Unit
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( classOf[java.lang.Throwable] )
-
final
def
getClass(): Class[_]
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
def
hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
final
def
isInstanceOf[T0]: Boolean
- Definition Classes
- Any
- def lockPath(path: String): Path
-
val
logger: Logger
- Attributes
- protected
- Definition Classes
- StrictLogging
-
def
metricsPath(path: String): Path
Function to build the metrics save path
Function to build the metrics save path
- path
: path where metrics are stored
- returns
: path where the metrics for the specified schema are stored
-
def
name: String
- Definition Classes
- MetricsJob → JobBase
-
final
def
ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
final
def
notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
final
def
notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
def
parseViewDefinition(valueWithEnv: String): (SinkType, Option[JdbcConfigName], String)
- valueWithEnv
in the form [SinkType:[configName:]]viewName
- returns
(SinkType, configName, viewName)
- Attributes
- protected
- Definition Classes
- JobBase
-
def
partitionDataset(dataset: DataFrame, partition: List[String]): DataFrame
- Attributes
- protected
- Definition Classes
- SparkJob
-
def
partitionedDatasetWriter(dataset: DataFrame, partition: List[String]): DataFrameWriter[Row]
Partition a dataset using dataset columns.
Partition a dataset using dataset columns. To partition the dataset using the ingestion time, use the reserved column names :
- comet_date
- comet_year
- comet_month
- comet_day
- comet_hour
- comet_minute These columns are renamed to "date", "year", "month", "day", "hour", "minute" in the dataset and their values is set to the current date/time.
- dataset
: Input dataset
- partition
: list of columns to use for partitioning.
- returns
The Spark session used to run this job
- Attributes
- protected
- Definition Classes
- SparkJob
-
def
registerUdf(udf: String): Unit
- Attributes
- protected
- Definition Classes
- SparkJob
- def run(dataUse: DataFrame, timestamp: Timestamp): Try[SparkJobResult]
-
def
run(): Try[JobResult]
Just to force any spark job to implement its entry point using within the "run" method
Just to force any spark job to implement its entry point using within the "run" method
- returns
: Spark Session used for the job
- Definition Classes
- MetricsJob → JobBase
-
lazy val
session: SparkSession
- Definition Classes
- SparkJob
-
implicit
val
settings: Settings
- Definition Classes
- MetricsJob → JobBase
-
lazy val
sparkEnv: SparkEnv
- Definition Classes
- SparkJob
-
final
def
synchronized[T0](arg0: ⇒ T0): T0
- Definition Classes
- AnyRef
-
def
toString(): String
- Definition Classes
- AnyRef → Any
-
def
unionDisContMetric(discreteDataset: Option[DataFrame], continuousDataset: Option[DataFrame], domain: Domain, schema: Schema, count: Long, ingestionTime: Timestamp, stageState: Stage): MetricsDatasets
Function Function that unifies discrete and continuous metrics dataframe, then write save the result to parquet
Function Function that unifies discrete and continuous metrics dataframe, then write save the result to parquet
- discreteDataset
: dataframe that contains all the discrete metrics
- continuousDataset
: dataframe that contains all the continuous metrics
- domain
: name of the domain
- schema
: schema of the initial data
- ingestionTime
: time which correspond to the ingestion
- stageState
: stage (unit / global)
-
final
def
wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()