Packages

c

org.checkita.dqf.context

DQStreamJob

final case class DQStreamJob(jobConfig: JobConfig, sources: Seq[Source], metrics: Seq[RegularMetricConfig], composedMetrics: Seq[ComposedMetricConfig] = Seq.empty, trendMetrics: Seq[TrendMetricConfig] = Seq.empty, loadChecks: Seq[LoadCheckConfig] = Seq.empty, checks: Seq[CheckConfig] = Seq.empty, targets: Seq[TargetConfig] = Seq.empty, schemas: Map[String, SourceSchema] = Map.empty, connections: Map[String, DQConnection] = Map.empty, storageManager: Option[DqStorageManager] = None, bufferCheckpoint: Option[ProcessorBuffer] = None)(implicit jobId: String, settings: AppSettings, spark: SparkSession, fs: FileSystem) extends Logging with Product with Serializable

Data Quality Streaming Job: provides all required functionality to calculate quality metrics, perform checks, save results and send targets for streaming data sources.

sources

Sequence of sources to process

metrics

Sequence of metrics to calculate

composedMetrics

Sequence of composed metrics to calculate

trendMetrics

Sequence of trend metrics to calculate

loadChecks

Sequence of load checks to perform

checks

Sequence of checks to perform

targets

Sequence of targets to send

schemas

Map of user-defined schemas (used for load checks evaluation)

connections

Map of connections to external systems (used to send targets)

storageManager

Data Quality Storage manager (used to save results)

bufferCheckpoint

Buffer state from last checkpoint for this job.

jobId

Implicit job ID

settings

Implicit application settings object

spark

Implicit spark session object

fs

Implicit hadoop file system object

Linear Supertypes
Serializable, Serializable, Product, Equals, Logging, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. DQStreamJob
  2. Serializable
  3. Serializable
  4. Product
  5. Equals
  6. Logging
  7. AnyRef
  8. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new DQStreamJob(jobConfig: JobConfig, sources: Seq[Source], metrics: Seq[RegularMetricConfig], composedMetrics: Seq[ComposedMetricConfig] = Seq.empty, trendMetrics: Seq[TrendMetricConfig] = Seq.empty, loadChecks: Seq[LoadCheckConfig] = Seq.empty, checks: Seq[CheckConfig] = Seq.empty, targets: Seq[TargetConfig] = Seq.empty, schemas: Map[String, SourceSchema] = Map.empty, connections: Map[String, DQConnection] = Map.empty, storageManager: Option[DqStorageManager] = None, bufferCheckpoint: Option[ProcessorBuffer] = None)(implicit jobId: String, settings: AppSettings, spark: SparkSession, fs: FileSystem)

    sources

    Sequence of sources to process

    metrics

    Sequence of metrics to calculate

    composedMetrics

    Sequence of composed metrics to calculate

    trendMetrics

    Sequence of trend metrics to calculate

    loadChecks

    Sequence of load checks to perform

    checks

    Sequence of checks to perform

    targets

    Sequence of targets to send

    schemas

    Map of user-defined schemas (used for load checks evaluation)

    connections

    Map of connections to external systems (used to send targets)

    storageManager

    Data Quality Storage manager (used to save results)

    bufferCheckpoint

    Buffer state from last checkpoint for this job.

    jobId

    Implicit job ID

    settings

    Implicit application settings object

    spark

    Implicit spark session object

    fs

    Implicit hadoop file system object

Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  5. val bufferCheckpoint: Option[ProcessorBuffer]
  6. implicit val caseSensitive: Boolean
  7. val checks: Seq[CheckConfig]
  8. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()
  9. val composedMetrics: Seq[ComposedMetricConfig]
  10. val connections: Map[String, DQConnection]
  11. implicit val dumpSize: Int
  12. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  13. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  14. final def getClass(): Class[_]
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  15. def initLogger(lvl: Level): Unit

    Initialises logger:

    Initialises logger:

    • gets log4j properties with following priority: resources directory -> working directory -> default settings
    • updates root logger level with verbosity level defined at application start
    • reconfigures Logger
    lvl

    Root logger level defined at application start

    Definition Classes
    Logging
  16. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  17. val jobConfig: JobConfig
  18. val loadChecks: Seq[LoadCheckConfig]
  19. lazy val log: Logger
    Definition Classes
    Logging
    Annotations
    @transient()
  20. def logPreMsg(): Unit
    Attributes
    protected
  21. val metrics: Seq[RegularMetricConfig]
  22. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  23. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  24. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  25. def run(): Result[Unit]

    Runs Data Quality job.

    Runs Data Quality job. Job include following stages (some of them can be omitted depending on the configuration):

    • processing load checks
    • calculating regular metrics
    • calculating composed composed metrics
    • processing checks
    • saving results
    • sending/saving targets
    returns

    Either a set of job results or a list of errors that occurred during job run.

  26. val schemas: Map[String, SourceSchema]
  27. val sources: Seq[Source]
  28. val storageManager: Option[DqStorageManager]
  29. final def synchronized[T0](arg0: ⇒ T0): T0
    Definition Classes
    AnyRef
  30. val targets: Seq[TargetConfig]
  31. val trendMetrics: Seq[TrendMetricConfig]
  32. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  33. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  34. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()

Inherited from Serializable

Inherited from Serializable

Inherited from Product

Inherited from Equals

Inherited from Logging

Inherited from AnyRef

Inherited from Any

Ungrouped