c

org.checkita.dqf.config.jobconf.Sources

KafkaSourceConfig

final case class KafkaSourceConfig(id: ID, description: Option[NonEmptyString], connection: ID, topics: Seq[NonEmptyString] = Seq.empty, topicPattern: Option[NonEmptyString], startingOffsets: Option[NonEmptyString], endingOffsets: Option[NonEmptyString], persist: Option[StorageLevel], windowBy: StreamWindowing = ProcessingTime, keyFormat: KafkaTopicFormat = KafkaTopicFormat.String, valueFormat: KafkaTopicFormat = KafkaTopicFormat.String, keySchema: Option[ID] = None, valueSchema: Option[ID] = None, subtractSchemaId: Boolean = false, options: Seq[SparkParam] = Seq.empty, keyFields: Seq[NonEmptyString] = Seq.empty, metadata: Seq[SparkParam] = Seq.empty) extends SourceConfig with Product with Serializable

Kafka source configuration

id

Source ID

description

Source description

connection

Connection ID (must be a Kafka Connection)

topics

Sequence of topics to read

topicPattern

Pattern that defined topics to read

startingOffsets

Json-string defining starting offsets. If none is set, then "earliest" is used in batch jobs and "latest is used in streaming jobs.

endingOffsets

Json-string defining ending offset. Applicable only to batch jobs. If none is set then "latest" is used.

persist

Spark storage level in order to persist dataframe during job execution.

windowBy

Source of timestamp used to build windows. Applicable only for streaming jobs! Default: processingTime - uses current timestamp at the moment when Spark processes row. Other options are:

  • eventTime - uses Kafka message creation timestamp.
  • customTime(columnName) - uses arbitrary user-defined column from kafka message (column must be of TimestampType)
keyFormat

Message key format. Default: string.

valueFormat

Message value format. Default: string.

keySchema

Schema ID. Used to parse message key. Ignored when keyFormat is string. Mandatory for other formats.

valueSchema

Schema ID. Used to parse message value. Ignored when valueFormat is string. Mandatory for other formats. Used to parse kafka message value.

subtractSchemaId

Boolean flag indicating whether a kafka message schema ID encoded into its value, i.e. [1 Magic Byte] + [4 Schema ID Bytes] + [Message Value Binary Data]. If set to true, then first five bytes are subtracted before value parsing. Default: false

options

Sequence of additional Kafka options

keyFields

Sequence of key fields (columns that identify data row)

metadata

List of metadata parameters specific to this source

Linear Supertypes
Product, Equals, SourceConfig, JobConfigEntity, Serializable, Serializable, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. KafkaSourceConfig
  2. Product
  3. Equals
  4. SourceConfig
  5. JobConfigEntity
  6. Serializable
  7. Serializable
  8. AnyRef
  9. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new KafkaSourceConfig(id: ID, description: Option[NonEmptyString], connection: ID, topics: Seq[NonEmptyString] = Seq.empty, topicPattern: Option[NonEmptyString], startingOffsets: Option[NonEmptyString], endingOffsets: Option[NonEmptyString], persist: Option[StorageLevel], windowBy: StreamWindowing = ProcessingTime, keyFormat: KafkaTopicFormat = KafkaTopicFormat.String, valueFormat: KafkaTopicFormat = KafkaTopicFormat.String, keySchema: Option[ID] = None, valueSchema: Option[ID] = None, subtractSchemaId: Boolean = false, options: Seq[SparkParam] = Seq.empty, keyFields: Seq[NonEmptyString] = Seq.empty, metadata: Seq[SparkParam] = Seq.empty)

    id

    Source ID

    description

    Source description

    connection

    Connection ID (must be a Kafka Connection)

    topics

    Sequence of topics to read

    topicPattern

    Pattern that defined topics to read

    startingOffsets

    Json-string defining starting offsets. If none is set, then "earliest" is used in batch jobs and "latest is used in streaming jobs.

    endingOffsets

    Json-string defining ending offset. Applicable only to batch jobs. If none is set then "latest" is used.

    persist

    Spark storage level in order to persist dataframe during job execution.

    windowBy

    Source of timestamp used to build windows. Applicable only for streaming jobs! Default: processingTime - uses current timestamp at the moment when Spark processes row. Other options are:

    • eventTime - uses Kafka message creation timestamp.
    • customTime(columnName) - uses arbitrary user-defined column from kafka message (column must be of TimestampType)
    keyFormat

    Message key format. Default: string.

    valueFormat

    Message value format. Default: string.

    keySchema

    Schema ID. Used to parse message key. Ignored when keyFormat is string. Mandatory for other formats.

    valueSchema

    Schema ID. Used to parse message value. Ignored when valueFormat is string. Mandatory for other formats. Used to parse kafka message value.

    subtractSchemaId

    Boolean flag indicating whether a kafka message schema ID encoded into its value, i.e. [1 Magic Byte] + [4 Schema ID Bytes] + [Message Value Binary Data]. If set to true, then first five bytes are subtracted before value parsing. Default: false

    options

    Sequence of additional Kafka options

    keyFields

    Sequence of key fields (columns that identify data row)

    metadata

    List of metadata parameters specific to this source

Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  5. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()
  6. val connection: ID
  7. val description: Option[NonEmptyString]
    Definition Classes
    KafkaSourceConfig → JobConfigEntity
  8. val endingOffsets: Option[NonEmptyString]
  9. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  10. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  11. final def getClass(): Class[_]
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  12. val id: ID
    Definition Classes
    KafkaSourceConfig → JobConfigEntity
  13. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  14. val keyFields: Seq[NonEmptyString]
    Definition Classes
    KafkaSourceConfigSourceConfig
  15. val keyFormat: KafkaTopicFormat
  16. val keySchema: Option[ID]
  17. val metadata: Seq[SparkParam]
    Definition Classes
    KafkaSourceConfig → JobConfigEntity
  18. val metadataString: Option[String]
    Definition Classes
    JobConfigEntity
  19. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  20. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  21. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  22. val options: Seq[SparkParam]
  23. val persist: Option[StorageLevel]
    Definition Classes
    KafkaSourceConfigSourceConfig
  24. val startingOffsets: Option[NonEmptyString]
  25. val streamable: Boolean
    Definition Classes
    KafkaSourceConfigSourceConfig
  26. val subtractSchemaId: Boolean
  27. final def synchronized[T0](arg0: ⇒ T0): T0
    Definition Classes
    AnyRef
  28. val topicPattern: Option[NonEmptyString]
  29. val topics: Seq[NonEmptyString]
  30. val valueFormat: KafkaTopicFormat
  31. val valueSchema: Option[ID]
  32. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  33. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  34. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()
  35. val windowBy: StreamWindowing

Inherited from Product

Inherited from Equals

Inherited from SourceConfig

Inherited from JobConfigEntity

Inherited from Serializable

Inherited from Serializable

Inherited from AnyRef

Inherited from Any

Ungrouped