Skip to main content
Version: Current

Kafka (Source)

Mode: Streaming

Description

Kafka topics are a common streaming data input to the Nussknacker scenarios. You need to define Kafka integration to make Kafka source and sink components available in the Component Palette. Check Kafka Integration page for the list of available Kafka or Kafka compatible integrations and integration configuration info.

Parameters and configuration

NameDescription
Topic nameKafka topic to write to
Schema versionSchema version

Advanced parameters

NameDescription
Event timeExpression which evaluates to the time when the event was created. For Kafka sources, creation timestamp is available in the Kafka event and can be accessed in SpEL as #inputMeta.timestamp. Check here
Max out-of-ordernessThe maximum amount of time an element is allowed to be late before being ignored when computing the result for time-based stream transformations: aggregates in time windows and joins. To read more about this mechanism see Flink documentation
IdlenessThe time period after which partition is marked as idle if no events are received from it. To read more about this mechanism see Flink documentation.

Additional considerations

Schema Registry and its role

If Schema Registry contains schema for the Kafka source topic, Nussknacker will be able to help with field names and data type related validation of SpEL expressions.

If schema for the Kafka topic is not defined or Kafka messages are transmitted in plain text and are in the JSON format, it is possible to use dynamic field access to access fields of the Kafka messages.