Skip to main content
Pentaho Documentation

Kafka Producer

The Kafka Producer allows you to publish messages in near-real-time across worker nodes where multiple, subscribed members have access. A Kafka Producer step publishes a stream of records to one Kafka topic.

General

Enter the following information in the transformation step name field.

  • Step name: Specifies the unique name of the transformation on the canvas. The Step Name is set to Kafka Producer by default.

Options

The Kafka Producer step features a Kafka connection setup tab and a configuration property options tab. Each tab is described below.

Setup Tab

PDI_TransStep_Dialog_KafkaProducer.png

Fill in the following fields.

Option Description

Connection

Select a connection type:

  • Direct - Specify the Bootstrap servers from which you want to receive the Kafka streaming data.
  • Cluster - Specify the Hadoop cluster configuration from which you want to retrieve the Kafka streaming data. In a Hadoop cluster configuration, you can specify information like host names and ports for HDFS, Job Tracker, security, and other big data cluster components. Multiple servers can be specified if these are part of the same cluster. For information on Hadoop clusters, see Set Up Pentaho to Connect to a Hadoop Cluster.

Client ID

The unique Client identifier, used to identify and set up a durable connection path to the server to make requests and to distinguish between different clients.

Topic

The category to which records are published.

Key Field

In Kafka, all messages can be keyed, allowing for messages to be distributed to partitions based on their keys in a default routing scheme. If no key is present, messages are randomly distributed to partitions.

Message Field

The individual record contained in a topic.

Options Tab

PDI_TransStep_Tab_KafkaProducer_Options.png

Use this tab to configure the Kafka producer broker sources.  For further information on these input names, see the Apache Kafka documentation site: https://kafka.apache.org/documentation/.

Metadata Injection Support

All fields of this step support metadata injection. You can use this step with ETL Metadata Injection to pass metadata to your transformation at runtime.

See Also

Kafka Consumer