tKafkaOutput properties - 6.1

Talend Components Reference Guide

Talend Big Data
Talend Big Data Platform
Talend Data Fabric
Talend Data Integration
Talend Data Management Platform
Talend Data Services Platform
Talend ESB
Talend MDM Platform
Talend Open Studio for Big Data
Talend Open Studio for Data Integration
Talend Open Studio for Data Quality
Talend Open Studio for ESB
Talend Open Studio for MDM
Talend Real-Time Big Data Platform
Talend Studio
Data Governance
Data Quality and Preparation
Design and Development

Component family



Basic settings

Schema and Edit schema

A schema is a row description. It defines the number of fields (columns) to be processed and passed on to the next component. The schema is either Built-In or stored remotely in the Repository.

Note that the schema of this component is read-only. It stores the messages to be published.


Use an existing connection

Select this check box and in the Component List click the relevant connection component to reuse the connection details you already defined.



Select the version of the Kafka cluster to be used.


Broker list

Enter the addresses of the broker nodes of the Kafka cluster to be used.

The form of this address should be hostname:port. This information is the name and the port of the hosting node in this Kafka cluster.

If you need to specify several addresses, separate them using a comma (,).

 Topic name

Enter the name of the topic you want to publish messages to. This topic must already exist.


Compress the data

Select the Compress the data check box to compress the output data.

Advanced settings

Kafka properties

Add the Kafka new producer properties you need to customize to this table.

For further information about the new producer properties you can define in this table, see the section describing the new producer configuration in Kafka's documentation in


tStatCatcher Statistics

Select this check box to gather the processing metadata at the Job level as well as at each component level.


This component is an end component. It requires a tJavaRow or tJava component to transform the incoming data into serialized byte arrays.

The following sample shows how to construct a statement to perform this transformation:

output_row.serializedValue = input_row.users.getBytes();

In this code, the output_row variable represents the schema of the data to be output to tKafkaOutput and output_row.serializedValue the single read-only column of that schema; the input_row variable represents the schema of the incoming data and input_row.users the input column called users to be transformed to byte arrays by the getBytes() method.