tKafkaOutput Standard properties
These properties are used to configure tKafkaOutput running in the Standard Job framework.
The Standard tKafkaOutput component belongs to the Internet family.
The component in this framework is available in all Talend products with Big Data and in Talend Data Fabric.
Basic settings
Schema and Edit schema |
A schema is a row description. It defines the number of fields (columns) to be processed and passed on to the next component. When you create a Spark Job, avoid the reserved word line when naming the fields. Note that the schema of this component is read-only. It stores the messages to be published. |
Use an existing connection |
Select this check box and in the Component List drop-down list, select the desired connection component to reuse the connection details you already defined. |
Version |
Select the version of the Kafka cluster to be used. |
Broker list |
Enter the addresses of the broker nodes of the Kafka cluster to be used. The form of this address should be hostname:port. This information is the name and the port of the hosting node in this Kafka cluster. If you need to specify several addresses, separate them using a comma (,). |
Topic name |
Enter the name of the topic you want to publish messages to. This topic must already exist. |
Compress the data |
Select the Compress the data check box to compress the output data. |
Use SSL/TLS |
Select this check box to enable the SSL or TLS encrypted connection. Then you need to use the tSetKeystore component in the same Job to specify the encryption information. This check box is available since Kafka 0.9.0.1. |
Use Kerberos authentication |
If the Kafka cluster to be used is secured with Kerberos, select this check box to display the related parameters to be defined:
For further information about how a Kafka cluster is secured with Kerberos, see Authenticating using SASL. This check box is available since Kafka 0.9.0.1. |
Advanced settings
Kafka properties |
Add the Kafka new producer properties you need to customize to this table. For further information about the new producer properties you can define in this table, see the section describing the new producer configuration in Kafka's documentation in http://kafka.apache.org/documentation.html#newproducerconfigs. |
Set Headers |
Select this check box to add headers to messages to be sent. This feature is available to Kafka 1.1.0 onwards. |
tStatCatcher Statistics |
Select this check box to gather the processing metadata at the Job level as well as at each component level. |
Usage
Usage rule |
This component is an end component. It requires a tJavaRow or tJava component to transform the incoming data into serialized byte arrays. The following sample shows how to construct a statement to perform this transformation:
In this code, the output_row variable represents the schema of the data to be output to tKafkaOutput and output_row.serializedValue the single read-only column of that schema; the input_row variable represents the schema of the incoming data and input_row.users the input column called users to be transformed to byte arrays by the getBytes() method. |