Skip to main content

Reading and writing Avro data from Kafka using Standard Jobs

This scenario explains how to use schema registry with deserializer and how to handle Avro data using ConsumerRecord and ProducerRecord from Kafka components in your Standard Jobs.

For more technologies supported by Talend, see Talend components.

In this scenario, you first create a Standard Job to read Avro data using ConsumerRecord that will be called the reading Job:
And then you create another Standard Job to write Avro data using ProducerRecord that will be called the writing Job:

This scenario applies only to Talend products with Big Data and Talend Data Fabric.

Did this page help you?

If you find any issues with this page or its content – a typo, a missing step, or a technical error – let us know how we can improve!