Writing and reading data from Cloudera Kudu using a Spark Batch Job - 7.3

Kudu

Version
7.3
Language
English
Product
Talend Big Data
Talend Big Data Platform
Talend Data Fabric
Talend Real-Time Big Data Platform
Module
Talend Studio
Content
Data Governance > Third-party systems > Database components (Integration) > Kudu components
Data Quality and Preparation > Third-party systems > Database components (Integration) > Kudu components
Design and Development > Third-party systems > Database components (Integration) > Kudu components
Last publication date
2024-02-21

In this scenario, you create a Spark Batch Job using the Kudu components to partition and write data in a Kudu table and then read some of the data from Kudu.

This scenario applies only to Talend products with Big Data.

For more technologies supported by Talend, see Talend components.

The sample data reads as follows:
01;ychen;30
02;john;40
03;yoko;20
04;tom;60
05;martin;50

This data contains the names of some persons, the ID numbers distributed to these persons and their ages.

The ages are made distinct one from another on purpose because this column is the primary key column and is used for range partitioning in this scenario.

Note that the sample data is created for demonstration purposes only.

Prerequisites:
  • Ensure that the Spark cluster and the Cloudera Kudu database to be used have been properly installed and are running.

  • Ensure that the client machine on which the Talend Jobs are executed can recognize the host names of the nodes of the Hadoop cluster to be used. For this purpose, add the IP address/hostname mapping entries for the services of that Hadoop cluster in the hosts file of the client machine.

    For example, if the host name of the Hadoop Namenode server is talend-cdh550.weave.local and its IP address is 192.168.x.x, the mapping entry reads 192.168.x.x talend-cdh550.weave.local.

  • If the cluster to be used is secured with Kerberos, ensure that you have properly installed and configured Kerberos on the machine on which your Talend Job is executed. Depending on the Kerberos mode being used, the Kerberos kinit ticket or keytab must be available on that machine.

    For further information, search for how to use Kerberos in the Studio with Big Data on Talend Help Center (https://help.talend.com).

Recommendation:
  • Define the Hadoop connection metadata from the Hadoop cluster node in Repository. This way, you can not only reuse this connection in different Jobs but also ensure that the connection to your Hadoop cluster has been properly set up and is working well when you use this connection in your Jobs.

    For further information about setting up a reusable Hadoop connection, see Centralizing a Hadoop connection of the Talend Studio User Guide.

    For further information about setting up a reusable Hadoop connection, search for centralizing a Hadoop connection on Talend Help Center (https://help.talend.com).