tJDBCOutput MapReduce properties

JDBC

author
Talend Documentation Team
EnrichVersion
6.5
EnrichProdName
Talend Big Data Platform
Talend Big Data
Talend Data Integration
Talend Data Services Platform
Talend ESB
Talend Open Studio for ESB
Talend Data Fabric
Talend Open Studio for Data Integration
Talend Open Studio for MDM
Talend Real-Time Big Data Platform
Talend MDM Platform
Talend Data Management Platform
Talend Open Studio for Big Data
task
Design and Development > Third-party systems > Database components > JDBC components
Data Quality and Preparation > Third-party systems > Database components > JDBC components
Data Governance > Third-party systems > Database components > JDBC components
EnrichPlatform
Talend Studio

These properties are used to configure tJDBCOutput running in the MapReduce Job framework.

The MapReduce tJDBCOutput component belongs to the MapReduce and the Databases families.

The component in this framework is available only if you have subscribed to one of the Talend solutions with Big Data.

Basic settings

Property type

Either Built-In or Repository.

 

Built-In: No property data stored centrally.

 

Repository: Select the repository file where the properties are stored.

Click this icon to open a database connection wizard and store the database connection parameters you set in the component Basic settings view.

For more information about setting up and storing database connection parameters, see Talend Studio User Guide.

JDBC URL

Specify the JDBC URL of the database to be used. For example, the JDBC URL for the Amazon Redshift database is jdbc:redshift://endpoint:port/database.

Driver JAR

Complete this table to load the driver JARs needed. To do this, click the [+] button under the table to add as many rows as needed, each row for a driver JAR, then select the cell and click the [...] button at the right side of the cell to open the Select Module wizard from which you can select the driver JAR of your interest. For example, the driver jar RedshiftJDBC41-1.1.13.1013.jar for the Redshift database.

Class Name

Enter the class name for the specified driver between double quotation marks. For example, for the RedshiftJDBC41-1.1.13.1013.jar driver, the name to be entered is com.amazon.redshift.jdbc41.Driver.

Username and Password

Enter the authentication information to the database you need to connect to.

To enter the password, click the [...] button next to the password field, and then in the pop-up dialog box enter the password between double quotes and click OK to save the settings.

Table name

Name of the table to be written. Note that only one table can be written at a time.

Schema and Edit schema

A schema is a row description. It defines the number of fields (columns) to Repository. When you create a Spark Job, avoid the reserved word line when naming the fields.

 

Built-In: You create and store the schema locally for this component only. Related topic: see Talend Studio User Guide.

 

Repository: You have already created the schema and stored it in the Repository. You can reuse it in various projects and Job designs. Related topic: see Talend Studio User Guide.

Advanced settings

Use Batch Size

When selected, enables you to define the number of lines in each processed batch.

Usage

Usage rule

In a Talend Map/Reduce Job, it is used as an end component and requires a transformation component as input link. The other components used along with it must be Map/Reduce components, too. They generate native Map/Reduce code that can be executed directly in Hadoop.

This component, along with the MapReduce family it belongs to, appears only when you are creating a Map/Reduce Job.

Note that in this documentation, unless otherwise explicitly stated, a scenario presents only Standard Jobs, that is to say traditional Talend data integration Jobs, and non Map/Reduce Jobs.

Hadoop Connection

You need to use the Hadoop Configuration tab in the Run view to define the connection to a given Hadoop distribution for the whole Job.

This connection is effective on a per-Job basis.

Limitation

We recommend using the following databases with the Map/Reduce version of this component: DB2, Informix, MSSQL, MySQL, Netezza, Oracle, Postgres, Teradata and Vertica.

It may work with other databases as well, but these may not necessarily have been tested.