Skip to main content Skip to complementary content

tELTTeradataOutput Standard properties

These properties are used to configure tELTTeradataOutput running in the Standard Job framework.

The Standard tELTTeradataOutput component belongs to the ELT family.

The component in this framework is available in all Talend products.

Basic settings

Action on data

On the data of the table defined, you can perform the following operation:

Insert: Add new entries to the table. If duplicates are found, Job stops.

Update: Updates entries in the table.

Delete: Deletes the entries which correspond to the entry flow.

Merge: Merges the incoming data in one of the following ways in the table.
  • Use merge insert. Select this option to create new entries and insert the incoming data to the specified columns of the new entries if the existing entries do not match the condition. To insert the incoming data to specific columns, select the columns in the table next to this option.
  • Use merge update. Select this option to update specified columns of the existing entries that match the condition. You can select the columns to update in the table next to this option.
  • Use merge delete. Select this option to remove the entries that match the condition.
Information noteNote:
  • The three Merge actions require that at least one column in the schema is set as the key column.
  • The Merge option is available only when you have installed the 8.0.1-R2022-11 Talend Studio Monthly update or a later one delivered by Talend. For more information, check with your administrator.

Schema and Edit schema

A schema is a row description. It defines the number of fields (columns) to be processed and passed on to the next component. When you create a Spark Job, avoid the reserved word line when naming the fields.

  • Built-in: The schema is created and stored locally for this component only. For more information about a component schema in its Basic settings tab, see Basic settings tab.
  • Repository: The schema already exists and is stored in the Repository, hence can be reused. For more information about a component schema in its Basic settings tab, see Basic settings tab.

Click Edit schema to make changes to the schema. If the current schema is of the Repository type, three options are available:

  • View schema: choose this option to view the schema only.

  • Change to built-in property: choose this option to change the schema to Built-in for local changes.

  • Update repository connection: choose this option to change the schema stored in the repository and decide whether to propagate the changes to all the Jobs upon completion.

    If you just want to propagate the changes to the current Job, you can select No upon completion and choose this schema metadata again in the Repository Content window.

Where clauses (for UPDATE and DELETE only)

Enter a clause to filter the data to be updated or deleted during the update or delete operations.

This field is available when Update or Delete is selected from the Action on data drop-down list and Use WHERE conditions table is not selected.

Use WHERE conditions table

Select this option to invoke the Where conditions table table, where you can set filter conditions by providing values in the Column, Function, and Value fields of the Where conditions table.

This option is available when Update or Delete is selected from the Action on data drop-down list.

Where conditions table

Add rows and enter conditions as prompted to filter the data to be updated or deleted during the update or delete operation. A data record is selected for the update or delete operation only when it matches all the conditions set in this table.

In addition to the commonly used operators (such as =, >=, <>, and so on), the Function column of this table also provides the following operators:
  • BETWEEN, which filters the data using an interval given in the Value field. The interval needs to be in the format of <start> and <end>, for example: 10 and 20.
  • LIKE, which filters the data using a string given in the Value field. The string needs to be in the format of sample_string%.
  • IN, which filters the data using the elements in a set given in the Value field. The set needs to be in the format of (element1, element2, element3, ...), for example: (1,2,5).

This option is available when the Use WHERE conditions table option is selected.

Default Table Name

Enter a default name for the table, between double quotation marks.

Note that the table must exist already. If it does not exist, you can use tCreateTable to create one first. For more information about tCreateTable, see tCreateTable.

Table name from connection name is variable

Select this check box when the name of the connection to this component is set to a variable, such as a context variable.

Use different table name

Select this check box to enter a different output table name, between double quotation marks, in the Table name field which appears.

Mapping

Specify the metadata mapping file for the database to be used. The metadata mapping file is used for the data type conversion between database and Java. For more information about the metadata mapping, see the related documentation for Type mapping.
Information noteNote: You can use Hive mapping to support Databricks Delta Lake.

Advanced settings

Clause SET

Select the column names that will be used to generate the SET clauses.

SET clauses will not be generated for the columns that are not selected.

This field appears when Update is selected from the Action on data drop-down list in the Basic settings view.

tStatCatcher Statistics

Select this check box to gather the Job processing metadata at a Job level as well as at component level.

Global Variables

Global Variables

NB_LINE: the number of rows read by an input component or transferred to an output component. This is an After variable and it returns an integer.

NB_LINE_INSERTED: the number of rows inserted. This is an After variable and it returns an integer.

ERROR_MESSAGE: the error message generated by the component when an error occurs. This is an After variable and it returns a string. This variable functions only if the Die on error check box is cleared, if the component has this check box.

A Flow variable functions during the execution of a component while an After variable functions after the execution of the component.

To fill up a field or expression with a variable, press Ctrl+Space to access the variable list and choose the variable to use from it.

For more information about variables, see Using contexts and variables.

Usage

Usage rule

tELTTeradataOutput is to be used along with the tELTTeradataMap. Note that the Output link to be used with these components must correspond strictly to the syntax of the table name.

Information noteNote:

Note that the ELT components do not handle actual data flow but only schema information.

Limitation

Due to license incompatibility, one or more JARs required to use this component are not provided. You can install the missing JARs for this particular component by clicking the Install button on the Component tab view. You can also find out and add all missing JARs easily on the Modules tab in the Integration perspective of Talend Studio. For details, see Installing external modules.

Did this page help you?

If you find any issues with this page or its content – a typo, a missing step, or a technical error – let us know how we can improve!