Skip to main content Skip to complementary content

tBigtableOutput Standard properties

These properties are used to configure tBigtableOutput running in the Standard Job framework.

The Standard tBigtableOutput component belongs to the Cloud family.

The component in this framework is available in all Talend products.

Basic settings

Property Type

Select the way the connection details will be set.

  • Built-In: The connection details will be set locally for this component. You need to specify the values for all related connection properties manually.

  • Repository: The connection details stored centrally in Repository > Metadata will be reused by this component.

    You need to click the [...] button next to it and in the pop-up Repository Content dialog box, select the connection details to be reused, and all related connection properties will be automatically filled in.

This option is not available when you select the Use an existing connection option and select a component from the Component List drop-down list.

Use an existing connection

Select this check box and in the Component List drop-down list, select the desired connection component to reuse the connection details you already defined.

Project name

Enter the unique identifier of your Google Cloud Platform project. This information is available on the Dashboard page of your Google Cloud console.

Instance ID

Enter the permanent identifier of the Bigtable instance.

Authentication method Select the mode to be used to authenticate to your project.
  • Service account key: authenticate using a Google account that is associated with your Google Cloud Platform project. When selecting this mode, the parameter to be defined is Google credentials.
  • Application Default Credentials: authenticate using the Application Default Credentials. When selecting this mode, no additional parameters need to be defined as credentials are automatically found based on the application environment.

To know more about the Google Cloud authentication process, read the Google Cloud documentation.

Google credentials

Paste the content of the JSON file that contains your service account key.

This property is only available when you authenticate using Service account key.

Table name

Enter the name of the Bigtable table you need to transfer data to.


A schema is a row description. It defines the number of fields (columns) to be processed and passed on to the next component. When you create a Spark Job, avoid the reserved word line when naming the fields.

  • Built-In: You create and store the schema locally for this component only.

  • Repository: You have already created the schema and stored it in the Repository. You can reuse it in various projects and Job designs.

Edit schema

Click Edit schema to make changes to the schema. If you make changes, the schema automatically becomes built-in.

Action on data

Select the action to be performed from the drop-down list when transferring data to the target table:

  • Insert: inserts new items from the input flow.

  • Delete: removes existing items according to the input flow.

Advanced settings

tStatCatcher Statistics

Select this check box to gather the Job processing metadata at the Job level as well as at each component level.

Max batch size

Set the maximum number of lines allowed in each batch.

Global Variables


The error message generated by the component when an error occurs. This is an After variable and it returns a string.


The number of rows processed. This is an After variable and it returns an integer.


Usage rule

tBigtableOuput is used as an end component and requires an input link such as tBigtableConnection.

Did this page help you?

If you find any issues with this page or its content – a typo, a missing step, or a technical error – let us know how we can improve!