These properties are used to configure tAzureStorageOutputTable running in the Standard Job framework.
The Standard tAzureStorageOutputTable component belongs to the Cloud family.
The component in this framework is available in all Talend products with Big Data and in Talend Data Fabric.
Basic settings
Property Type |
Select the way the connection details will be set.
This property is not available when other connection component is selected from the Connection Component drop-down list. |
Connection Component |
Select the component whose connection details will be used to set up the connection to Azure storage from the drop-down list. |
Account Name |
Enter the name of the storage account you need to access. A storage account name can be found in the Storage accounts dashboard of the Microsoft Azure Storage system to be used. Ensure that the administrator of the system has granted you the appropriate access permissions to this storage account. |
Account Key |
Enter the key associated with the storage account you need to access. Two keys are available for each account and by default, either of them can be used for this access. |
Protocol |
Select the protocol for this connection to be created. |
Use Azure Shared Access Signature |
Select this check box to use a shared access signature (SAS) to access the storage resources without need for the account key. For more information, see Using Shared Access Signatures (SAS). In the Azure Shared Access Signature field displayed,
enter your account SAS URL between double quotation marks. You can get the
SAS URL for each allowed service on Microsoft Azure portal after generating
SAS. The SAS URL format is
Note that the SAS has valid period, you can set the start time at which the SAS becomes valid and the expiry time after which the SAS is no longer valid when generating it, and you need to make sure your SAS is still valid when running your Job. |
Table name |
Specify the name of the table into which the entities will be written. |
Schema and Edit schema |
A schema is a row description. It defines the number of fields (columns) to be processed and passed on to the next component. When you create a Spark Job, avoid the reserved word line when naming the fields.
Click Edit schema to make changes to the schema. If the current schema is of the Repository type, three options are available:
|
Partition Key |
Select the schema column that holds the partition key value from the drop-down list. |
Row Key |
Select the schema column that holds the row key value from the drop-down list. |
Action on data |
Select an action to be performed on data of the table defined.
For performance reasons, the incoming data is processed in parallel and in random order. Therefore, it is not recommended to perform any order-sensitive data operation (for example, insert or replace) if there are duplicated rows in your data. |
Action on table |
Select an operation to be performed on the table defined.
|
Process in batch |
Select this check box to process the input entities in batch. Note that the entities to be processed in batch should belong to the same partition group, which means, they should have the same partition key value. |
Die on error |
Select the check box to stop the execution of the Job when an error occurs. |
Advanced settings
Name mappings |
Complete this table to map the column name of the component schema with the property name of the Azure table entity if they are different.
For example, if there are three schema columns CompanyID, EmployeeID, and EmployeeName that are used to feed the values for the PartitionKey, RowKey, and Name entity properties respectively, then you need to add the following rows for the mapping when writing data into the Azure table.
|
tStatCatcher Statistics |
Select this check box to gather the Job processing metadata at the Job level as well as at each component level. |
Global variables
NB_LINE |
The number of rows processed. This is an After variable and it returns an integer. |
NB_SUCCESS |
The number of rows successfully processed. This is an After variable and it returns an integer. |
NB_REJECT |
The number of rows rejected. This is an After variable and it returns an integer. |
ERROR_MESSAGE |
The error message generated by the component when an error occurs. This is an After variable and it returns a string. |
Usage
Usage rule |
This component is usually used as an end component of a Job or Subjob and it always needs an input link. |