Scenario: Inserting transformed bulk data into your Salesforce.com - 6.1

Talend Components Reference Guide

EnrichVersion
6.1
EnrichProdName
Talend Big Data
Talend Big Data Platform
Talend Data Fabric
Talend Data Integration
Talend Data Management Platform
Talend Data Services Platform
Talend ESB
Talend MDM Platform
Talend Open Studio for Big Data
Talend Open Studio for Data Integration
Talend Open Studio for Data Quality
Talend Open Studio for ESB
Talend Open Studio for MDM
Talend Real-Time Big Data Platform
task
Data Governance
Data Quality and Preparation
Design and Development
EnrichPlatform
Talend Studio

This scenario describes a six-component Job that transforms .csv data suitable for bulk processing, load them in Salesforce.com and then displays the Job execution results in the console.

This Job is composed of two steps: preparing data by transformation and processing the transformed data.

Before starting this scenario, you need to prepare the input file which offers the data to be processed by the Job. In this use case, this file is sforcebulk.txt, containing some customer information.

Then to create and execute this Job, operate as follows:

Setting up the Job

  1. Drop tFileInputDelimited, tMap, tSalesforceOutputBulk, tSalesforceBulkExec and tLogRow from the Palette onto the workspace of your studio.

  2. Use a Row > Main connection to connect tFileInputDelimited to tMap, and Row > out1 from tMap to tSalesforceOutputBulk.

  3. Use a Row > Main connection and a Row > Reject connection to connect tSalesforceBulkExec respectively to the two tLogRow components.

  4. Use a Trigger > OnSubjobOk connection to connect tFileInputDelimited and tSalesforceBulkExec.

Configuring the input component

  1. Double-click tFileInputDelimited to display its Basic settings view and define the component properties.

  2. From the Property Type list, select Repository if you have already stored the connection to the salesforce server in the Metadata node of the Repository tree view. The property fields that follow are automatically filled in. If you have not defined the server connection locally in the Repository, fill in the details manually after selecting Built-in from the Property Type list.

    For more information about how to create the delimited file metadata, see Talend Studio User Guide.

  3. Next to the File name/Stream field, click the [...] button to browse to the input file you prepared for the scenario, for example, sforcebulk.txt.

  4. From the Schema list, select Repository and then click the three-dot button to open a dialog box where you can select the repository schema you want to use for this component. If you have not defined your schema locally in the metadata, select Built-in from the Schema list and then click the three-dot button next to the Edit schema field to open the dialog box to set the schema manually. In this scenario, the schema is made of four columns: Name, ParentId, Phone and Fax.

  5. According to your input file to be used by the Job, set the other fields like Row Separator, Field Separator...

Setting up the mapping

  1. Double-click the tMap component to open its editor and set the transformation.

  2. Drop all columns from the input table to the output table.

  3. Add .toUpperCase() behind the Name column.

  4. Click OK to validate the transformation.

Defining the output path

  1. Double-click tSalesforceOutputBulk to display its Basic settings view and define the component properties.

  2. In the File Name field, type in or browse to the directory where you want to store the generated .csv data for bulk processing.

  3. Click Sync columns to import the schema from its preceding component.

Setting up the connection to the Salesforce server

  1. Double-click tSalesforceBulkExect to display its Basic settings view and define the component properties.

  2. Use the by-default URL of the Salesforce Web service or enter the URL you want to access.

  3. In the Username and Password fields, enter your username and password for the Web service.

  4. In the Bulk file path field, browse to the directory where is stored the generated .csv file by tSalesforceOutputBulk.

  5. From the Action list, select the action you want to carry out on the prepared bulk data. In this use case, insert.

  6. From the Module list, select the object you want to access, Account in this example.

  7. From the Schema list, select Repository and then click the three-dot button to open a dialog box where you can select the repository schema you want to use for this component. If you have not defined your schema locally in the metadata, select Built-in from the Schema list and then click the three-dot button next to the Edit schema field to open the dialog box to set the schema manually. In this example, edit it conforming to the schema defined previously.

Configuring the output component

  1. Double-click tLogRow_1 to display its Basic settings view and define the component properties.

  2. Click Sync columns to retrieve the schema from the preceding component.

  3. Select Table mode to display the execution result.

  4. Do the same with tLogRow_2.

Job execution

  1. Press CTRL+S to save your Job.

  2. Press F6 to execute it.

    You can check the execution result on the Run console.

    In the tLogRow_1 table, you can read the data inserted into your Salesforce.com.

    In the tLogRow_2 table, you can read the rejected data due to the incompatibility with the Account objects you have accessed.

    All the customer names are written in upper case.