Scenario: Inserting bulk data into Salesforce - 6.3

Talend Open Studio for Big Data Components Reference Guide

EnrichVersion
6.3
EnrichProdName
Talend Open Studio for Big Data
task
Data Governance
Data Quality and Preparation
Design and Development
EnrichPlatform
Talend Studio

This scenario describes a four-component Job that submits bulk data in the file SalesforceAccount.txt used in Scenario 2: Gathering erroneous data while inserting data into a Salesforce object into Salesforce, executs your intended actions on the data, and ends up with displaying the Job execution results for your reference.

Setting up the Job

  1. Create a new Job and add a tFileInputDelimited component, a tSalesforceOutputBulkExec component, and two tLogRow components by typing their names on the design workspace or dropping them from the Palette.

  2. Link the tFileInputDelimited component to the tSalesforceOutputBulkExec component using a Row > Main connection.

  3. Link the tSalesforceOutputBulkExec component to the first tLogRow component using a Row > Main connection.

    Link the tSalesforceOutputBulkExec component to the second tLogRow component using a Row > Reject connection.

Configuring the components

  1. Double-click the tFileInputDelimited component to open its Basic settings view.

  2. In the File name/Stream field, browse to or enter the path to the input data file. In this example, it is D:/SalesforceAccount.txt.

  3. Click the [...] button next to Edit schema and in the pop-up schema dialog box, define the schema by adding four columns Name, ParentId, Phone and Fax of String type.

    Click OK to save the changes and close the dialog box.

Setting up the connection to Salesforce

  1. Double-click the tSalesforceOutputBulkExec component to open its Basic settings view.

  2. In the User Id, Password and Security Key fields, enter the user authentication information required to access Salesforce.

  3. Click the [...] button next to the Module Name field and in the pop-up dialog box, select the object you want to access. In this example, it is Account.

  4. In the Bulk File Path field, browse to or enter the path to the CSV file that stores the data for bulk processing.

    Note that the bulk file here to be processed must be in .csv format.

  5. Double-click the first tLogRow component to open its Basic settings view.

  6. In the Mode area, select Table (print values in cells of a table) for better readability of the results.

  7. Do the same to configure the second tLogRow component.

Executing the Job

  1. Press Ctrl + S to save the Job.

  2. Press F6 to execute the Job.

    On the console of the Run view, you can check the execution result.

    In the tLogRow_1 table, you can read the data inserted into Salesforce.

    In the tLogRow_2 table, you can read the rejected data due to the incompatibility with the Account objects you have accessed.

    Note that if you want to transform the input data before loading them into Salesforce, you need to use tSalesforceOutputBulk and tSalesforceBulkExec in cooperation to achieve this purpose. For further information about the usage of these two components, see Scenario: Inserting transformed bulk data into Salesforce.