Write the sample data to S3 - Cloud - 8.0

Databricks

Version
Cloud
8.0
Language
English
Product
Talend Big Data
Talend Big Data Platform
Talend Data Fabric
Talend Real-Time Big Data Platform
Module
Talend Studio
Content
Design and Development > Designing Jobs > Hadoop distributions > Databricks
Design and Development > Designing Jobs > Serverless > Databricks
Last publication date
2024-02-20

Procedure

  1. Double-click the tFixedFlowIput component to open its Component view.

    Example

  2. Click the [...] button next to Edit schema to open the schema editor.
  3. Click the [+] button to add the schema columns as shown in this image.

    Example

  4. Click OK to validate these changes and accept the propagation prompted by the pop-up dialog box.
  5. In the Mode area, select the Use Inline Content radio button and paste the previously mentioned sample data into the Content field that is displayed.
  6. In the Field separator field, enter a semicolon (;).
  7. Double-click the tFileOutputParquet component to open its Component view.

    Example

  8. Select the Define a storage configuration component check box and then select the tS3Configuration component you configured in the previous steps.
  9. Click Sync columns to ensure that tFileOutputParquet has the same schema as tFixedFlowInput.
  10. In the Folder/File field, enter the name of the S3 folder to be used to store the sample data. For example, enter /sample_user, then as you have specified my_bucket/my_folder to use in tS3Configuration to store the business data on S3, the eventual directory on S3 becomes my_bucket/my_folder/sample_user.
  11. From the Action drop-down list, select Create if the sample_user folder does not exist yet; if this folder already exists, select Overwrite.