Writing and reading data from S3 (Databricks on AWS) - 7.1

Databricks

author
Talend Documentation Team
EnrichVersion
Cloud
7.1
EnrichProdName
Talend Big Data
Talend Big Data Platform
Talend Data Fabric
Talend Real-Time Big Data Platform
task
Design and Development > Designing Jobs > Hadoop distributions > Databricks
Design and Development > Designing Jobs > Serverless > Databricks
EnrichPlatform
Talend Studio

In this scenario, you create a Spark Batch Job using tS3Configuration and the Parquet components to write data on S3 and then read the data from S3.

This scenario applies only to a subscription-based Talend solution with Big data.

For more technologies supported by Talend, see Talend components.

The sample data reads as follows:
01;ychen

This data contains a user name and the ID number distributed to this user.

Note that the sample data is created for demonstration purposes only.