Loading/unloading data to/from Amazon S3 - Cloud - 8.0

Amazon Redshift

Version
Cloud
8.0
Language
English
Product
Talend Big Data
Talend Big Data Platform
Talend Data Fabric
Talend Data Integration
Talend Data Management Platform
Talend Data Services Platform
Talend ESB
Talend MDM Platform
Talend Real-Time Big Data Platform
Module
Talend Studio
Content
Data Governance > Third-party systems > Amazon services (Integration) > Amazon Redshift components
Data Quality and Preparation > Third-party systems > Amazon services (Integration) > Amazon Redshift components
Design and Development > Third-party systems > Amazon services (Integration) > Amazon Redshift components
Last publication date
2024-02-20

This scenario describes a Job that generates a delimited file and uploads the file to S3, loads data from the file on S3 to Redshift and displays the data on the console, then unloads the data from Redshift to files on S3 per slice of the Redshift cluster, and finally lists and gets the unloaded files on S3.

For more technologies supported by Talend, see Talend components.

Prerequisites:

The following context variables have been created and saved in the Repository tree view. For more information about context variables, see Using contexts and variables.

  • redshift_host: the connection endpoint URL of the Redshift cluster.

  • redshift_port: the listening port number of the database server.

  • redshift_database: the name of the database.

  • redshift_username: the username for the database authentication.

  • redshift_password: the password for the database authentication.

  • redshift_schema: the name of the schema.

  • s3_accesskey: the access key for accessing Amazon S3.

  • s3_secretkey: the secret key for accessing Amazon S3.

  • s3_bucket: the name of the Amazon S3 bucket.

Note that all context values in the above screenshot are for demonstration purposes only.