Saving and executing the Job - Cloud - 8.0

Amazon Redshift

Version
Cloud
8.0
Language
English
Product
Talend Big Data
Talend Big Data Platform
Talend Data Fabric
Talend Data Integration
Talend Data Management Platform
Talend Data Services Platform
Talend ESB
Talend MDM Platform
Talend Open Studio for Big Data
Talend Open Studio for Data Integration
Talend Open Studio for ESB
Talend Real-Time Big Data Platform
Module
Talend Studio
Content
Data Governance > Third-party systems > Amazon services (Integration) > Amazon Redshift components
Data Quality and Preparation > Third-party systems > Amazon services (Integration) > Amazon Redshift components
Design and Development > Third-party systems > Amazon services (Integration) > Amazon Redshift components

Procedure

  1. Press Ctrl + S to save the Job.
  2. Execute the Job by pressing F6 or clicking Run on the Run tab.
    As shown above, the generated data is written into the local file redshift_bulk.txt, the file is uploaded on S3 with the new name person_load, and then the data is loaded from the file on S3 to the table person in Redshift and displayed on the console. After that, the data is unloaded from the table person in Redshift to two files person_unload_0000_part_00 and person_unload_0001_part_00 on S3 per slice of the Redshift cluster, and finally the unloaded files on S3 are listed and retrieved in the local folder.