Reading the output data - Cloud - 8.0

Azure Data Lake Store

Version
Cloud
8.0
Language
English
Product
Talend Big Data
Talend Big Data Platform
Talend Data Fabric
Talend Data Integration
Talend Data Management Platform
Talend Data Services Platform
Talend ESB
Talend MDM Platform
Talend Open Studio for Big Data
Talend Open Studio for Data Integration
Talend Open Studio for ESB
Talend Real-Time Big Data Platform
Module
Talend Studio
Content
Data Governance > Third-party systems > Cloud storages > Azure components > Azure Data Lake Storage Gen2 components
Data Quality and Preparation > Third-party systems > Cloud storages > Azure components > Azure Data Lake Storage Gen2 components
Design and Development > Third-party systems > Cloud storages > Azure components > Azure Data Lake Storage Gen2 components
Last publication date
2023-06-07

Procedure

  1. In the design workspace, select tLogRow and click the Component tab to define its basic settings.
  2. Click the Sync columns button to retrieve the schema from the previous component.
  3. In the Mode selection list, select the Basic option.
  4. Switch to Run tab of the Job.
  5. Save your Job and press F6 to execute it.

Results

The rows extracted from your Delta table, and modified using a compatible preparation, are now displayed in the execution logs.

You can use the same Job, but with a different output component to send the cleaned data to the destination of your choice.