Using HDFS components to work with Azure Data Lake Storage (ADLS) - 7.3

HDFS

Version
7.3
Language
English
Product
Talend Big Data
Talend Big Data Platform
Talend Data Fabric
Talend Real-Time Big Data Platform
Module
Talend Studio
Content
Data Governance > Third-party systems > File components (Integration) > HDFS components
Data Quality and Preparation > Third-party systems > File components (Integration) > HDFS components
Design and Development > Third-party systems > File components (Integration) > HDFS components
Last publication date
2024-02-21

This scenario describes how to use the HDFS components to read data from and write data to Azure Data Lake Storage.

For more technologies supported by Talend, see Talend components.

This scenario applies only to Talend products with Big Data.

  • tFixedFlowInput: it provides sample data to the Job.

  • tHDFSOutput: it writes sample data to Azure Data Lake Store.

  • tHDFSInput: it reads sample data from Azure Data Lake Store.

  • tLogRow: it displays the output of the Job on the console of the Run view of the Job.