Skip to main content

Using HDFS components to work with Azure Data Lake Storage (ADLS)

This scenario describes how to use the HDFS components to read data from and write data to Azure Data Lake Storage.

For more technologies supported by Talend, see Talend components.

This scenario applies only to Talend products with Big Data.

  • tFixedFlowInput: it provides sample data to the Job.

  • tHDFSOutput: it writes sample data to Azure Data Lake Store.

  • tHDFSInput: it reads sample data from Azure Data Lake Store.

  • tLogRow: it displays the output of the Job on the console of the Run view of the Job.

Did this page help you?

If you find any issues with this page or its content – a typo, a missing step, or a technical error – let us know how we can improve!