Converting java types using Map/Reduce components - 7.0

Processing (Integration)

author
Talend Documentation Team
EnrichVersion
7.0
EnrichProdName
Talend Big Data
Talend Big Data Platform
Talend Data Fabric
Talend Data Integration
Talend Data Management Platform
Talend Data Services Platform
Talend ESB
Talend MDM Platform
Talend Open Studio for Big Data
Talend Open Studio for Data Integration
Talend Open Studio for ESB
Talend Open Studio for MDM
Talend Real-Time Big Data Platform
task
Data Governance > Third-party systems > Processing components (Integration)
Data Quality and Preparation > Third-party systems > Processing components (Integration)
Design and Development > Third-party systems > Processing components (Integration)
EnrichPlatform
Talend Studio

This scenario applies only to subscription-based Talend products with Big Data.

For more technologies supported by Talend, see Talend components.

If you are a subscription-based Big Data users, you can produce the Map/Reduce version of the Job described earlier using Map/Reduce components. This Talend Map/Reduce Job generates Map/Reduce code and is run natively in Hadoop.

The sample data used in this scenario is the same as in the scenario explained earlier.
3;123;456.21

Since Talend Studio allows you to convert a Job between its Map/Reduce and Standard (Non Map/Reduce) versions, you can convert the previous scenario to create this Map/Reduce Job. This way, many components used can keep their original settings so as to reduce your workload in designing this Job.

Before starting to replicate this scenario, ensure that you have appropriate rights and permissions to access the Hadoop distribution to be used. Then proceed as follows: