These properties are used to configure tHMapFile running in the Spark Batch Job framework.
The Spark Batch tHMapFile component belongs to the Processing family.
This component is available in Talend Platform products with Big Data and in Talend Data Fabric.
Basic settings
Storage |
To connect to an HDFS installation, select the Define a storage configuration component check box and then select the name of the component to use from those available in the drop-down list. This option requires you to have previously configured the connection to the HDFS installation to be used, as described in the documentation for the tHDFSConfiguration component. If you leave the Define a storage configuration component check box unselected, you can only convert files locally. |
Configure Component |
To configure the component, click the [...] button and, in the [Component Configuration] window, perform the following actions.
|
Input |
Click the [...] button to define the path to where the input file is stored. |
Output |
Click the [...] button to define the path to where the output files will be stored. |
Action |
From the drop-down list, select:
|
Open Map Editor |
Click the [...] button to open the map for editing in the Map Editor of Talend Data Mapper . For more information, see Talend Data Mapper User Guide. |
Advanced settings
Die on error |
Select the check box to stop the execution of the Job when an error occurs. Clear the check box to skip any rows on error and complete the process for error-free rows. |
Usage
Usage rule |
This component is used with a tHDFSConfiguration component which defines the connection to the HDFS storage, or as a standalone component for mapping local files only. |