tRunJob properties for Apache Spark Batch

System

author
Talend Documentation Team
EnrichVersion
6.4
EnrichProdName
Talend Big Data
Talend Big Data Platform
Talend Data Fabric
Talend Data Integration
Talend Data Management Platform
Talend Data Services Platform
Talend ESB
Talend MDM Platform
Talend Open Studio for Big Data
Talend Open Studio for Data Integration
Talend Open Studio for ESB
Talend Open Studio for MDM
Talend Real-Time Big Data Platform
task
Data Governance > Third-party systems > System components
Data Quality and Preparation > Third-party systems > System components
Design and Development > Third-party systems > System components
EnrichPlatform
Talend Studio

These properties are used to configure tRunJob running in the Spark Batch Job framework.

The Spark Batch tRunJob component belongs to the System family.

The component in this framework is available in all subscription-based Talend products with Big Data and Talend Data Fabric.

Basic settings

Use dynamic job

Select this check box to allow multiple Jobs to be called and processed. When this option is enabled, only the latest version of the Jobs can be called and processed. An independent process will be used to run the subjob. The Context and the Use an independent process to run subjob options disappear.

Warning:

The Use dynamic job option is not compatible with the Jobserver cache. Therefore, the execution may fail if you run a Job that contains tRunjob with this check box selected in Talend Administration Center.

Context job

This field is visible only when the Use dynamic job option is selected. Enter the name of the Job that you want to call from the list of Jobs selected.

Job

Select the Job to be called in and processed. Make sure you already executed once the Job called, beforehand, in order to ensure a smooth run through tRunJob.

Version

Select the child Job version that you want to use.

Context

If you defined contexts and variables for the Job to be run by the tRunJob, select the applicable context entry on the list.

Die on child error

Clear this check box to execute the parent Job even though there is an error when executing the child Job.

Transmit whole context

Select this check box to get all the context variables from the parent Job. Deselect it to get all the context variables from the child Job.

If this check box is selected when the parent and child Jobs have the same context variables defined:
  • variable values for the parent Job will be used during the child Job execution if no relevant values are defined in the Context Param table.

  • otherwise, values defined in the Context Param table will be used during the child Job execution.

Context Param

You can change the value of selected context parameters. Click the [+] button to add the parameters defined in the Context tab of the child Job. For more information on context parameters, see Talend Studio User Guide.

The values defined here will be used during the child Job execution even if Transmit whole context is selected.

Advanced settings

Print Parameters

Select this check box to display the internal and external parameters in the Console.

Usage

Usage rule

This component is used with no need to be connected to other components.

This component, along with the Spark Batch component Palette it belongs to, appears only when you are creating a Spark Batch Job.

Note that in this documentation, unless otherwise explicitly stated, a scenario presents only Standard Jobs, that is to say traditional Talend data integration Jobs.

Spark Connection

You need to use the Spark Configuration tab in the Run view to define the connection to a given Spark cluster for the whole Job. In addition, since the Job expects its dependent jar files for execution, you must specify the directory in the file system to which these jar files are transferred so that Spark can access these files:
  • Yarn mode: when using Google Dataproc, specify a bucket in the Google Storage staging bucket field in the Spark configuration tab; when using other distributions, use a tHDFSConfiguration component to specify the directory.

  • Standalone mode: you need to choose the configuration component depending on the file system you are using, such as tHDFSConfiguration or tS3Configuration.

This connection is effective on a per-Job basis.