Skip to main content Skip to complementary content

tRunJob MapReduce properties (deprecated)

Availability-noteDeprecated

These properties are used to configure tRunJob running in the MapReduce Job framework.

The MapReduce tRunJob component belongs to the System family.

The component in this framework is available in all Talend products with Big Data and Talend Data Fabric.

Availability-noteDeprecated
The MapReduce framework is deprecated from Talend 7.3 onwards. Use Talend Jobs for Apache Spark to accomplish your integration tasks.

Basic settings

Use dynamic job

Select this check box to allow multiple Jobs to be called and processed. When this option is enabled, only the latest version of the Jobs can be called and processed. An independent process will be used to run the subJob. The Context and the Use an independent process to run subJob options disappear.

Information noteWarning: The Use dynamic job option is not compatible with the Jobserver cache. Therefore, the execution may fail if you run a Job that contains tRunjob with this check box selected in Talend Administration Center.
Information noteWarning: This option is incompatible with the Use or register a shared DB Connection option of database connection components. When tRunJob works together with a database connection component, enabling both options will cause your Job to fail.

Context job

This field is visible only when the Use dynamic job option is selected. Enter the name of the Job that you want to call from the list of Jobs selected.

Job

Select the Job to be called in and processed. Make sure you already executed once the Job called, beforehand, in order to ensure a smooth run through tRunJob.

Version

Select the child Job version that you want to use.

Context

If you defined contexts and variables for the Job to be run by the tRunJob, select the applicable context entry on the list.

Die on child error

Clear this check box to execute the parent Job even though there is an error when executing the child Job.

Transmit whole context

Select this check box to get all the context variables from the parent Job. Deselect it to get all the context variables from the child Job.

If this check box is selected when the parent and child Jobs have the same context variables defined:
  • variable values for the parent Job will be used during the child Job execution if no relevant values are defined in the Context Param table.

  • otherwise, values defined in the Context Param table will be used during the child Job execution.

Context Param

You can change the value of selected context parameters. Click the [+] button to add the parameters defined in the Context tab of the child Job. For more information on context parameters, see Talend Studio User Guide.

The values defined here will be used during the child Job execution even if Transmit whole context is selected.

Advanced settings

Print Parameters

Select this check box to display the internal and external parameters in the Console.

Global Variables

Global Variables

ERROR_MESSAGE: the error message generated by the component when an error occurs. This is an After variable and it returns a string. This variable functions only if the Die on error check box is cleared, if the component has this check box.

CHILD_RETURN_CODE: the return code of a child Job. This is an After variable and it returns an integer.

CHILD_EXCEPTION_STACKTRACE: the exception stack trace from a child Job. This is an After variable and it returns a string.

A Flow variable functions during the execution of a component while an After variable functions after the execution of the component.

To fill up a field or expression with a variable, press Ctrl + Space to access the variable list and choose the variable to use from it.

For further information about variables, see Talend Studio User Guide.

Usage

Usage rule

In a Talend Map/Reduce Job, this component is used standalone. It generates native Map/Reduce code that can be executed directly in Hadoop.

You need to use the Hadoop Configuration tab in the Run view to define the connection to a given Hadoop distribution for the whole Job.

This connection is effective on a per-Job basis.

For further information about a Talend Map/Reduce Job, see the sections describing how to create, convert and configure a Talend Map/Reduce Job of the Talend Big Data Getting Started Guide .

Note that in this documentation, unless otherwise explicitly stated, a scenario presents only Standard Jobs, that is to say traditional Talend data integration Jobs, and non Map/Reduce Jobs.

Did this page help you?

If you find any issues with this page or its content – a typo, a missing step, or a technical error – please let us know!