tOracleInput MapReduce properties (deprecated)
These properties are used to configure tOracleInput running in the MapReduce Job framework.
The MapReduce tOracleInput component belongs to the Databases family.
The component in this framework is available in all Talend products with Big Data and Talend Data Fabric.
The MapReduce framework is deprecated from Talend 7.3 onwards. Use Talend Jobs for Apache Spark to accomplish your integration tasks.Basic settings
Property type |
Either Built-In or Repository. Built-In: No property data stored centrally. Repository: Select the repository file where the properties are stored. |
Click this icon to open a database connection wizard and store the database connection parameters you set in the component Basic settings view. For more information about setting up and storing database connection parameters, see Talend Studio User Guide. |
|
Connection type |
Drop-down list of available drivers: Oracle OCI: Select this connection type to use Oracle Call Interface with a set of C-language software APIs that provide an interface to the Oracle database. Oracle Custom: Select this connection type to access a clustered database. Oracle Service Name: Select this connection type to use the TNS alias that you give when you connect to the remote database. WALLET: Select this connection type to store credentials in an Oracle wallet. Oracle SID: Select this connection type to uniquely identify a particular database on a system. |
DB Version |
Select the Oracle version in use. |
Host |
Database server IP address. |
Port |
Listening port number of DB server. |
Database |
Name of the database. |
Oracle schema |
Oracle schema name. |
Username and Password |
DB user authentication data. To enter the password, click the [...] button next to the password field, and then in the pop-up dialog box enter the password between double quotes and click OK to save the settings. |
Schema and Edit Schema |
A schema is a row description, it defines the number of fields to be processed and passed on to the next component. The schema is either Built-in or stored remotely in the Repository. Click Edit schema to make changes to the schema. If the current schema is of the Repository type, three options are available:
|
|
Built-In: You create and store the schema locally for this component only. |
|
Repository: You have already created the schema and stored it in the Repository. You can reuse it in various projects and Job designs. |
Table name |
Database table name. |
Query type and Query |
Enter your DB query paying particularly attention to properly sequence the fields in order to match the schema definition. |
Die on error |
Select the check box to stop the execution of the Job when an error occurs. Clear the check box to skip any rows on error and complete the process for error-free rows. When errors are skipped, you can collect the rows on error using a Row > Reject link. |
Usage
Usage rule |
In a Talend Map/Reduce Job, it is used as a start component and requires a transformation component as output link. The other components used along with it must be Map/Reduce components, too. They generate native Map/Reduce code that can be executed directly in Hadoop. You need to use the Hadoop Configuration tab in the Run view to define the connection to a given Hadoop distribution for the whole Job. Note that in this documentation, unless otherwise explicitly stated, a scenario presents only Standard Jobs, that is to say traditional Talend data integration Jobs, and non Map/Reduce Jobs. |
Hadoop Connection |
You need to use the Hadoop Configuration tab in the Run view to define the connection to a given Hadoop distribution for the whole Job. This connection is effective on a per-Job basis. |