tELTTeradataOutput Standard properties
These properties are used to configure tELTTeradataOutput running in the Standard Job framework.
The Standard tELTTeradataOutput component belongs to the ELT family.
The component in this framework is available in all Talend products.
Basic settings
Action on data |
On the data of the table defined, you can perform the following operation: Insert: Add new entries to the table. If duplicates are found, Job stops. Update: Updates entries in the table. Delete: Deletes the entries which correspond to the entry flow.
Merge: Merges the incoming data in one of the following
ways in the table.
Information noteNote:
|
Schema and Edit schema |
A schema is a row description. It defines the number of fields (columns) to be processed and passed on to the next component. When you create a Spark Job, avoid the reserved word line when naming the fields.
Click Edit schema to make changes to the schema. If the current schema is of the Repository type, three options are available:
|
Where clauses (for UPDATE and DELETE only) |
Enter a clause to filter the data to be updated or deleted during the update or delete operations. This field is available when Update or Delete is selected from the Action on data drop-down list and Use WHERE conditions table is not selected. |
Use WHERE conditions table |
Select this option to invoke the Where conditions table table, where you can set filter conditions by providing values in the Column, Function, and Value fields of the Where conditions table. This option is available when Update or Delete is selected from the Action on data drop-down list. |
Where conditions table |
Add rows and enter conditions as prompted to filter the data to be updated or deleted during the update or delete operation. A data record is selected for the update or delete operation only when it matches all the conditions set in this table. In addition to the commonly used operators (such as =, >=,
<>, and so on), the Function column of this table also provides
the following operators:
This option is available when the Use WHERE conditions table option is selected. |
Default Table Name |
Enter a default name for the table, between double quotation marks. Note that the table must exist already. If it does not exist, you can use tCreateTable to create one first. For more information about tCreateTable, see tCreateTable. |
Table name from connection name is variable |
Select this check box when the name of the connection to this component is set to a variable, such as a context variable. |
Use different table name |
Select this check box to enter a different output table name, between double quotation marks, in the Table name field which appears. |
Mapping |
Specify the metadata mapping file
for the database to be used. The metadata mapping file is used for the data
type conversion between database and Java. For more information about the
metadata mapping, see the related documentation for Type mapping.
Information noteNote: You can use Hive mapping to support Databricks Delta
Lake.
|
Advanced settings
Clause SET |
Select the column names that will be used to generate the SET clauses. SET clauses will not be generated for the columns that are not selected. This field appears when Update is selected from the Action on data drop-down list in the Basic settings view. |
tStatCatcher Statistics |
Select this check box to gather the Job processing metadata at a Job level as well as at component level. |
Global Variables
Global Variables |
NB_LINE: the number of rows read by an input component or transferred to an output component. This is an After variable and it returns an integer. NB_LINE_INSERTED: the number of rows inserted. This is an After variable and it returns an integer. ERROR_MESSAGE: the error message generated by the component when an error occurs. This is an After variable and it returns a string. This variable functions only if the Die on error check box is cleared, if the component has this check box. A Flow variable functions during the execution of a component while an After variable functions after the execution of the component. To fill up a field or expression with a variable, press Ctrl+Space to access the variable list and choose the variable to use from it. For more information about variables, see Using contexts and variables. |
Usage
Usage rule |
tELTTeradataOutput is to be used along with the tELTTeradataMap. Note that the Output link to be used with these components must correspond strictly to the syntax of the table name. Information noteNote:
Note that the ELT components do not handle actual data flow but only schema information. |
Limitation |
Due to license incompatibility, one or more JARs required to use this component are not provided. You can install the missing JARs for this particular component by clicking the Install button on the Component tab view. You can also find out and add all missing JARs easily on the Modules tab in the Integration perspective of Talend Studio. For details, see Installing external modules. |