Skip to main content Skip to complementary content

tELTGreenplumOutput Standard properties

These properties are used to configure tELTGreenplumOutput running in the Standard Job framework.

The Standard tELTGreenplumOutput component belongs to the ELT family.

The component in this framework is available in all Talend products.

Basic settings

Action on data

On the data of the table defined, you can perform the following operation:

Insert: Adds new entries to the table.

Update: Updates entries in the table.

Delete: Deletes the entries which correspond to the entry flow.

Schema and Edit Schema

A schema is a row description. It defines the number of fields (columns) to be processed and passed on to the next component. When you create a Spark Job, avoid the reserved word line when naming the fields.

Click Edit schema to make changes to the schema. If the current schema is of the Repository type, three options are available:

  • View schema: choose this option to view the schema only.

  • Change to built-in property: choose this option to change the schema to Built-in for local changes.

  • Update repository connection: choose this option to change the schema stored in the repository and decide whether to propagate the changes to all the Jobs upon completion.

    If you just want to propagate the changes to the current Job, you can select No upon completion and choose this schema metadata again in the Repository Content window.

 

Built-in: The schema is created and stored locally for this component only. For more information about a component schema in its Basic settings tab, see Basic settings tab.

 

Repository: The schema already exists and is stored in the Repository, hence can be reused. For more information about a component schema in its Basic settings tab, see Basic settings tab.

Where clauses (for UPDATE and DELETE only)

Enter a clause to filter the data to be updated or deleted during the update or delete operations.

This field is available when Update or Delete is selected from the Action on data drop-down list and Use WHERE conditions table is not selected.

Use WHERE conditions table

Select this option to invoke the Where conditions table table, where you can set filter conditions by providing values in the Column, Function, and Value fields of the Where conditions table.

This option is available when Update or Delete is selected from the Action on data drop-down list.

Where conditions table

Add rows and enter conditions as prompted to filter the data to be updated or deleted during the update or delete operation. A data record is selected for the update or delete operation only when it matches all the conditions set in this table.

In addition to the commonly used operators (such as =, >=, <>, and so on), the Function column of this table also provides the following operators:
  • BETWEEN, which filters the data using an interval given in the Value field. The interval needs to be in the format of <start> and <end>, for example: 10 and 20.
  • LIKE, which filters the data using a string given in the Value field. The string needs to be in the format of sample_string%.
  • IN, which filters the data using the elements in a set given in the Value field. The set needs to be in the format of (element1, element2, element3, ...), for example: (1,2,5).

This option is available when the Use WHERE conditions table option is selected.

Default Table Name

Enter the default table name, between double quotation marks.

Default Schema Name

Enter the default schema name,between double quotation marks.

Table name from connection name is variable

Select this check box when the name of the connection to this component is set to a variable, such as a context variable.

Use different table name

Select this check box to define a different output table name, between double quotation marks, in the Table name field which appears.

Mapping

Specify the metadata mapping file for the database to be used. The metadata mapping file is used for the data type conversion between database and Java. For more information about the metadata mapping, see the related documentation for Type mapping.
Information noteNote: You can use Hive mapping to support Databricks Delta Lake.

Advanced settings

tStatCatcher Statistics

Select this check box to gather the Job processing metadata at a Job level as well as at each component level.

Global Variables

Global Variables

NB_LINE: the number of rows read by an input component or transferred to an output component. This is an After variable and it returns an integer.

NB_LINE_INSERTED: the number of rows inserted. This is an After variable and it returns an integer.

ERROR_MESSAGE: the error message generated by the component when an error occurs. This is an After variable and it returns a string. This variable functions only if the Die on error check box is cleared, if the component has this check box.

A Flow variable functions during the execution of a component while an After variable functions after the execution of the component.

To fill up a field or expression with a variable, press Ctrl+Space to access the variable list and choose the variable to use from it.

For more information about variables, see Using contexts and variables.

Usage

Usage rule

tELTGreenplumOutput is to be used along with the tELTGreenplumMap. Note that the Output link to be used with these components must correspond strictly to the syntax of the table name.

Information noteNote:

Note that the ELT components do not handle actual data flow but only schema information.

Did this page help you?

If you find any issues with this page or its content – a typo, a missing step, or a technical error – please let us know!