tEXABulkExec Standard properties
These properties are used to configure tEXABulkExec running in the Standard Job framework.
The Standard tEXABulkExec component belongs to the Databases family.
The component in this framework is available in all Talend products.
Basic settings
Use an existing connection |
Select this check box and from the list displayed select the relevant connection component to reuse the connection details you have already defined. Information noteNote: When a Job contains the parent Job and the child Job, if you
need to share an existing connection between the two levels, for example, to share the
connection created by the parent Job with the child Job, you have to:
For an example about how to share a database connection across Job levels, see Talend Studio User Guide. |
Property Type |
Either Built-In or Repository.
|
Host |
Enter the host or host list of the EXASol database servers. EXASol can run in a cluster environment. The valid value can be a simple IP address (for example, 172.16.173.128), an IP range list (for example, 172.16.173.128..130 that represents three servers 172.16.173.128, 172.16.173.129, and 172.16.173.130), or a comma-separated host list (for example, server1,server2,server3) of the Exasol database cluster. |
Port |
Enter the listening port number of the Exasol database cluster. |
Schema |
Enter the name of the schema you want to use. |
User and Password |
Enter the user authentication data to access the Exasol database. To enter the password, click the [...] button next to the password field, and then in the pop-up dialog box enter the password between double quotes and click OK to save the settings. |
Table |
Enter the name of the table to be written. Information noteNote:
Typically the table names are stored in upper case. If you need mixed case identifiers, you have to enter the name in double quotes. For example, "\"TEST_data_LOAD\"". |
Action on table |
On the table defined, you can perform one of the following operations before running the import:
|
Schema and Edit schema |
A schema is a row description. It defines the number of fields (columns) to be processed and passed on to the next component. When you create a Spark Job, avoid the reserved word line when naming the fields.
Information noteNote:
The columns in the schema must be in the same order as they are in the CSV file. It is not necessary to fill all columns of the defined table unless the use case or table definition expects that. |
|
Click Edit schema to make changes to the schema. If the current schema is of the Repository type, three options are available:
|
Advanced settings
Additional JDBC Parameters |
Specify additional connection properties for the database connection you are creating. The properties are separated by semicolon and each property is a key-value pair, for example, encryption=1;clientname=Talend. This field is not available if the Use an existing connection check box is selected. |
Column Formats |
Specify the format for Date and numeric columns if the default can not be applied.
|
Source table columns |
If the source is a database, configure the mapping between the source columns and the target columns in this table. Specifically configuring the mapping is optional. If you set nothing here, it is assumed that the source table has the same structure as the target table.
|
Column Separator |
Enter the separator for the columns of a row in the local file. |
Column Delimiter |
Enter the delimiter that encapsulates the field content in the local file. |
Row Separator |
Enter the char used to separate the rows in the local file. |
Null representation |
Enter the string that represents a NULL value in the local file. If not specified, NULL values are represented as the empty string. |
Skip rows |
Enter the number of rows (for example, header or any other prefix rows) to be omitted. |
Encoding |
Enter the character set used in the local file. By default, it is UTF8. |
Trim column values |
Specify whether spaces are deleted at the border of CSV columns.
|
Default Date Format |
Specify the format for datetime values. By default, it is YYYY-MM-DD. |
Default Timestamp Format |
Specify the timestamp format used. By default, it is YYYY-MM-DD HH24:MI:SS.FF3. |
Thousands Separator |
Specify the character used to separate thousand groups in a numeric text value. In the numeric format, the character will be applied to the placeholder G. If the text values contain this char, you have to configure it also in the Column Formats table. Note that this setting affects the connection property NLS_NUMERIC_CHARACTERS that defines the decimal and group characters used for representing numbers. |
Decimal Separator |
Specify the character used to separate the integer part of a number from the fraction. In the numeric format, the character will be applied to the placeholder D. Note that this setting affects the connection property NLS_NUMERIC_CHARACTERS that defines the decimal and group characters used for representing numbers. |
Minimal number errors to reject the transfer |
Specify the maximum number of invalid rows allowed during the data loading process. For example, the value 2 means the loading process will stop if the third error occurs. |
Log Error Destination |
Specify the location where error messages will be stored.
|
Transfer files secure |
Select this check box to transfer the file over HTTPS instead of HTTP. |
Test mode (no statements are executed) |
Select this check box to have the component running in test mode, where no statements are executed. |
Use precision and length from schema |
Select this check box to check column values that are of numeric types (that is, Double, Float, BigDecimal, Integer, Long, and Short) against the Length setting (which sets the number of integer digits) and the Precision setting (which sets the number of decimal digits) in the schema. Only the values with neither their number of integer digits nor number of decimal digits larger than the Length setting and the Precision setting are loaded. For example, with Length set to 4 and Precision set to 3, the values 8888.8888 and 88888.888 will be dropped; the values 8888.88 and 888.888 will be loaded. |
tStatCatcher Statistics |
Select this check box to gather the Job processing metadata at the Job level as well as at each component level. |
Global Variables
Global Variables |
NB_LINE_INSERTED: the number of rows inserted. This is an After variable and it returns an integer. NB_LINE_DELETED: the number of rows deleted. This is an After variable and it returns an integer. FILENAME: the name of the file processed. This is an After variable and it returns a string. ERROR_LOG_FILE: the path to the local log file. This is an After variable and it returns a string. ERROR_MESSAGE: the error message generated by the component when an error occurs. This is an After variable and it returns a string. This variable functions only if the Die on error check box is cleared, if the component has this check box. A Flow variable functions during the execution of a component while an After variable functions after the execution of the component. To fill up a field or expression with a variable, press Ctrl + Space to access the variable list and choose the variable to use from it. For further information about variables, see Talend Studio User Guide. |
Usage
Usage rule |
This component is usually used as a standalone component. |
Dynamic settings |
Click the [+] button to add a row in the table and fill the Code field with a context variable to choose your database connection dynamically from multiple connections planned in your Job. This feature is useful when you need to access database tables having the same data structure but in different databases, especially when you are working in an environment where you cannot change your Job settings, for example, when your Job has to be deployed and executed independent of Talend Studio. The Dynamic settings table is available only when the Use an existing connection check box is selected in the Basic settings view. Once a dynamic parameter is defined, the Component List box in the Basic settings view becomes unusable. For examples on using dynamic parameters, see Reading data from databases through context-based dynamic connections and Reading data from different MySQL databases using dynamically loaded connection parameters. For more information on Dynamic settings and context variables, see Talend Studio User Guide. |