tHiveWarehouseOutput properties for Apache Spark Batch
These properties are used to configure tHiveWarehouseOutput running in the Spark Batch Job framework.
The Spark Batch tHiveWarehouseOutput component belongs to the Storage family.
The component in this framework is available in all subscription-based Talend products with Big Data and Talend Data Fabric.
Basic settings
Property Type |
Select the way the connection details will be set.
|
Hive Storage Configuration | Select the tHiveWarehouseConfiguration component from which you want Spark to use the configuration details to connect to Hive. |
HDFS Storage Configuration |
Select the tHDFSConfiguration component from which you want Spark to use the configuration details to connect to a given HDFS system and transfer the dependent jar files to this HDFS system. This field is relevant only when you are using an on-premises distribution. |
Schema and Edit Schema |
A schema is a row description. It defines the number of fields (columns) to be processed and passed on to the next component. When you create a Spark Job, avoid the reserved word line when naming the fields. Always use lowercase when naming a field because the processing behind the scene could force the field names to be lowercase. Select the type of schema you want to use from the
Schema drop-down list:
Click Edit schema to make changes to the schema. If the current schema is of the Repository type, three options are available:
|
Output source | Select the type of the output data you want
tHiveWarehouseOutput to change:
|
Save mode | Select the type of changes you want to make regarding the target Hive
table:
|
Enable Hive partitions |
Select the Enable Hive partitions check box and in the Partition keys table, define partitions for the Hive table you are creating or changing. In the Partition keys table, select columns from the input schema of tHiveOutput to use as partition keys. Bear in mind that:
|
Advanced settings
Sort columns alphabetically | Select this check box to sort the schema columns in the alphabetical order. If you leave this check box clear, these columns stick to the order defined in the schema editor. |
Global Variables
Global Variables |
ERROR_MESSAGE: the error message generated by the component when an error occurs. This is an After variable and it returns a string. This variable functions only if the Die on error check box is cleared, if the component has this check box. A Flow variable functions during the execution of a component while an After variable functions after the execution of the component. To fill up a field or expression with a variable, press Ctrl+Space to access the variable list and choose the variable to use from it. For more information about variables, see Using contexts and variables. |
Usage
Usage rule |
This component is used as an end component and requires an input link. This component should use a tHiveWarehouseConfiguration component present in the same Job to connect to Hive. This component, along with the Spark Batch component Palette it belongs to, appears only when you are creating a Spark Batch Job. |