This section provides a list of core features that are installed by default and a
list of optional features that need to be installed using the Feature Manager.
A component is a functional element that performs a single data integration
operation in a Job. Only some basic Data Integration components are installed by
default.
Talend Studio
provides a range of SQL templates to simplify the most common data query and update,
schema creation and modification, and data access control tasks. It also comprises a
SQL editor which allows you to customize or design your own SQL templates to meet
less common requirements.
The metadata wizard allows you to store reusable information on databases,
files, and/or systems in the Repository tree view. The
information can be reused later to set the connection parameters of the relevant
input or output components and the data schema in a centralized manner.
This feature allows you to deploy and execute your Jobs on a remote JobServer
when you work on either a local project or on a remote one on the condition that you
are connected with Talend Administration Center.
This feature allows you to publish Jobs, Routes and Data Services (artifacts)
created in Talend Studio to
Talend Cloud and
make them available to specific or all users of Talend Management Console.
Talend Project Audit
transforms project data flows to valuable business information. It introduces an
auditing approach for evaluating various aspects of Jobs implemented in your
Talend Studio.
Talend
Metadata Bridge accelerates the implementation, maintenance and continuous
improvement of integration scenarios by allowing the synchronization, the sharing
and the conversion of metadata across the different components.
This feature allows you to create resources and use them in your Jobs for file
handling. This way, when exporting your Jobs, for example, you can pack the resource
files as Job dependencies and deploy your Jobs without having to copy the files to
the target system.
This feature allows you to create test cases to test your Jobs and Services
during Continuous Integration development to make sure they will function as
expected when they are actually executed to handle large datasets.
A validation rule is a basic or integrity rule that you can apply to metadata
items to check the validity of your data. It can be basic check for correct values
or referential integrity check, both applicable to database tables or individual
columns, file metadata or any relevant metadata item.
Data Integration > Components
Amazon DocumentDB
This feature installs Amazon DocumentDB components, including
tAmazonDocumentDBConnection, tAmazonDocumentDBInput, tAmazonDocumentDBOutput, and
tAmazonDocumentDBClose.
This feature installs Google drive components, including
tGoogleDriveConnection, tGoogleDriveCopy, tGoogleDriveCreate, tGoogleDriveDelete,
tGoogleDriveGet, tGoogleDriveList, tGoogleDrivePut.
Data Integration > Components
Google Bigtable
This feature installs Google Bigtable components, including
tBigtableConnection, tBigtableInput, tBigtableOutput, and tBigtableClose.
This feature installs MarkLogic components, including tMarkLogicBulkLoad,
tMarkLogicClose, tMarkLogicConnection, tMarkLogicInput, tMarkLogicOutput.
Data Integration > Components
Neo4j
This feature installs Neo4j components, including tNeo4JClose,
tNeo4JConnection, tNeo4JInput, tNeo4JOutput, tNeo4JRow, tNeo4JBatchOutput,
tNeo4JBatchOutputRelationship, and tNeo4JBatchSchema.
This feature installs NetSuite components, including tNetsuiteConnection,
tNetsuiteInput, tNetsuiteOutput.
Available in:
Big Data
Big Data Platform
Cloud Big Data
Cloud Big Data Platform
Cloud Data Fabric
Data Fabric
Real-Time Big Data Platform
Data Integration > Components
Available in:
Big Data
Big Data Platform
Cloud Big Data
Cloud Big Data Platform
Cloud Data Fabric
Data Fabric
Real-Time Big Data Platform
NoSQL / Big Data
Available in:
Big Data
Big Data Platform
Cloud Big Data
Cloud Big Data Platform
Cloud Data Fabric
Data Fabric
Real-Time Big Data Platform
This feature installs the NoSQL / Big Data components, including
Cassandra, CosmosDB, DBFS, DynamoDB, ELTHive, HBase, HCatalog, HDFS, Hive,
Iceberg, Impala, Kafka, MapRDB, MongoDB, Neo4j, SAP HANA, and Sqoop related
components.
Available in:
Big Data
Big Data Platform
Cloud Big Data
Cloud Big Data Platform
Cloud Data Fabric
Data Fabric
Real-Time Big Data Platform
Data Integration > Components
Available in:
Big Data
Big Data Platform
Cloud Big Data
Cloud Big Data Platform
Cloud Data Fabric
Data Fabric
Real-Time Big Data Platform
Partitioner
Available in:
Big Data
Big Data Platform
Cloud Big Data
Cloud Big Data Platform
Cloud Data Fabric
Data Fabric
Real-Time Big Data Platform
This feature installs the Partitioner components, including tCollector, tDepartitoner, tPartitioner, and tRecollector.
This feature installs Splunk components, including
tSplunkEventCollector.
Available in:
Big Data
Big Data Platform
Data Fabric
Data Integration
Data Management Platform
Data Services Platform
ESB
MDM Platform
Real-Time Big Data Platform
Data Integration > Components
Available in:
Big Data
Big Data Platform
Data Fabric
Data Integration
Data Management Platform
Data Services Platform
ESB
MDM Platform
Real-Time Big Data Platform
Talend Data Preparation
Available in:
Big Data
Big Data Platform
Data Fabric
Data Integration
Data Management Platform
Data Services Platform
ESB
MDM Platform
Real-Time Big Data Platform
The Talend Data Preparation
components apply preparations, create datasets in Talend Data Preparation
or create flows with data from Talend Data Preparation
datasets.
Available in:
Big Data
Big Data Platform
Data Fabric
Data Integration
Data Management Platform
Data Services Platform
ESB
MDM Platform
Real-Time Big Data Platform
Data Integration > Components
Available in:
Big Data
Big Data Platform
Data Fabric
Data Integration
Data Management Platform
Data Services Platform
ESB
MDM Platform
Real-Time Big Data Platform
Talend Data Stewardship
Available in:
Big Data
Big Data Platform
Data Fabric
Data Integration
Data Management Platform
Data Services Platform
ESB
MDM Platform
Real-Time Big Data Platform
The Talend Data Stewardship
components load data into Talend Data Stewardship
campaigns and retrieve or delete data in the form of tasks in Talend Data Stewardshipcampaigns.
This feature helps you set up a CDC environment on a dedicated database
connection, which can quickly identify and capture data that has been added to,
updated in, or removed from database tables and make this change data available for
future use by applications or individuals. It is available for Oracle, MySQL, DB2,
PostgreSQL, Sybase, MS SQL Server, Informix, Ingres, Teradata, and AS/400.
The EDIFACT metadata wizard helps you create a schema to be used for the
tExtractEDIField component to read and extract data from UN/EDIFACT message
files.
The SAP metadata wizard helps you create a connection to an SAP BW system and
an SAP HANA database and store this connection in the
Repository tree view.
Data Service combines data integration with Web services and enables the
graphical design of a Service which includes a WSDL file and one or more Jobs that
addresses all of the different sources and targets required to publish the Web
service. Route defines how messages will be moved from one service (or endpoint) to
another.
Available in:
Big Data
Big Data Platform
Cloud Big Data
Cloud Big Data Platform
Cloud Data Fabric
Data Fabric
Real-Time Big Data Platform
Big Data
Available in:
Big Data
Big Data Platform
Cloud Big Data
Cloud Big Data Platform
Cloud Data Fabric
Data Fabric
Real-Time Big Data Platform
Spark Batch
Available in:
Big Data
Big Data Platform
Cloud Big Data
Cloud Big Data Platform
Cloud Data Fabric
Data Fabric
Real-Time Big Data Platform
This feature enables you to create Spark Batch Jobs.
Available in:
Cloud Data Fabric
Data Fabric
Real-Time Big Data Platform
Big Data
Available in:
Cloud Data Fabric
Data Fabric
Real-Time Big Data Platform
Spark Streaming
Available in:
Cloud Data Fabric
Data Fabric
Real-Time Big Data Platform
This feature enables you to create Spark Streaming Jobs.
Big Data > Distributions
Amazon EMR 5.29.0
This feature enables you to run your Spark Jobs on the Amazon EMR 5.29.0
distribution.
Big Data > Distributions
Amazon EMR 6.2.0
This feature enables you to run your Spark Jobs on the Amazon EMR 6.2.0
distribution.
Big Data > Distributions
Azure Synapse
This feature enables you to run your Spark Jobs on Azure Synapse Analytics with
Apache Spark pools as a distribution.
Big Data > Distributions
Cloudera CDH Dynamic Distribution
This feature enables you to run your Spark Jobs on Cloudera CDH using either
Static (CDH 6.1, CDH 6.2 and CDH 6.3) or Dynamic distributions.
Big Data > Distributions
Cloudera Data Platform Dynamic Distribution
This feature enables you to run your Spark Jobs on Cloudera Data Platform using
either Static (CDP 7.1) or Dynamic distributions.
Big Data > Distributions
Databricks 5.5
This feature enables you to run your Spark Jobs on the Databricks 5.5
distribution.
Big Data > Distributions
Databricks 6.4
This feature enables you to run your Spark Jobs on the Databricks 6.4
distribution.
Big Data > Distributions
Databricks 7.3 LTS
This feature enables you to run your Spark Jobs on the Databricks 7.3 LTS
distribution.
Big Data > Distributions
Hortonworks HDP Dynamic Distribution
This feature enables you to run your Spark Jobs on Hortonworks HDP using either
Static or Dynamic distributions.
Big Data > Distributions
Microsoft Azure HDInsight 4.0
This feature enables you to run your Spark Jobs on the Microsoft Azure
HDInsight 4.0 distribution.
Big Data > Universal Distribution (Recommended)
Universal Distribution (Spark 2.4.x)
This feature enables you to run your Spark Jobs on Universal distribution with
Spark 2.4.x.
Big Data > Universal Distribution (Recommended)
Universal Distribution (Spark 3.0.x)
This feature enables you to run your Spark Jobs on Universal distribution with
Spark 3.0.x.
Big Data > Universal Distribution (Recommended)
Universal Distribution (Spark 3.1.x)
This feature enables you to run your Spark Jobs on Universal distribution with
Spark 3.1.x.
Big Data > Universal Distribution (Recommended)
Universal Distribution (Spark 3.2.x)
This feature enables you to run your Spark Jobs on Universal distribution with
Spark 3.2.x.
Big Data > Universal Distribution (Recommended)
Universal Distribution (Spark 3.3.x)
This feature enables you to run your Spark Jobs on Universal distribution with
Spark 3.3.x.
Big Data > Universal Distribution (Recommended)
Universal Distribution (Spark 3.2.x)
This feature enables you to run your Spark Jobs on Universal distribution with
Spark 3.2.x.
Big Data > Universal Distribution (Recommended)
Universal Distribution (Spark 3.3.x)
This feature enables you to run your Spark Jobs on Universal distribution with
Spark 3.3.x.
Talend MDM
addresses the challenges of creating and managing master data for all types of
organizations where data is hosted under various formats in various systems. It
groups all master data of the company in a central hub and has all the core features
a user needs for an MDM application: advanced modeling, model-driven dynamic web
interface, full-text search, event triggering, role-based security, etc.
Did this page help you?
If you find any issues with this page or its content – a typo, a missing step, or a technical error – please let us know!