Configuring a HDFS connection to run on Spark
Using the tHDFSConfiguration component, you can connect your HDFS filesystem to Spark.
Before you begin
- This tutorial makes use of a Hadoop cluster. You must have a Hadoop cluster available to you.
- You must also have HDFS metadata configured (see Creating a Hadoop cluster metadata definition and Importing a Hadoop cluster metadata definition).
Procedure
-
In the Repository, expand
, then expand the Hadoop cluster
metadata of your choice.
- Click OK.