/
Talend - Download a file OpenData and write file with HDFS
Talend - Download a file OpenData and write file with HDFS
Github Project : example-talend-download-file-opendata-and-write-file-with-HDFS
Preamble
Configuration: Context
To create the different jobs displayed in this article, you have to create a repository: With VALUES
Download a file Open Data and write file with HDFS (with all versions of Data Fabric)
- Create a new job
- Add the component "HDFSConnection" : Allows the creation of a HDFS connection.
- Add the component "tREST" : Send a message for RESTful webservice and retrieve the response.
- Add the component "tHDFSOutput" : Writes data to HDFS.
- Create links:
- "HDFSConnection" is connected with "tREST" (through "OnSubjobOk")
- "tREST" is connected with "tHDFSOutput" (through "Main")
- Double click on "tHDFSConnection" and set its properties:
- Add a "Cloudera" distribution and select the latest version of Cloudera
- Enter the Name Node URL.
The URL has to respect this format : "hdfs://ip_hdfs:port_hdfs/"
Use context variables if possible : "hdfs://"+context.IP_HDFS+":"+context.Port_HDFS+"/" - Add the user
- Uncheck "Use Datanode Hostname"
- Double click on "tREST" and enter a URL
- Double click on "tHDFSOutput" :
- Check "Use an existing connection"
- Enter a name file
- Run a job
, multiple selections available,
Related content
Talend - Write files with HDFS
Talend - Write files with HDFS
More like this
Talend - Read files with HDFS
Talend - Read files with HDFS
More like this
Talend - List file in HDFS
Talend - List file in HDFS
More like this
Talend - HDFS with high availability
Talend - HDFS with high availability
More like this
Talend
Talend
More like this
R - Read & Write files from HDFS
R - Read & Write files from HDFS
More like this