Manage your HDFS data in SQL Server with Talend

Apache HDFS (Hadoop Distributed File Systems) provides a software framework for distributed storage and processing of big data. In combination with tools such as MapReduce, Yarn, and other core modules, HDFS lets organizations build Apache Hadoop clusters of hundreds or thousands of nodes that can handle datasets of terabyte size. A robust ecosystem of other tools can take advantage of data stored in HDFS. Manage HDFS data in SQL Server with Talend's suite of data integration tools.

Connecting to HDFS

To connect to HDFS, use the Component tab of the tHDFSExist component. Enter the Hadoop distribution and version, the HDFS directory, and name of the file you want to use.

Learn more about connecting to HDFS

More about integrating HDFS data

Talend has detailed documentation on how to ETL your HDFS data for a better view of the business.

Connecting to SQL Server

SQL Server is a popular relational database management system created by Microsoft.

Connect to your Microsoft SQL Server database with our native connector and easily carry out the following actions on your data:

  • Write, update, make changes or suppress entries defined on a table and/or on the data contained in a table
  • Prepare a file to be used as parameter in an INSERT query to feed a MS SQL database.
  • Execute a SQL query stated onto a specified Microsoft SQL Server or Azure SQL database.
  • Offer a convenient way to centralize multiple or complex queries in a database and calls them easily.

Work with your HDFS data

Computing data with Hadoop distributed file system Create a file in a defined directory, get it into and out of HDFS, store it to another local directory, and read it at the end of the Job. See how >

Get more from your HDFS data

Deliver data your organization can trust... Get started today.

Explore Talend's full suite of apps