Manage your Hive data in Databricks with Talend

Apache Hive is a data warehouse built on top of Apache Hadoop and used for querying, summarizing and analyzing large datasets. Businesses can run Hive on hardware in their own data centers or on cloud platforms. Its HiveQL query language uses syntax similar to that of SQL. Manage Hive data in Databricks with Talend's suite of data integration tools.

Connecting to Hive

To connect to Hive, use the tHiveConnection component. Connection parameters differ depending on where your Hive data is hosted.

Find out more about connecting to Apache Hive

More about integrating Hive data

Talend has detailed documentation on how to ETL your Hive data for a better view of the business.

Connecting to Databricks

Databricks is the leader in unified analytics and founded by the original creators of Apache Spark™.

The Talend tDBFSConnection module connects to DBFS (the Databricks Filesystem) system. DBFS components are designed for quick and straightforward data transferring with Databricks. For more sophisticated scenarios you can also use Spark Jobs with Databricks.

Get more from your Hive data

Deliver data your organization can trust... Get started today.

Explore Talend's full suite of apps