HCatalog Connector requirements

Before you can use the HCatalog Connector, both your OpenText™ Analytics Database and Hadoop installations must meet the following requirements.

Before you can use the HCatalog Connector, both your OpenText™ Analytics Database and Hadoop installations must meet the following requirements.

Database requirements

All of the nodes in your cluster must have a Java Virtual Machine (JVM) installed. You must use the same Java version that the Hadoop cluster uses. See Installing Java Runtime on your database cluster.

You must also add certain libraries distributed with Hadoop and Hive to your database installation directory. See Configuring OpenText™ Analytics Database for HCatalog.

Hadoop requirements

Your Hadoop cluster must meet several requirements to operate correctly with the database Connector for HCatalog:

  • It must have Hive, HiveServer2, and HCatalog installed and running. See Apache's HCatalog page for more information.

  • The HiveServer2 server and all of the HDFS nodes that store HCatalog data must be directly accessible from all of the hosts in your database. Verify that any firewall separating the Hadoop cluster and the database cluster will pass HiveServer2, metastore database, and HDFS traffic.

  • The data that you want to query must be in an internal or external Hive table.

  • If a table you want to query uses a non-standard SerDe, you must install the SerDe's classes on your database cluster before you can query the data. See Using nonstandard SerDes.