Model Context Protocol (MCP) finally gives AI models a way to access the business data needed to make them really useful at work. CData MCP Servers have the depth and performance to make sure AI has access to all of the answers.
Try them now for free →A PostgreSQL Interface for Hive Data
Use the Remoting features of the Hive JDBC Driver to create a PostgreSQL entry-point for data access.
There are a vast number of PostgreSQL clients available on the Internet. From standard Drivers to BI and Analytics tools, PostgreSQL is a popular interface for data access. Using our JDBC Drivers, you can now create PostgreSQL entry-points that you can connect to from any standard client.
To access Hive data as a PostgreSQL database, use the CData JDBC Driver for Hive and a JDBC foreign data wrapper (FDW). In this article, we compile the FDW, install it, and query Hive data from PostgreSQL Server.
Connect to Hive Data as a JDBC Data Source
To connect to Hive as a JDBC data source, you will need the following:
- Driver JAR path: The JAR is located in the lib subfolder of the installation directory.
Driver class:
cdata.jdbc.apachehive.ApacheHiveDriver
- JDBC URL:
The URL must start with "jdbc:apachehive:" and can include any of the connection properties in name-value pairs separated with semicolons.
Set the Server, Port, TransportMode, and AuthScheme connection properties to connect to Hive.
Built-in Connection String Designer
For assistance in constructing the JDBC URL, use the connection string designer built into the Hive JDBC Driver. Either double-click the JAR file or execute the jar file from the command-line.
java -jar cdata.jdbc.apachehive.jar
Fill in the connection properties and copy the connection string to the clipboard.
A typical JDBC URL is below:
jdbc:apachehive:Server=127.0.0.1;Port=10000;TransportMode=BINARY;
Build the JDBC Foreign Data Wrapper
The Foreign Data Wrapper can be installed as an extension to PostgreSQL, without recompiling PostgreSQL. The jdbc2_fdw extension is used as an example (downloadable here).
- Add a symlink from the shared object for your version of the JRE to /usr/lib/libjvm.so. For example:
ln -s /usr/lib/jvm/java-6-openjdk/jre/lib/amd64/server/libjvm.so /usr/lib/libjvm.so
- Start the build:
make install USE_PGXS=1
Query Hive Data as a PostgreSQL Database
After you have installed the extension, follow the steps below to start executing queries to Hive data:
- Log into your database.
-
Load the extension for the database:
CREATE EXTENSION jdbc2_fdw;
-
Create a server object for Hive:
CREATE SERVER ApacheHive FOREIGN DATA WRAPPER jdbc2_fdw OPTIONS ( drivername 'cdata.jdbc.apachehive.ApacheHiveDriver', url 'jdbc:apachehive:Server=127.0.0.1;Port=10000;TransportMode=BINARY;', querytimeout '15', jarfile '/home/MyUser/CData/CData\ JDBC\ Driver\ for\ Salesforce MyDriverEdition/lib/cdata.jdbc.apachehive.jar');
-
Create a user mapping for the username and password of a user known to the MySQL daemon.
CREATE USER MAPPING for postgres SERVER ApacheHive OPTIONS ( username 'admin', password 'test');
-
Create a foreign table in your local database:
postgres=# CREATE FOREIGN TABLE customers ( customers_id text, customers_City text, customers_CompanyName numeric) SERVER ApacheHive OPTIONS ( table_name 'customers');
postgres=# SELECT * FROM customers;