Hive JDBC Driver#
Instructions#
Kyuubi is fully compatible with Hive JDBC and ODBC drivers that let you connect to popular Business Intelligence (BI) tools to query, analyze and visualize data though Spark SQL engines.
It’s recommended to use Kyuubi JDBC driver for new applications.
Install Hive JDBC#
For programing, the easiest way to get hive-jdbc
is from the maven central. For example,
The following sections demonstrate how to use Hive JDBC driver 2.3.8 to connect Kyuubi Server, actually, any version less or equals 3.1.x should work fine.
maven
<dependency>
<groupId>org.apache.hive</groupId>
<artifactId>hive-jdbc</artifactId>
<version>2.3.8</version>
</dependency>
sbt
libraryDependencies += "org.apache.hive" % "hive-jdbc" % "2.3.8"
gradle
implementation group: 'org.apache.hive', name: 'hive-jdbc', version: '2.3.8'
For BI tools, please refer to Quick Start to check the guide for the BI tool used. If you find there is no specific document for the BI tool that you are using, don’t worry, the configuration part for all BI tools are basically the same. Also, we will appreciate if you can help us to improve the document.
JDBC URL#
JDBC URLs have the following format:
jdbc:hive2://<host>:<port>/<dbName>;<sessionVars>?<kyuubiConfs>#<[spark|hive]Vars>
JDBC Parameter | Description |
---|---|
host | The cluster node hosting Kyuubi Server. |
port | The port number to which is Kyuubi Server listening. |
dbName | Optional database name to set the current database to run the query against, use default if absent. |
sessionVars | Optional Semicolon(;) separated key=value parameters for the JDBC/ODBC driver. Such as user , password and hive.server2.proxy.user . |
kyuubiConfs | Optional Semicolon(;) separated key=value parameters for Kyuubi server to create the corresponding engine, dismissed if engine exists. |
[spark|hive]Vars | Optional Semicolon(;) separated key=value parameters for Spark/Hive variables used for variable substitution. |
Example#
jdbc:hive2://localhost:10009/default;hive.server2.proxy.user=proxy_user?kyuubi.engine.share.level=CONNECTION;spark.ui.enabled=false#var_x=y