Hive JDBC Driver#

Instructions#

Kyuubi is fully compatible with Hive JDBC and ODBC drivers that let you connect to popular Business Intelligence (BI) tools to query, analyze and visualize data though Spark SQL engines.

It’s recommended to use Kyuubi JDBC driver for new applications.

Install Hive JDBC#

For programing, the easiest way to get hive-jdbc is from the maven central. For example,

The following sections demonstrate how to use Hive JDBC driver 2.3.8 to connect Kyuubi Server, actually, any version less or equals 3.1.x should work fine.

  • maven

<dependency>
    <groupId>org.apache.hive</groupId>
    <artifactId>hive-jdbc</artifactId>
    <version>2.3.8</version>
</dependency>
  • sbt

libraryDependencies += "org.apache.hive" % "hive-jdbc" % "2.3.8"
  • gradle

implementation group: 'org.apache.hive', name: 'hive-jdbc', version: '2.3.8'

For BI tools, please refer to Quick Start to check the guide for the BI tool used. If you find there is no specific document for the BI tool that you are using, don’t worry, the configuration part for all BI tools are basically the same. Also, we will appreciate if you can help us to improve the document.

JDBC URL#

JDBC URLs have the following format:

jdbc:hive2://<host>:<port>/<dbName>;<sessionVars>?<kyuubiConfs>#<[spark|hive]Vars>
JDBC Parameter Description
host The cluster node hosting Kyuubi Server.
port The port number to which is Kyuubi Server listening.
dbName Optional database name to set the current database to run the query against, use default if absent.
sessionVars Optional Semicolon(;) separated key=value parameters for the JDBC/ODBC driver. Such as user, password and hive.server2.proxy.user.
kyuubiConfs Optional Semicolon(;) separated key=value parameters for Kyuubi server to create the corresponding engine, dismissed if engine exists.
[spark|hive]Vars Optional Semicolon(;) separated key=value parameters for Spark/Hive variables used for variable substitution.

Example#

jdbc:hive2://localhost:10009/default;hive.server2.proxy.user=proxy_user?kyuubi.engine.share.level=CONNECTION;spark.ui.enabled=false#var_x=y