Apache Hudi (pronounced “hoodie”) is the next generation streaming data lake platform. Apache Hudi brings core warehouse and database functionality directly to a data lake.


This article assumes that you have mastered the basic knowledge and operation of Hudi. For the knowledge about Hudi not mentioned in this article, you can obtain it from its Official Documentation.

By using Kyuubi, we can run SQL queries towards Hudi which is more convenient, easy to understand, and easy to expand than directly using Spark to manipulate Hudi.

Hudi Integration#

To enable the integration of kyuubi spark sql engine and Hudi through Catalog APIs, you need to:


The classpath of kyuubi spark sql engine with Hudi supported consists of

  1. kyuubi-spark-sql-engine-1.7.0-SNAPSHOT_2.12.jar, the engine jar deployed with Kyuubi distributions

  2. a copy of spark distribution

  3. hudi-spark<spark.version>-bundle_<scala.version>-<hudi.version>.jar (example: hudi-spark3.2-bundle_2.12-0.11.1.jar), which can be found in the Maven Central

In order to make the Hudi packages visible for the runtime classpath of engines, we can use one of these methods:

  1. Put the Hudi packages into $SPARK_HOME/jars directly

  2. Set spark.jars=/path/to/hudi-spark-bundle


To activate functionality of Hudi, we can set the following configurations:

Hudi Operations#

Taking Create Table as a example,

CREATE TABLE hudi_cow_nonpcf_tbl (
  uuid INT,
  name STRING,
  price DOUBLE

Taking Query Data as a example,

SELECT * FROM hudi_cow_nonpcf_tbl WHERE id < 20;

Taking Insert Data as a example,

INSERT INTO hudi_cow_nonpcf_tbl SELECT 1, 'a1', 20;

Taking Update Data as a example,

UPDATE hudi_cow_nonpcf_tbl SET name = 'foo', price = price * 2 WHERE id = 1;

Taking Delete Data as a example,

DELETE FROM hudi_cow_nonpcf_tbl WHERE uuid = 1;