This is an automated email from the ASF dual-hosted git repository. shaofengshi pushed a commit to branch update_flink_connector_doc in repository https://gitbox.apache.org/repos/asf/gravitino.git
commit c593c47c49ced77e18107b41eacdf5e80fa2a831 Author: Shaofeng Shi <[email protected]> AuthorDate: Mon Aug 4 19:21:42 2025 +0800 update flink connector document --- docs/flink-connector/flink-catalog-hive.md | 16 +++++++++++----- docs/flink-connector/flink-catalog-jdbc.md | 6 ++++++ docs/flink-connector/flink-connector.md | 15 ++++++++++----- 3 files changed, 27 insertions(+), 10 deletions(-) diff --git a/docs/flink-connector/flink-catalog-hive.md b/docs/flink-connector/flink-catalog-hive.md index 9fc9349e35..057cd4d0d0 100644 --- a/docs/flink-connector/flink-catalog-hive.md +++ b/docs/flink-connector/flink-catalog-hive.md @@ -39,18 +39,24 @@ USE CATALOG hive_a; CREATE DATABASE IF NOT EXISTS mydatabase; USE mydatabase; +SET 'execution.runtime-mode' = 'batch'; +-- [INFO] Execute statement succeed. + +SET 'sql-client.execution.result-mode' = 'tableau'; +-- [INFO] Execute statement succeed. + // Create table CREATE TABLE IF NOT EXISTS employees ( id INT, name STRING, - date INT + dt INT ) -PARTITIONED BY (date); +PARTITIONED BY (dt); -DESC TABLE EXTENDED employees; +DESC EXTENDED employees; -INSERT INTO TABLE employees VALUES (1, 'John Doe', 20240101), (2, 'Jane Smith', 20240101); -SELECT * FROM employees WHERE date = '20240101'; +INSERT INTO employees VALUES (1, 'John Doe', 20240101), (2, 'Jane Smith', 20240101); +SELECT * FROM employees WHERE dt = 20240101; ``` ## Catalog properties diff --git a/docs/flink-connector/flink-catalog-jdbc.md b/docs/flink-connector/flink-catalog-jdbc.md index 4414500f4e..31c084abda 100644 --- a/docs/flink-connector/flink-catalog-jdbc.md +++ b/docs/flink-connector/flink-catalog-jdbc.md @@ -23,6 +23,12 @@ Place the following JAR files in the lib directory of your Flink installation: - [`gravitino-flink-connector-runtime-1.18_2.12-${gravitino-version}.jar`](https://mvnrepository.com/artifact/org.apache.gravitino/gravitino-flink-connector-runtime-1.18) - JDBC driver +Next, edit the JDBC catalog in Gravitino, add the `flink.bypass.default-database` property, and give it an value of the default database. + +```json +flink.bypass.default-database=db +``` + ### SQL Example ```sql diff --git a/docs/flink-connector/flink-connector.md b/docs/flink-connector/flink-connector.md index 84067ecbeb..da49bee38d 100644 --- a/docs/flink-connector/flink-connector.md +++ b/docs/flink-connector/flink-connector.md @@ -38,28 +38,33 @@ This capability allows users to perform federation queries, accessing data from Set the flink configuration in flink-conf.yaml. ```yaml table.catalog-store.kind: gravitino -table.catalog-store.gravitino.gravitino.metalake: test +table.catalog-store.gravitino.gravitino.metalake: metalake_demo table.catalog-store.gravitino.gravitino.uri: http://localhost:8090 ``` Or you can set the flink configuration in the `TableEnvironment`. ```java final Configuration configuration = new Configuration(); configuration.setString("table.catalog-store.kind", "gravitino"); -configuration.setString("table.catalog-store.gravitino.gravitino.metalake", "test"); +configuration.setString("table.catalog-store.gravitino.gravitino.metalake", "metalake_demo"); configuration.setString("table.catalog-store.gravitino.gravitino.uri", "http://localhost:8090"); EnvironmentSettings.Builder builder = EnvironmentSettings.newInstance().withConfiguration(configuration); TableEnvironment tableEnv = TableEnvironment.create(builder.inBatchMode().build()); ``` -3. Execute the Flink SQL query. +3. Add necessary jar files to Flink's `lib` sub-folder. +To run Flink with Gravitino connector and then access the data sources like Hive, Iceberg and others, you need to put the necessary jar files into Flink's `lib` folder. For example, to access Hive, you need to add at least these jar files: `flink-connector-hive_{version}.jar`, `hive-metastore-{version}.jar`, `hive-exec-{version}-core.jar`, `hive-serde-{version}.jar`, `hive-shims-common-{version}.jar`, `hive-storage-api-{version}.jar`, etc. -Suppose there is only one hive catalog with the name `hive` in the metalake `test`. +4. Execute the Flink SQL query. + +Suppose there is only one hive catalog with the name `catalog_hive` in the metalake `metalake_demo`. ```sql // use hive catalog -USE hive; +USE CATALOG catalog_hive; CREATE DATABASE db; USE db; +SET 'execution.runtime-mode' = 'batch'; +SET 'sql-client.execution.result-mode' = 'tableau'; CREATE TABLE hive_students (id INT, name STRING); INSERT INTO hive_students VALUES (1, 'Alice'), (2, 'Bob'); SELECT * FROM hive_students;
