This is an automated email from the ASF dual-hosted git repository.

chengpan pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/kyuubi.git


The following commit(s) were added to refs/heads/master by this push:
     new 96498844d [KYUUBI #6239] Rename `beeline` to `kyuubi-beeline`
96498844d is described below

commit 96498844d367e6c664c9cf864e66ec7c8136b1aa
Author: Cheng Pan <[email protected]>
AuthorDate: Wed Apr 3 18:35:38 2024 +0800

    [KYUUBI #6239] Rename `beeline` to `kyuubi-beeline`
    
    # :mag: Description
    
    Discussion in mailing thread: 
https://lists.apache.org/thread/tnmz71o3rypy7qvs3899p3jkkq4xqb4r
    
    I propose to rename the `bin/beeline` to `bin/kyuubi-beeline`, while for 
compatibility, we may still want to keep the alias `bin/beeline` for a while.
    
    In a real Hadoop cluster, it’s likely to add `$HIVE_HOME/bin`, 
`$SPARK_HOME/bin`, `$KYUUBI_HOME/bin` to the `$PATH`, at the current state, 
when performing `beeline`, which one is called depends on the declaration order.
    
    It does not matter for Spark’s `bin/beeline` because it’s a vanilla Hive 
BeeLine, but in Kyuubi, we have made some improvements based on vanilla Hive 
BeeLine, so the behavior is not exactly same as Hive’s BeeLine.
    
    An identical name would solve this problem. And I saw some vendors[1] who 
shippes Kyuubi already have done the same thing.
    
    [1] https://help.aliyun.com/zh/emr/emr-on-ecs/user-guide/connect-to-kyuubi
    
    ## Types of changes :bookmark:
    
    - [ ] Bugfix (non-breaking change which fixes an issue)
    - [ ] New feature (non-breaking change which adds functionality)
    - [ ] Breaking change (fix or feature that would cause existing 
functionality to change)
    
    ## Test Plan 🧪
    
    Manual test.
    
    ```
    $ bin/beeline -u 'jdbc:kyuubi://0.0.0.0:10009/'
    Warning: beeline is deprecated and will be removed in the future, please 
use kyuubi-beeline instead.
    Connecting to jdbc:kyuubi://0.0.0.0:10009/
    Connected to: Spark SQL (version 3.4.1)
    Driver: Kyuubi Project Hive JDBC Client (version 1.10.0-SNAPSHOT)
    Beeline version 1.10.0-SNAPSHOT by Apache Kyuubi
    0: jdbc:kyuubi://0.0.0.0:10009/>
    ```
    
    ```
    $ bin/kyuubi-beeline -u 'jdbc:kyuubi://0.0.0.0:10009/'
    Connecting to jdbc:kyuubi://0.0.0.0:10009/
    Connected to: Spark SQL (version 3.4.1)
    Driver: Kyuubi Project Hive JDBC Client (version 1.10.0-SNAPSHOT)
    Beeline version 1.10.0-SNAPSHOT by Apache Kyuubi
    0: jdbc:kyuubi://0.0.0.0:10009/>
    ```
    
    ---
    
    # Checklist 📝
    
    - [x] This patch was not authored or co-authored using [Generative 
Tooling](https://www.apache.org/legal/generative-tooling.html)
    
    **Be nice. Be informative.**
    
    Closes #6239 from pan3793/kyuubi-beeline.
    
    Closes #6239
    
    cec8f56e2 [Cheng Pan] docs
    b3446baf1 [Cheng Pan] docs
    46a115077 [Cheng Pan] Remove `bin/beeline` to `bin/kyuubi-beeline`
    
    Authored-by: Cheng Pan <[email protected]>
    Signed-off-by: Cheng Pan <[email protected]>
---
 bin/beeline                                        | 38 ++--------------------
 bin/{beeline => kyuubi-beeline}                    |  0
 docs/client/advanced/features/plan_only.md         | 18 +++++-----
 docs/client/advanced/kerberos.md                   |  4 +--
 .../connector/spark/delta_lake_with_azure_blob.rst |  6 ++--
 docs/connector/spark/kudu.md                       | 12 +++----
 docs/deployment/engine_on_yarn.md                  |  2 +-
 docs/deployment/high_availability_guide.md         |  2 +-
 docs/deployment/hive_metastore.md                  | 21 ++++++------
 docs/deployment/kyuubi_on_kubernetes.md            |  6 ++--
 docs/deployment/migration-guide.md                 |  4 +++
 docs/deployment/spark/incremental_collection.md    | 16 ++++-----
 docs/extensions/server/configuration.rst           |  2 +-
 docs/monitor/logging.md                            |  6 ++--
 docs/overview/architecture.md                      |  2 +-
 docs/quick_start/quick_start.rst                   | 24 +++++++-------
 16 files changed, 66 insertions(+), 97 deletions(-)

diff --git a/bin/beeline b/bin/beeline
index ada7fecfa..9fa281009 100755
--- a/bin/beeline
+++ b/bin/beeline
@@ -16,44 +16,10 @@
 # limitations under the License.
 #
 
-## Kyuubi BeeLine Entrance
-CLASS="org.apache.hive.beeline.KyuubiBeeLine"
+echo "Warning: beeline is deprecated and will be removed in the future, please 
use kyuubi-beeline instead."
 
 if [ -z "${KYUUBI_HOME}" ]; then
   KYUUBI_HOME="$(cd "`dirname "$0"`"/..; pwd)"
 fi
 
-. "${KYUUBI_HOME}/bin/load-kyuubi-env.sh" -s
-
-if [[ -z ${JAVA_HOME} ]]; then
-  echo "Error: JAVA_HOME IS NOT SET! CANNOT PROCEED."
-  exit 1
-fi
-
-RUNNER="${JAVA_HOME}/bin/java"
-
-# Append jline option to enable the Beeline process to run in background.
-if [[ ( ! $(ps -o stat= -p $$) =~ "+" ) && ! ( -p /dev/stdin ) ]]; then
-  export KYUUBI_BEELINE_OPTS="$KYUUBI_BEELINE_OPTS 
-Djline.terminal=jline.UnsupportedTerminal"
-fi
-
-## Find the Kyuubi beeline Jar
-if [[ -z "$KYUUBI_BEELINE_JARS" ]]; then
-  KYUUBI_BEELINE_JARS="$KYUUBI_HOME/beeline-jars"
-  if [[ ! -d ${KYUUBI_BEELINE_JARS} ]]; then
-    echo -e "\nCandidate Kyuubi beeline jars $KYUUBI_BEELINE_JARS doesn't 
exist, using $KYUUBI_HOME/kyuubi-hive-beeline/target/"
-    KYUUBI_BEELINE_JARS="$KYUUBI_HOME/kyuubi-hive-beeline/target"
-  fi
-fi
-
-if [[ -z ${YARN_CONF_DIR} ]]; then
-  KYUUBI_BEELINE_CLASSPATH="${KYUUBI_BEELINE_JARS}/*:${HADOOP_CONF_DIR}"
-else
-  
KYUUBI_BEELINE_CLASSPATH="${KYUUBI_BEELINE_JARS}/*:${HADOOP_CONF_DIR}:${YARN_CONF_DIR}"
-fi
-
-if [[ -f ${KYUUBI_CONF_DIR}/log4j2-repl.xml ]]; then
-  KYUUBI_CTL_JAVA_OPTS="${KYUUBI_CTL_JAVA_OPTS} 
-Dlog4j2.configurationFile=log4j2-repl.xml"
-fi
-
-exec ${RUNNER} ${KYUUBI_BEELINE_OPTS} -cp ${KYUUBI_BEELINE_CLASSPATH} $CLASS 
"$@"
+exec "${KYUUBI_HOME}/bin/kyuubi-beeline" "$@"
diff --git a/bin/beeline b/bin/kyuubi-beeline
similarity index 100%
copy from bin/beeline
copy to bin/kyuubi-beeline
diff --git a/docs/client/advanced/features/plan_only.md 
b/docs/client/advanced/features/plan_only.md
index bcedb2f02..4d1d1737a 100644
--- a/docs/client/advanced/features/plan_only.md
+++ b/docs/client/advanced/features/plan_only.md
@@ -27,17 +27,17 @@ You can add parameters to the URL when establishing a JDBC 
connection, the param
 JDBC URLs have the following format:
 
 ```shell
-jdbc:hive2://<host>:<port>/<dbName>;<sessionVars>?kyuubi.operation.plan.only.mode=parse/analyze/optimize/optimize_with_stats/physical/execution/none;<kyuubiConfs>#<[spark|hive]Vars>
+jdbc:kyuubi://<host>:<port>/<dbName>;<sessionVars>?kyuubi.operation.plan.only.mode=parse/analyze/optimize/optimize_with_stats/physical/execution/none;<kyuubiConfs>#<[spark|hive]Vars>
 ```
 
 Refer to [hive_jdbc doc](../../jdbc/hive_jdbc.md) for details of others 
parameters
 
 ### Example:
 
-Using beeline tool to connect to the local service, the Shell command is:
+Using `kyuubi-beeline` to connect to the local service, the Shell command is:
 
 ```shell
-beeline -u 
'jdbc:hive2://0.0.0.0:10009/default?kyuubi.operation.plan.only.mode=parse' -n 
{user_name}
+kyuubi-beeline -u 
'jdbc:kyuubi://0.0.0.0:10009/default?kyuubi.operation.plan.only.mode=parse' -n 
{user_name}
 ```
 
 Running the following SQL:
@@ -50,7 +50,7 @@ The results are as follows:
 
 ```shell
 # SQL:
-0: jdbc:hive2://0.0.0.0:10009/default> SELECT * FROM t1 LEFT JOIN t2 ON t1.id 
= t2.id;
+0: jdbc:kyuubi://0.0.0.0:10009/default> SELECT * FROM t1 LEFT JOIN t2 ON t1.id 
= t2.id;
 
 #Result:
 +----------------------------------------------------+
@@ -63,7 +63,7 @@ The results are as follows:
  |
 +----------------------------------------------------+
 1 row selected (3.008 seconds)
-0: jdbc:hive2://0.0.0.0:10009/default>
+0: jdbc:kyuubi://0.0.0.0:10009/default>
 ```
 
 ## Session-level parameter setting
@@ -71,7 +71,7 @@ The results are as follows:
 You can also set the kyuubi.operation.plan.only.mode parameter by executing 
the set command after the connection has been established
 
 ```shell
-beeline -u 'jdbc:hive2://0.0.0.0:10009/default' -n {user_name}
+kyuubi-beeline -u 'jdbc:kyuubi://0.0.0.0:10009/default' -n {user_name}
 ```
 
 Running the following SQL:
@@ -85,7 +85,7 @@ The results are as follows:
 
 ```shell
 #set command:
-0: jdbc:hive2://0.0.0.0:10009/default> set 
kyuubi.operation.plan.only.mode=parse;
+0: jdbc:kyuubi://0.0.0.0:10009/default> set 
kyuubi.operation.plan.only.mode=parse;
 
 #set command result:
 +----------------------------------+--------+
@@ -96,7 +96,7 @@ The results are as follows:
 1 row selected (0.568 seconds)
 
 #execute SQL:
-0: jdbc:hive2://0.0.0.0:10009/default> SELECT * FROM t1 LEFT JOIN t2 ON t1.id 
= t2.id;
+0: jdbc:kyuubi://0.0.0.0:10009/default> SELECT * FROM t1 LEFT JOIN t2 ON t1.id 
= t2.id;
 
 # SQL result:
 +----------------------------------------------------+
@@ -109,6 +109,6 @@ The results are as follows:
  |
 +----------------------------------------------------+
 1 row selected (0.404 seconds)
-0: jdbc:hive2://0.0.0.0:10009/default>
+0: jdbc:kyuubi://0.0.0.0:10009/default>
 ```
 
diff --git a/docs/client/advanced/kerberos.md b/docs/client/advanced/kerberos.md
index a9cb55812..c7bd643ad 100644
--- a/docs/client/advanced/kerberos.md
+++ b/docs/client/advanced/kerberos.md
@@ -229,13 +229,13 @@ The last step is to connect to Kyuubi with the right JDBC 
URL.
 The JDBC URL should be in format:
 
 ```
-jdbc:hive2://<kyuubi_server_address>:<kyuubi_server_port>/<db>;principal=<kyuubi_server_principal>
+jdbc:kyuubi://<kyuubi_server_address>:<kyuubi_server_port>/<db>;principal=<kyuubi_server_principal>
 ```
 
 or
 
 ```
-jdbc:hive2://<kyuubi_server_address>:<kyuubi_server_port>/<db>;kyuubiServerPrincipal=<kyuubi_server_principal>
+jdbc:kyuubi://<kyuubi_server_address>:<kyuubi_server_port>/<db>;kyuubiServerPrincipal=<kyuubi_server_principal>
 ```
 
 **Note**:  
diff --git a/docs/connector/spark/delta_lake_with_azure_blob.rst 
b/docs/connector/spark/delta_lake_with_azure_blob.rst
index 67065641d..f9e07afa7 100644
--- a/docs/connector/spark/delta_lake_with_azure_blob.rst
+++ b/docs/connector/spark/delta_lake_with_azure_blob.rst
@@ -190,7 +190,7 @@ Check kyuubi log, in order to check kyuubi start status and 
find the jdbc connec
 
 .. code-block:: log
 
-   2021-11-26 17:49:50.235 INFO service.ThriftFrontendService: Starting and 
exposing JDBC connection at: jdbc:hive2://HOST:10009/
+   2021-11-26 17:49:50.235 INFO service.ThriftFrontendService: Starting and 
exposing JDBC connection at: jdbc:kyuubi://HOST:10009/
    2021-11-26 17:49:50.265 INFO client.ServiceDiscovery: Created a 
/kyuubi/serviceUri=host:10009;version=1.3.1-incubating;sequence=0000000037 on 
ZooKeeper for KyuubiServer uri: host:10009
    2021-11-26 17:49:50.267 INFO server.KyuubiServer: Service[KyuubiServer] is 
started.
 
@@ -199,11 +199,11 @@ You can get the jdbc connection url by the log above.
 Test The Connectivity Of Kyuubi And Delta Lake
 **********************************************
 
-Use ``$KYUUBI_HOME/bin/beeline`` tool,
+Use ``$KYUUBI_HOME/bin/kyuubi-beeline`` tool,
 
 .. code-block:: shell
 
-   ./bin//beeline -u 'jdbc:hive2://<YOUR_HOST>:10009/'
+   ./bin/kyuubi-beeline -u 'jdbc:kyuubi://<YOUR_HOST>:10009/'
 
 At the same time, you can also check whether the engine is running on the 
spark UI:
 
diff --git a/docs/connector/spark/kudu.md b/docs/connector/spark/kudu.md
index ce2d1e88c..c38e7a8d3 100644
--- a/docs/connector/spark/kudu.md
+++ b/docs/connector/spark/kudu.md
@@ -50,7 +50,7 @@ Now, you can start Kyuubi server with this kudu embedded 
Spark distribution.
 #### Start Beeline Or Other Client You Prefer
 
 ```shell
-bin/beeline -u 'jdbc:hive2://<host>:<port>/;principal=<if 
kerberized>;#spark.yarn.queue=kyuubi_test'
+bin/kyuubi-beeline -u 'jdbc:kyuubi://<host>:<port>/;principal=<if 
kerberized>;#spark.yarn.queue=kyuubi_test'
 ```
 
 #### Register Kudu table as Spark Temporary view
@@ -64,7 +64,7 @@ options (
 ```
 
 ```sql
-0: jdbc:hive2://spark5.jd.163.org:10009/> show tables;
+0: jdbc:kyuubi://spark5.jd.163.org:10009/> show tables;
 19/07/09 15:28:03 INFO ExecuteStatementInClientMode: Running query 'show 
tables' with 1104328b-515c-4f8b-8a68-1c0b202bc9ed
 19/07/09 15:28:03 INFO KyuubiSparkUtil$: Application 
application_1560304876299_3805060 has been activated
 19/07/09 15:28:03 INFO ExecuteStatementInClientMode: Executing query in 
incremental mode, running 1 jobs before optimization
@@ -82,7 +82,7 @@ options (
 #### Query Kudu Table
 
 ```sql
-0: jdbc:hive2://spark5.jd.163.org:10009/> select * from kudutest;
+0: jdbc:kyuubi://spark5.jd.163.org:10009/> select * from kudutest;
 19/07/09 15:25:17 INFO ExecuteStatementInClientMode: Running query 'select * 
from kudutest' with ac3e8553-0d79-4c57-add1-7d3ffe34ba16
 19/07/09 15:25:17 INFO KyuubiSparkUtil$: Application 
application_1560304876299_3805060 has been activated
 19/07/09 15:25:17 INFO ExecuteStatementInClientMode: Executing query in 
incremental mode, running 3 jobs before optimization
@@ -103,7 +103,7 @@ options (
 #### Join Kudu table with Hive table
 
 ```sql
-0: jdbc:hive2://spark5.jd.163.org:10009/> select t1.*, t2.* from hive_tbl t1 
join kudutest t2 on t1.userid=t2.userid+1;
+0: jdbc:kyuubi://spark5.jd.163.org:10009/> select t1.*, t2.* from hive_tbl t1 
join kudutest t2 on t1.userid=t2.userid+1;
 19/07/09 15:31:01 INFO ExecuteStatementInClientMode: Running query 'select 
t1.*, t2.* from hive_tbl t1 join kudutest t2 on t1.userid=t2.userid+1' with 
6982fa5c-29fa-49be-a5bf-54c935bbad18
 19/07/09 15:31:01 INFO KyuubiSparkUtil$: Application 
application_1560304876299_3805060 has been activated
 <omitted lines.... >
@@ -123,7 +123,7 @@ options (
 You should notice that only `INSERT INTO` is supported by Kudu, `OVERWRITE` 
data is not supported
 
 ```sql
-0: jdbc:hive2://spark5.jd.163.org:10009/> insert overwrite table kudutest 
select *  from hive_tbl;
+0: jdbc:kyuubi://spark5.jd.163.org:10009/> insert overwrite table kudutest 
select *  from hive_tbl;
 19/07/09 15:35:29 INFO ExecuteStatementInClientMode: Running query 'insert 
overwrite table kudutest select *  from hive_tbl' with 
1afdb791-1aa7-4ceb-8ba8-ff53c17615d1
 19/07/09 15:35:29 INFO KyuubiSparkUtil$: Application 
application_1560304876299_3805060 has been activated
 19/07/09 15:35:30 ERROR ExecuteStatementInClientMode:
@@ -163,7 +163,7 @@ java.lang.UnsupportedOperationException: overwrite is not 
yet supported
 ```
 
 ```sql
-0: jdbc:hive2://spark5.jd.163.org:10009/> insert into table kudutest select * 
from hive_tbl;
+0: jdbc:kyuubi://spark5.jd.163.org:10009/> insert into table kudutest select * 
from hive_tbl;
 19/07/09 15:36:26 INFO ExecuteStatementInClientMode: Running query 'insert 
into table kudutest select *  from hive_tbl' with 
f7460400-0564-4f98-93b6-ad76e579e7af
 19/07/09 15:36:26 INFO KyuubiSparkUtil$: Application 
application_1560304876299_3805060 has been activated
 <omitted lines ...>
diff --git a/docs/deployment/engine_on_yarn.md 
b/docs/deployment/engine_on_yarn.md
index 1025418d9..523b3da03 100644
--- a/docs/deployment/engine_on_yarn.md
+++ b/docs/deployment/engine_on_yarn.md
@@ -61,7 +61,7 @@ These properties are defined by Spark and Kyuubi will pass 
them to `spark-submit
 
 **Note:** None of these would take effect if the application for a particular 
user already exists.
 
-- Specify it in the JDBC connection URL, e.g. 
`jdbc:hive2://localhost:10009/;#spark.master=yarn;spark.yarn.queue=thequeue`
+- Specify it in the JDBC connection URL, e.g. 
`jdbc:kyuubi://localhost:10009/;#spark.master=yarn;spark.yarn.queue=thequeue`
 - Specify it in `$KYUUBI_HOME/conf/kyuubi-defaults.conf`
 - Specify it in `$SPARK_HOME/conf/spark-defaults.conf`
 
diff --git a/docs/deployment/high_availability_guide.md 
b/docs/deployment/high_availability_guide.md
index ca2ffa9fb..8cf729eeb 100644
--- a/docs/deployment/high_availability_guide.md
+++ b/docs/deployment/high_availability_guide.md
@@ -71,7 +71,7 @@ With [Kyuubi Hive JDBC 
Driver](https://mvnrepository.com/artifact/org.apache.kyu
 For example,
 
 ```shell
-bin/beeline -u 
'jdbc:hive2://10.242.189.214:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=kyuubi'
 -n kentyao
+bin/kyuubi-beeline -u 
'jdbc:kyuubi://10.242.189.214:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=kyuubi'
 -n kentyao
 ```
 
 ## How to Hot Upgrade Kyuubi Server
diff --git a/docs/deployment/hive_metastore.md 
b/docs/deployment/hive_metastore.md
index f60465a1a..fd16023a6 100644
--- a/docs/deployment/hive_metastore.md
+++ b/docs/deployment/hive_metastore.md
@@ -41,20 +41,19 @@ By default, Kyuubi launches Spark SQL engines pointing to a 
dummy embedded [Apac
 and this metadata can only be seen by one user at a time, e.g.
 
 ```shell script
-bin/beeline -u 'jdbc:hive2://localhost:10009/' -n kentyao
-Connecting to jdbc:hive2://localhost:10009/
-Connected to: Spark SQL (version 1.0.0-SNAPSHOT)
-Driver: Hive JDBC (version 2.3.7)
-Transaction isolation: TRANSACTION_REPEATABLE_READ
-Beeline version 2.3.7 by Apache Hive
-0: jdbc:hive2://localhost:10009/> show databases;
+bin/kyuubi-beeline -u 'jdbc:kyuubi://localhost:10009/' -n kentyao
+Connecting to jdbc:kyuubi://localhost:10009/
+Connected to: Spark SQL (version 3.4.2)
+Driver: Kyuubi Project Hive JDBC Client (version 1.9.0)
+Beeline version 1.9.0 by Apache Kyuubi
+0: jdbc:kyuubi://localhost:10009/> show databases;
 2020-11-16 23:50:50.388 INFO operation.ExecuteStatement:
            Spark application name: 
kyuubi_kentyao_spark_2020-11-16T15:50:08.968Z
                  application ID:  local-1605541809797
                  application web UI: http://192.168.1.14:60165
                  master: local[*]
                  deploy mode: client
-                 version: 3.0.1
+                 version: 3.4.2
            Start time: 2020-11-16T15:50:09.123Z
            User: kentyao
 2020-11-16 23:50:50.404 INFO metastore.HiveMetaStore: 2: get_databases: *
@@ -66,14 +65,14 @@ Beeline version 2.3.7 by Apache Hive
 | default    |
 +------------+
 1 row selected (0.122 seconds)
-0: jdbc:hive2://localhost:10009/> show tables;
+0: jdbc:kyuubi://localhost:10009/> show tables;
 2020-11-16 23:50:52.957 INFO operation.ExecuteStatement:
            Spark application name: 
kyuubi_kentyao_spark_2020-11-16T15:50:08.968Z
                  application ID:  local-1605541809797
                  application web UI: http://192.168.1.14:60165
                  master: local[*]
                  deploy mode: client
-                 version: 3.0.1
+                 version: 3.4.2
            Start time: 2020-11-16T15:50:09.123Z
            User: kentyao
 2020-11-16 23:50:52.968 INFO metastore.HiveMetaStore: 2: get_database: default
@@ -139,7 +138,7 @@ This version of configuration has lower priority than those 
in `$KYUUBI_HOME/con
 We can pass _**Hive primitives**_ or **_Spark derivatives_** directly in the 
JDBC connection URL, e.g.
 
 ```
-jdbc:hive2://localhost:10009/;#hive.metastore.uris=thrift://localhost:9083
+jdbc:kyuubi://localhost:10009/;#hive.metastore.uris=thrift://localhost:9083
 ```
 
 This will override the defaults in `$SPARK_HOME/conf/hive-site.xml` and 
`$KYUUBI_HOME/conf/kyuubi-defaults.conf` for each _**user account**_.
diff --git a/docs/deployment/kyuubi_on_kubernetes.md 
b/docs/deployment/kyuubi_on_kubernetes.md
index f016a8bb0..8351055f1 100644
--- a/docs/deployment/kyuubi_on_kubernetes.md
+++ b/docs/deployment/kyuubi_on_kubernetes.md
@@ -100,13 +100,13 @@ You should connect like:
 
 ```shell
 kubectl exec -it kyuubi-example -- /bin/bash
-${KYUUBI_HOME}/bin/beeline -u 'jdbc:kyuubi://localhost:10009'
+${KYUUBI_HOME}/bin/kyuubi-beeline -u 'jdbc:kyuubi://localhost:10009'
 ```
 
-Or you can submit tasks directly through local beeline:
+Or you can submit tasks directly through kyuubi-beeline:
 
 ```shell
-${KYUUBI_HOME}/bin/beeline -u 'jdbc:kyuubi://${hostname}:${port}'
+${KYUUBI_HOME}/bin/kyuubi-beeline -u 'jdbc:kyuubi://${hostname}:${port}'
 ```
 
 As using service nodePort, port means nodePort and hostname means any hostname 
of kubernetes node.
diff --git a/docs/deployment/migration-guide.md 
b/docs/deployment/migration-guide.md
index 48ded45ad..a1586d445 100644
--- a/docs/deployment/migration-guide.md
+++ b/docs/deployment/migration-guide.md
@@ -17,6 +17,10 @@
 
 # Kyuubi Migration Guide
 
+## Upgrading from Kyuubi 1.9 to 1.10
+
+* Since Kyuubi 1.10, `beeline` is deprecated and will be removed in the 
future, please use `kyuubi-beeline` instead.
+
 ## Upgrading from Kyuubi 1.8 to 1.9
 
 * Since Kyuubi 1.9.0, `kyuubi.session.conf.advisor` can be set as a sequence, 
Kyuubi supported chaining SessionConfAdvisors.
diff --git a/docs/deployment/spark/incremental_collection.md 
b/docs/deployment/spark/incremental_collection.md
index 28fd4aa78..1210e331a 100644
--- a/docs/deployment/spark/incremental_collection.md
+++ b/docs/deployment/spark/incremental_collection.md
@@ -47,12 +47,12 @@ As above explains, the incremental collection mode is not 
suitable for common qu
 collection mode for specific queries by using
 
 ```
-beeline -u 
'jdbc:hive2://kyuubi:10009/?spark.driver.maxResultSize=8g;spark.driver.memory=12g#kyuubi.engine.share.level=CONNECTION;kyuubi.operation.incremental.collect=true'
 \
+kyuubi-beeline -u 
'jdbc:kyuubi://kyuubi:10009/?spark.driver.maxResultSize=8g;spark.driver.memory=12g#kyuubi.engine.share.level=CONNECTION;kyuubi.operation.incremental.collect=true'
 \
     --incremental=true \
     -f big_result_query.sql
 ```
 
-`--incremental=true` is required for beeline client, otherwise, the entire 
result sets is fetched and buffered before
+`--incremental=true` is required for kyuubi-beeline client, otherwise, the 
entire result sets is fetched and buffered before
 being displayed, which may cause client side OOM.
 
 ## Change incremental collection mode in session
@@ -60,10 +60,10 @@ being displayed, which may cause client side OOM.
 The configuration `kyuubi.operation.incremental.collect` can also be changed 
using `SET` in session.
 
 ```
-~ beeline -u 'jdbc:hive2://localhost:10009'
-Connected to: Apache Kyuubi (Incubating) (version 1.5.0-SNAPSHOT)
+~ kyuubi-beeline -u 'jdbc:kyuubi://localhost:10009'
+Connected to: Apache Kyuubi (version 1.9.0)
 
-0: jdbc:hive2://localhost:10009/> set 
kyuubi.operation.incremental.collect=true;
+0: jdbc:kyuubi://localhost:10009/> set 
kyuubi.operation.incremental.collect=true;
 +---------------------------------------+--------+
 |                  key                  | value  |
 +---------------------------------------+--------+
@@ -71,7 +71,7 @@ Connected to: Apache Kyuubi (Incubating) (version 
1.5.0-SNAPSHOT)
 +---------------------------------------+--------+
 1 row selected (0.039 seconds)
 
-0: jdbc:hive2://localhost:10009/> select /*+ REPARTITION(5) */ * from range(1, 
10);
+0: jdbc:kyuubi://localhost:10009/> select /*+ REPARTITION(5) */ * from 
range(1, 10);
 +-----+
 | id  |
 +-----+
@@ -88,7 +88,7 @@ Connected to: Apache Kyuubi (Incubating) (version 
1.5.0-SNAPSHOT)
 +-----+
 10 rows selected (1.929 seconds)
 
-0: jdbc:hive2://localhost:10009/> set 
kyuubi.operation.incremental.collect=false;
+0: jdbc:kyuubi://localhost:10009/> set 
kyuubi.operation.incremental.collect=false;
 +---------------------------------------+--------+
 |                  key                  | value  |
 +---------------------------------------+--------+
@@ -96,7 +96,7 @@ Connected to: Apache Kyuubi (Incubating) (version 
1.5.0-SNAPSHOT)
 +---------------------------------------+--------+
 1 row selected (0.027 seconds)
 
-0: jdbc:hive2://localhost:10009/> select /*+ REPARTITION(5) */ * from range(1, 
10);
+0: jdbc:kyuubi://localhost:10009/> select /*+ REPARTITION(5) */ * from 
range(1, 10);
 +-----+
 | id  |
 +-----+
diff --git a/docs/extensions/server/configuration.rst 
b/docs/extensions/server/configuration.rst
index 97f503426..b23cae6ae 100644
--- a/docs/extensions/server/configuration.rst
+++ b/docs/extensions/server/configuration.rst
@@ -68,6 +68,6 @@ If a user `uly` creates a connection with:
 
 .. code-block:: java
 
-   
jdbc:hive2://localhost:10009/;hive.server2.proxy.user=uly;#spark.driver.memory=2G
+   
jdbc:kyuubi://localhost:10009/;hive.server2.proxy.user=uly;#spark.driver.memory=2G
 
 The final Spark application will allocate ``1G`` rather than ``2G`` for the 
driver jvm.
diff --git a/docs/monitor/logging.md b/docs/monitor/logging.md
index 9dce6e22a..41a3c4342 100644
--- a/docs/monitor/logging.md
+++ b/docs/monitor/logging.md
@@ -176,16 +176,16 @@ Operation log will show how SQL queries are executed, 
such as query planning, ex
 
 Operation logs can reveal directly to end-users how their queries are being 
executed on the server/engine-side, including some process-oriented 
information, and why their queries are slow or in error.
 
-For example, when you, as an end-user, use `beeline` to connect a Kyuubi 
server and execute query like below.
+For example, when you, as an end-user, use `kyuubi-beeline` to connect a 
Kyuubi server and execute query like below.
 
 ```shell
-bin/beeline -u 
'jdbc:hive2://10.242.189.214:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=kyuubi'
 -n kent -e 'select * from src;'
+kyuubi-beeline -u 
'jdbc:kyuubi://10.242.189.214:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=kyuubi'
 -n kent -e 'select * from src;'
 ```
 
 You will both get the final results and the corresponding operation logs 
telling you the journey of the query.
 
 ```log
-0: jdbc:hive2://10.242.189.214:2181/> select * from src;
+0: jdbc:kyuubi://10.242.189.214:2181/> select * from src;
 2021-10-27 17:00:19.399 INFO operation.ExecuteStatement: Processing kent's 
query[fb5f57d2-2b50-4a46-961b-3a5c6a2d2597]: INITIALIZED_STATE -> 
PENDING_STATE, statement: select * from src
 2021-10-27 17:00:19.401 INFO operation.ExecuteStatement: Processing kent's 
query[fb5f57d2-2b50-4a46-961b-3a5c6a2d2597]: PENDING_STATE -> RUNNING_STATE, 
statement: select * from src
 2021-10-27 17:00:19.400 INFO operation.ExecuteStatement: Processing kent's 
query[26e169a2-6c06-450a-b758-e577ac673d70]: INITIALIZED_STATE -> 
PENDING_STATE, statement: select * from src
diff --git a/docs/overview/architecture.md b/docs/overview/architecture.md
index 4df5e24a4..f612f4868 100644
--- a/docs/overview/architecture.md
+++ b/docs/overview/architecture.md
@@ -116,7 +116,7 @@ With these features, Kyuubi provides a two-level elastic 
resource management arc
 For example,
 
 ```shell
-./beeline -u "jdbc:hive2://kyuubi.org:10009/;\
+bin/kyuubi-beeline -u "jdbc:kyuubi://kyuubi.org:10009/;\
 hive.server2.proxy.user=tom#\
 spark.yarn.queue=thequeue;\
 spark.dynamicAllocation.enabled=true;\
diff --git a/docs/quick_start/quick_start.rst b/docs/quick_start/quick_start.rst
index edb801feb..232903872 100644
--- a/docs/quick_start/quick_start.rst
+++ b/docs/quick_start/quick_start.rst
@@ -42,7 +42,7 @@ pre-installed and the ``JAVA_HOME`` is correctly set to each 
component.
   **Java**         JRE          8/11/17              Officially released 
against JDK8
   **Kyuubi**       Gateway      \ |release| \        - Kyuubi Server
                    Engine lib                        - Kyuubi Engine
-                   Beeline                           - Kyuubi Hive Beeline
+                   Beeline                           - Kyuubi Beeline
   **Spark**        Engine       3.1 to 3.5           A Spark distribution
   **Flink**        Engine       1.16 to 1.19         A Flink distribution
   **Trino**        Engine       N/A                  A Trino cluster allows to 
access via trino-client v411
@@ -187,7 +187,7 @@ And you are able to get the JDBC connection URL from the 
log file -
 
 For example,
 
-  Starting and exposing JDBC connection at: jdbc:hive2://localhost:10009/
+  Starting and exposing JDBC connection at: jdbc:kyuubi://localhost:10009/
 
 If something goes wrong, you shall be able to find some clues in the log file 
too.
 
@@ -206,7 +206,7 @@ If something goes wrong, you shall be able to find some 
clues in the log file to
 Operate Clients
 ---------------
 
-Kyuubi delivers a beeline client, enabling a similar experience to Apache Hive 
use cases.
+Kyuubi delivers a kyuubi-beeline client, enabling a similar experience to 
Apache Hive use cases.
 
 Open Connections
 ~~~~~~~~~~~~~~~~
@@ -216,21 +216,21 @@ for the following JDBC URL. The case below open a session 
for user named `apache
 
 .. code-block::
 
-   $ bin/beeline -u 'jdbc:hive2://localhost:10009/' -n apache
+   $ bin/kyuubi-beeline -u 'jdbc:kyuubi://localhost:10009/' -n apache
 
 .. note::
    :class: toggle
 
-   Use `--help` to display the usage guide for the beeline tool.
+   Use `--help` to display the usage guide for the kyuubi-beeline tool.
 
    .. code-block::
 
-      $ bin/beeline --help
+      $ bin/kyuubi-beeline --help
 
 Execute Statements
 ~~~~~~~~~~~~~~~~~~
 
-After successfully connected with the server, you can run sql queries in the 
beeline
+After successfully connected with the server, you can run sql queries in the 
kyuubi-beeline
 console. For instance,
 
 .. code-block::
@@ -238,7 +238,7 @@ console. For instance,
 
    > SHOW DATABASES;
 
-You will see a wall of operation logs, and a result table in the beeline 
console.
+You will see a wall of operation logs, and a result table in the 
kyuubi-beeline console.
 
 .. code-block::
 
@@ -264,19 +264,19 @@ started.
 
 .. code-block::
 
-   $ bin/beeline -u 'jdbc:hive2://localhost:10009/' -n kentyao
+   $ bin/kyuubi-beeline -u 'jdbc:kyuubi://localhost:10009/' -n kentyao
 
 This may change depending on the `engine share level`_ you set.
 
 Close Connections
 ~~~~~~~~~~~~~~~~~
 
-Close the session between beeline and Kyuubi server by executing `!quit`, for 
example,
+Close the session between kyuubi-beeline and Kyuubi server by executing 
`!quit`, for example,
 
 .. code-block::
 
    > !quit
-   Closing: 0: jdbc:hive2://localhost:10009/
+   Closing: 0: jdbc:kyuubi://localhost:10009/
 
 Stop Engines
 ~~~~~~~~~~~~
@@ -289,7 +289,7 @@ mean terminations of engines. It depends on both the 
`engine share level`_ and
 Stop Kyuubi
 -----------
 
-Stop Kyuubi by running the following in the `$KYUUBI_HOME` directory:
+Stop Kyuubi which running at the background by performing the following in the 
`$KYUUBI_HOME` directory:
 
 .. code-block::
 

Reply via email to