yaooqinn commented on a change in pull request #1133:
URL: https://github.com/apache/incubator-kyuubi/pull/1133#discussion_r714495511
##########
File path:
kyuubi-hive-jdbc/src/test/scala/org/apache/kyuubi/jdbc/KyuubiDriverSuite.scala
##########
@@ -19,33 +19,55 @@ package org.apache.kyuubi.jdbc
import java.util.Properties
+import org.apache.kyuubi.IcebergSuiteMixin
import org.apache.kyuubi.engine.spark.WithSparkSQLEngine
import org.apache.kyuubi.engine.spark.shim.SparkCatalogShim
+import org.apache.kyuubi.tags.IcebergTest
-class KyuubiDriverSuite extends WithSparkSQLEngine {
+@IcebergTest
+class KyuubiDriverSuite extends WithSparkSQLEngine with IcebergSuiteMixin {
+
+ override def withKyuubiConf: Map[String, String] = extraConfigs -
+ "spark.sql.defaultCatalog" -
+ "spark.sql.catalog.spark_catalog"
+
+ override def afterAll(): Unit = {
+ super.afterAll()
+ for ((k, _) <- withKyuubiConf) {
+ System.clearProperty(k)
+ }
+ }
test("get tables with kyuubi driver") {
val kyuubiDriver = new KyuubiDriver()
val connection = kyuubiDriver.connect(getJdbcUrl, new Properties())
+ assert(connection.isInstanceOf[KyuubiConnection])
+ val metaData = connection.getMetaData
+ assert(metaData.isInstanceOf[KyuubiDatabaseMetaData])
val statement = connection.createStatement()
- val table = s"${SparkCatalogShim.SESSION_CATALOG}.default.kyuubi_hive_jdbc"
+ val table1 =
s"${SparkCatalogShim.SESSION_CATALOG}.default.kyuubi_hive_jdbc"
+ val table2 = s"$catalog.default.hdp_cat_tbl"
try {
- statement.execute(s"CREATE TABLE ${table}(key int) using parquet")
- assert(connection.isInstanceOf[KyuubiConnection])
- val metaData = connection.getMetaData
- assert(metaData.isInstanceOf[KyuubiDatabaseMetaData])
- val resultSet = metaData.getTables(SparkCatalogShim.SESSION_CATALOG,
"default", "%", null)
- assert(resultSet.next())
- assert(resultSet.getString(1) === SparkCatalogShim.SESSION_CATALOG)
- assert(resultSet.getString(2) === "default")
- assert(resultSet.getString(3) === "kyuubi_hive_jdbc")
+ statement.execute(s"CREATE TABLE $table1(key int) using parquet")
+ statement.execute(s"CREATE TABLE $table2(key int) using $format")
+
+ val resultSet1 = metaData.getTables(SparkCatalogShim.SESSION_CATALOG,
"default", "%", null)
+ assert(resultSet1.next())
+ assert(resultSet1.getString(1) === SparkCatalogShim.SESSION_CATALOG)
+ assert(resultSet1.getString(2) === "default")
+ assert(resultSet1.getString(3) === "kyuubi_hive_jdbc")
+
+ val resultSet2 = metaData.getTables(catalog, "default", "%", null)
+ assert(resultSet2.next())
+ assert(resultSet2.getString(1) === catalog)
+ assert(resultSet2.getString(2) === "default")
+ assert(resultSet2.getString(3) === "hdp_cat_tbl")
} finally {
- statement.execute(s"DROP TABLE ${table}")
+ statement.execute(s"DROP TABLE $table1")
Review comment:
nit: IF EXISTS
##########
File path:
kyuubi-hive-jdbc/src/test/scala/org/apache/kyuubi/jdbc/KyuubiDriverSuite.scala
##########
@@ -19,33 +19,55 @@ package org.apache.kyuubi.jdbc
import java.util.Properties
+import org.apache.kyuubi.IcebergSuiteMixin
import org.apache.kyuubi.engine.spark.WithSparkSQLEngine
import org.apache.kyuubi.engine.spark.shim.SparkCatalogShim
+import org.apache.kyuubi.tags.IcebergTest
-class KyuubiDriverSuite extends WithSparkSQLEngine {
+@IcebergTest
+class KyuubiDriverSuite extends WithSparkSQLEngine with IcebergSuiteMixin {
+
+ override def withKyuubiConf: Map[String, String] = extraConfigs -
+ "spark.sql.defaultCatalog" -
+ "spark.sql.catalog.spark_catalog"
+
+ override def afterAll(): Unit = {
+ super.afterAll()
+ for ((k, _) <- withKyuubiConf) {
+ System.clearProperty(k)
+ }
+ }
test("get tables with kyuubi driver") {
val kyuubiDriver = new KyuubiDriver()
val connection = kyuubiDriver.connect(getJdbcUrl, new Properties())
+ assert(connection.isInstanceOf[KyuubiConnection])
+ val metaData = connection.getMetaData
+ assert(metaData.isInstanceOf[KyuubiDatabaseMetaData])
val statement = connection.createStatement()
- val table = s"${SparkCatalogShim.SESSION_CATALOG}.default.kyuubi_hive_jdbc"
+ val table1 =
s"${SparkCatalogShim.SESSION_CATALOG}.default.kyuubi_hive_jdbc"
+ val table2 = s"$catalog.default.hdp_cat_tbl"
try {
- statement.execute(s"CREATE TABLE ${table}(key int) using parquet")
- assert(connection.isInstanceOf[KyuubiConnection])
- val metaData = connection.getMetaData
- assert(metaData.isInstanceOf[KyuubiDatabaseMetaData])
- val resultSet = metaData.getTables(SparkCatalogShim.SESSION_CATALOG,
"default", "%", null)
- assert(resultSet.next())
- assert(resultSet.getString(1) === SparkCatalogShim.SESSION_CATALOG)
- assert(resultSet.getString(2) === "default")
- assert(resultSet.getString(3) === "kyuubi_hive_jdbc")
+ statement.execute(s"CREATE TABLE $table1(key int) using parquet")
+ statement.execute(s"CREATE TABLE $table2(key int) using $format")
+
+ val resultSet1 = metaData.getTables(SparkCatalogShim.SESSION_CATALOG,
"default", "%", null)
+ assert(resultSet1.next())
+ assert(resultSet1.getString(1) === SparkCatalogShim.SESSION_CATALOG)
+ assert(resultSet1.getString(2) === "default")
+ assert(resultSet1.getString(3) === "kyuubi_hive_jdbc")
+
+ val resultSet2 = metaData.getTables(catalog, "default", "%", null)
+ assert(resultSet2.next())
+ assert(resultSet2.getString(1) === catalog)
+ assert(resultSet2.getString(2) === "default")
+ assert(resultSet2.getString(3) === "hdp_cat_tbl")
} finally {
- statement.execute(s"DROP TABLE ${table}")
+ statement.execute(s"DROP TABLE $table1")
+ statement.execute(s"DROP TABLE $table2")
Review comment:
ditto
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]