This is an automated email from the ASF dual-hosted git repository.
chengpan pushed a commit to branch branch-1.10
in repository https://gitbox.apache.org/repos/asf/kyuubi.git
The following commit(s) were added to refs/heads/branch-1.10 by this push:
new 61a4641839 [KYUUBI #6925] Only run Paimon authz tests with Scala 2.12
61a4641839 is described below
commit 61a46418394c131beab96c1000e7facf884ba11e
Author: Cheng Pan <[email protected]>
AuthorDate: Wed Feb 19 14:19:22 2025 +0800
[KYUUBI #6925] Only run Paimon authz tests with Scala 2.12
### Why are the changes needed?
Paimon does not seem to support Scala 2.13
### How was this patch tested?
Pass GHA.
### Was this patch authored or co-authored using generative AI tooling?
No.
Closes #6925 from pan3793/authz-paimon-scala212.
Closes #6925
865a7dd72 [Cheng Pan] fix
971d23273 [Cheng Pan] Update
extensions/spark/kyuubi-spark-authz/src/test/scala/org/apache/kyuubi/plugin/spark/authz/ranger/PaimonCatalogRangerSparkExtensionSuite.scala
499f10ab0 [Cheng Pan] Only run Paimon authz tests with Scala 2.12
Authored-by: Cheng Pan <[email protected]>
Signed-off-by: Cheng Pan <[email protected]>
(cherry picked from commit 93ac1ee26980ef01793b100e382fdb531a9d956e)
Signed-off-by: Cheng Pan <[email protected]>
---
.../plugin/spark/authz/util/AuthZUtils.scala | 3 +-
.../PaimonCatalogRangerSparkExtensionSuite.scala | 39 ++++++++--------------
2 files changed, 16 insertions(+), 26 deletions(-)
diff --git
a/extensions/spark/kyuubi-spark-authz/src/main/scala/org/apache/kyuubi/plugin/spark/authz/util/AuthZUtils.scala
b/extensions/spark/kyuubi-spark-authz/src/main/scala/org/apache/kyuubi/plugin/spark/authz/util/AuthZUtils.scala
index 523b866beb..1c2ad5f5d3 100644
---
a/extensions/spark/kyuubi-spark-authz/src/main/scala/org/apache/kyuubi/plugin/spark/authz/util/AuthZUtils.scala
+++
b/extensions/spark/kyuubi-spark-authz/src/main/scala/org/apache/kyuubi/plugin/spark/authz/util/AuthZUtils.scala
@@ -92,7 +92,8 @@ private[authz] object AuthZUtils {
lazy val SCALA_RUNTIME_VERSION: SemanticVersion =
SemanticVersion(scala.util.Properties.versionNumberString)
- lazy val isScalaV213: Boolean = SCALA_RUNTIME_VERSION >= "2.13"
+ lazy val isScalaV212: Boolean = SCALA_RUNTIME_VERSION === "2.12"
+ lazy val isScalaV213: Boolean = SCALA_RUNTIME_VERSION === "2.13"
def derbyJdbcDriverClass: String = if (isSparkV40OrGreater) {
// SPARK-46257 (Spark 4.0.0) moves to Derby 10.16
diff --git
a/extensions/spark/kyuubi-spark-authz/src/test/scala/org/apache/kyuubi/plugin/spark/authz/ranger/PaimonCatalogRangerSparkExtensionSuite.scala
b/extensions/spark/kyuubi-spark-authz/src/test/scala/org/apache/kyuubi/plugin/spark/authz/ranger/PaimonCatalogRangerSparkExtensionSuite.scala
index 81fe228d22..9cbd534d0f 100644
---
a/extensions/spark/kyuubi-spark-authz/src/test/scala/org/apache/kyuubi/plugin/spark/authz/ranger/PaimonCatalogRangerSparkExtensionSuite.scala
+++
b/extensions/spark/kyuubi-spark-authz/src/test/scala/org/apache/kyuubi/plugin/spark/authz/ranger/PaimonCatalogRangerSparkExtensionSuite.scala
@@ -16,13 +16,12 @@
*/
package org.apache.kyuubi.plugin.spark.authz.ranger
-import scala.util.Properties
-
import org.scalatest.Outcome
import org.apache.kyuubi.Utils
import org.apache.kyuubi.plugin.spark.authz.AccessControlException
import org.apache.kyuubi.plugin.spark.authz.RangerTestUsers._
+import org.apache.kyuubi.plugin.spark.authz.util.AuthZUtils._
import org.apache.kyuubi.tags.PaimonTest
import org.apache.kyuubi.util.AssertionUtils._
@@ -32,9 +31,8 @@ import org.apache.kyuubi.util.AssertionUtils._
@PaimonTest
class PaimonCatalogRangerSparkExtensionSuite extends RangerSparkExtensionSuite
{
override protected val catalogImpl: String = "hive"
- private def isSupportedVersion = true
+ private def isSupportedVersion = isScalaV212
- val scalaVersion: String = Properties.versionString
val catalogV2 = "paimon_catalog"
val namespace1 = "paimon_ns"
val table1 = "table1"
@@ -51,9 +49,8 @@ class PaimonCatalogRangerSparkExtensionSuite extends
RangerSparkExtensionSuite {
s"spark.sql.catalog.$catalogV2.warehouse",
Utils.createTempDir(catalogV2).toString)
super.beforeAll()
+ doAs(admin, sql(s"CREATE DATABASE IF NOT EXISTS $catalogV2.$namespace1"))
}
-
- doAs(admin, sql(s"CREATE DATABASE IF NOT EXISTS $catalogV2.$namespace1"))
}
override def afterAll(): Unit = {
@@ -87,28 +84,20 @@ class PaimonCatalogRangerSparkExtensionSuite extends
RangerSparkExtensionSuite {
test("[KYUUBI #6541] INSERT/SELECT TABLE") {
val tName = "t_paimon"
+ withCleanTmpResources(Seq((s"$catalogV2.$namespace1.$tName", "table"))) {
+ doAs(bob, sql(createTableSql(namespace1, tName)))
- /**
- * paimon-spark run on Scala 2.12.
- */
- if (scalaVersion.startsWith("version 2.12")) {
- withCleanTmpResources(Seq((s"$catalogV2.$namespace1.$tName", "table"))) {
-
- doAs(bob, sql(createTableSql(namespace1, tName)))
-
- interceptEndsWith[AccessControlException] {
- doAs(someone, sql(s"INSERT INTO $catalogV2.$namespace1.$tName VALUES
(1, 'name_1')"))
- }(s"does not have [update] privilege on [$namespace1/$tName]")
- doAs(bob, sql(s"INSERT INTO $catalogV2.$namespace1.$tName VALUES (1,
'name_1')"))
- doAs(bob, sql(s"INSERT INTO $catalogV2.$namespace1.$tName VALUES (1,
'name_2')"))
+ interceptEndsWith[AccessControlException] {
+ doAs(someone, sql(s"INSERT INTO $catalogV2.$namespace1.$tName VALUES
(1, 'name_1')"))
+ }(s"does not have [update] privilege on [$namespace1/$tName]")
+ doAs(bob, sql(s"INSERT INTO $catalogV2.$namespace1.$tName VALUES (1,
'name_1')"))
+ doAs(bob, sql(s"INSERT INTO $catalogV2.$namespace1.$tName VALUES (1,
'name_2')"))
- interceptEndsWith[AccessControlException] {
- doAs(someone, sql(s"SELECT id FROM
$catalogV2.$namespace1.$tName").show())
- }(s"does not have [select] privilege on [$namespace1/$tName/id]")
- doAs(bob, sql(s"SELECT name FROM
$catalogV2.$namespace1.$tName").show())
- }
+ interceptEndsWith[AccessControlException] {
+ doAs(someone, sql(s"SELECT id FROM
$catalogV2.$namespace1.$tName").show())
+ }(s"does not have [select] privilege on [$namespace1/$tName/id]")
+ doAs(bob, sql(s"SELECT name FROM $catalogV2.$namespace1.$tName").show())
}
-
}
def createTableSql(namespace: String, table: String): String =