This is an automated email from the ASF dual-hosted git repository.
yao pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/kyuubi.git
The following commit(s) were added to refs/heads/master by this push:
new 9306e5762 [KYUUBI #5726][AUTHZ] Support optimize path-based table for
Delta Lake in Authz
9306e5762 is described below
commit 9306e5762683b59d9f02379853a53aabca46415d
Author: zml1206 <[email protected]>
AuthorDate: Tue Nov 21 17:18:02 2023 +0800
[KYUUBI #5726][AUTHZ] Support optimize path-based table for Delta Lake in
Authz
# :mag: Description
## Issue References ๐
This pull request fixes #5726.
## Describe Your Solution ๐ง
`org.apache.spark.sql.delta.commands.OptimizeTableCommand` add uriDescs.
## Types of changes :bookmark:
- [ ] Bugfix (non-breaking change which fixes an issue)
- [x] New feature (non-breaking change which adds functionality)
- [ ] Breaking change (fix or feature that would cause existing
functionality to change)
## Test Plan ๐งช
#### Behavior Without This Pull Request :coffin:
#### Behavior With This Pull Request :tada:
#### Related Unit Tests
org.apache.kyuubi.plugin.spark.authz.ranger.DeltaCatalogRangerSparkExtensionSuite.test("optimize
path-based table")
---
# Checklists
## ๐ Author Self Checklist
- [x] My code follows the [style
guidelines](https://kyuubi.readthedocs.io/en/master/contributing/code/style.html)
of this project
- [x] I have performed a self-review
- [ ] I have commented my code, particularly in hard-to-understand areas
- [ ] I have made corresponding changes to the documentation
- [x] My changes generate no new warnings
- [x] I have added tests that prove my fix is effective or that my feature
works
- [ ] New and existing unit tests pass locally with my changes
- [x] This patch was not authored or co-authored using [Generative
Tooling](https://www.apache.org/legal/generative-tooling.html)
## ๐ Committer Pre-Merge Checklist
- [x] Pull request title is okay.
- [x] No license issues.
- [x] Milestone correctly set?
- [x] Test coverage is ok
- [x] Assignees are selected.
- [x] Minimum number of approvals
- [x] No changes are requested
**Be nice. Be informative.**
Closes #5732 from zml1206/KYUUBI-5726.
Closes #5726
e4cab7a02 [zml1206] update
62c52b20b [zml1206] Support optimize path-based table for Delta Lake in
Authz
Authored-by: zml1206 <[email protected]>
Signed-off-by: Kent Yao <[email protected]>
---
...he.kyuubi.plugin.spark.authz.serde.URIExtractor | 3 +++
.../src/main/resources/table_command_spec.json | 14 ++++++++++++-
.../plugin/spark/authz/serde/tableExtractors.scala | 23 ++++++++++++---------
.../plugin/spark/authz/serde/uriExtractors.scala | 24 ++++++++++++++++++++++
.../plugin/spark/authz/gen/DeltaCommands.scala | 6 +++++-
.../DeltaCatalogRangerSparkExtensionSuite.scala | 21 ++++++++++++++++++-
6 files changed, 78 insertions(+), 13 deletions(-)
diff --git
a/extensions/spark/kyuubi-spark-authz/src/main/resources/META-INF/services/org.apache.kyuubi.plugin.spark.authz.serde.URIExtractor
b/extensions/spark/kyuubi-spark-authz/src/main/resources/META-INF/services/org.apache.kyuubi.plugin.spark.authz.serde.URIExtractor
index 60f761233..460dfeb01 100644
---
a/extensions/spark/kyuubi-spark-authz/src/main/resources/META-INF/services/org.apache.kyuubi.plugin.spark.authz.serde.URIExtractor
+++
b/extensions/spark/kyuubi-spark-authz/src/main/resources/META-INF/services/org.apache.kyuubi.plugin.spark.authz.serde.URIExtractor
@@ -23,7 +23,10 @@
org.apache.kyuubi.plugin.spark.authz.serde.IdentifierURIExtractor
org.apache.kyuubi.plugin.spark.authz.serde.PartitionLocsSeqURIExtractor
org.apache.kyuubi.plugin.spark.authz.serde.PropertiesLocationUriExtractor
org.apache.kyuubi.plugin.spark.authz.serde.PropertiesPathUriExtractor
+org.apache.kyuubi.plugin.spark.authz.serde.ResolvedTableURIExtractor
org.apache.kyuubi.plugin.spark.authz.serde.StringSeqURIExtractor
org.apache.kyuubi.plugin.spark.authz.serde.StringURIExtractor
org.apache.kyuubi.plugin.spark.authz.serde.SubqueryAliasURIExtractor
+org.apache.kyuubi.plugin.spark.authz.serde.TableIdentifierOptionURIExtractor
+org.apache.kyuubi.plugin.spark.authz.serde.TableIdentifierURIExtractor
org.apache.kyuubi.plugin.spark.authz.serde.TableSpecURIExtractor
diff --git
a/extensions/spark/kyuubi-spark-authz/src/main/resources/table_command_spec.json
b/extensions/spark/kyuubi-spark-authz/src/main/resources/table_command_spec.json
index 583ad29b4..a133c981c 100644
---
a/extensions/spark/kyuubi-spark-authz/src/main/resources/table_command_spec.json
+++
b/extensions/spark/kyuubi-spark-authz/src/main/resources/table_command_spec.json
@@ -2226,7 +2226,19 @@
} ],
"opType" : "ALTERTABLE_COMPACT",
"queryDescs" : [ ],
- "uriDescs" : [ ]
+ "uriDescs" : [ {
+ "fieldName" : "child",
+ "fieldExtractor" : "ResolvedTableURIExtractor",
+ "isInput" : false
+ }, {
+ "fieldName" : "tableId",
+ "fieldExtractor" : "TableIdentifierOptionURIExtractor",
+ "isInput" : false
+ }, {
+ "fieldName" : "path",
+ "fieldExtractor" : "StringURIExtractor",
+ "isInput" : false
+ } ]
}, {
"classname" : "org.apache.spark.sql.delta.commands.UpdateCommand",
"tableDescs" : [ {
diff --git
a/extensions/spark/kyuubi-spark-authz/src/main/scala/org/apache/kyuubi/plugin/spark/authz/serde/tableExtractors.scala
b/extensions/spark/kyuubi-spark-authz/src/main/scala/org/apache/kyuubi/plugin/spark/authz/serde/tableExtractors.scala
index adff93474..8a7bc4522 100644
---
a/extensions/spark/kyuubi-spark-authz/src/main/scala/org/apache/kyuubi/plugin/spark/authz/serde/tableExtractors.scala
+++
b/extensions/spark/kyuubi-spark-authz/src/main/scala/org/apache/kyuubi/plugin/spark/authz/serde/tableExtractors.scala
@@ -17,8 +17,7 @@
package org.apache.kyuubi.plugin.spark.authz.serde
-import java.util.{Map => JMap}
-import java.util.LinkedHashMap
+import java.util.{LinkedHashMap, Map => JMap}
import scala.collection.JavaConverters._
@@ -81,14 +80,18 @@ object TableExtractor {
class TableIdentifierTableExtractor extends TableExtractor {
override def apply(spark: SparkSession, v1: AnyRef): Option[Table] = {
val identifier = v1.asInstanceOf[TableIdentifier]
- val owner =
- try {
- val catalogTable =
spark.sessionState.catalog.getTableMetadata(identifier)
- Option(catalogTable.owner).filter(_.nonEmpty)
- } catch {
- case _: Exception => None
- }
- Some(Table(None, identifier.database, identifier.table, owner))
+ if (isPathIdentifier(identifier.table, spark)) {
+ None
+ } else {
+ val owner =
+ try {
+ val catalogTable =
spark.sessionState.catalog.getTableMetadata(identifier)
+ Option(catalogTable.owner).filter(_.nonEmpty)
+ } catch {
+ case _: Exception => None
+ }
+ Some(Table(None, identifier.database, identifier.table, owner))
+ }
}
}
diff --git
a/extensions/spark/kyuubi-spark-authz/src/main/scala/org/apache/kyuubi/plugin/spark/authz/serde/uriExtractors.scala
b/extensions/spark/kyuubi-spark-authz/src/main/scala/org/apache/kyuubi/plugin/spark/authz/serde/uriExtractors.scala
index fd0b19420..434cc7699 100644
---
a/extensions/spark/kyuubi-spark-authz/src/main/scala/org/apache/kyuubi/plugin/spark/authz/serde/uriExtractors.scala
+++
b/extensions/spark/kyuubi-spark-authz/src/main/scala/org/apache/kyuubi/plugin/spark/authz/serde/uriExtractors.scala
@@ -18,6 +18,7 @@
package org.apache.kyuubi.plugin.spark.authz.serde
import org.apache.spark.sql.SparkSession
+import org.apache.spark.sql.catalyst.TableIdentifier
import org.apache.spark.sql.catalyst.catalog.{CatalogStorageFormat,
CatalogTable}
import org.apache.spark.sql.catalyst.plans.logical.{LogicalPlan, SubqueryAlias}
import org.apache.spark.sql.connector.catalog.Identifier
@@ -133,3 +134,26 @@ class DataSourceV2RelationURIExtractor extends
URIExtractor {
}
}
}
+
+class ResolvedTableURIExtractor extends URIExtractor {
+ override def apply(spark: SparkSession, v1: AnyRef): Seq[Uri] = {
+ val identifier = invokeAs[AnyRef](v1, "identifier")
+ lookupExtractor[IdentifierURIExtractor].apply(spark, identifier)
+ }
+}
+
+class TableIdentifierURIExtractor extends URIExtractor {
+ override def apply(spark: SparkSession, v1: AnyRef): Seq[Uri] = v1 match {
+ case tableIdentifier: TableIdentifier if
isPathIdentifier(tableIdentifier.table, spark) =>
+ Seq(tableIdentifier.table).map(Uri)
+ case _ => Nil
+ }
+}
+
+class TableIdentifierOptionURIExtractor extends URIExtractor {
+ override def apply(spark: SparkSession, v1: AnyRef): Seq[Uri] = v1 match {
+ case Some(tableIdentifier: TableIdentifier) =>
+ lookupExtractor[TableIdentifierURIExtractor].apply(spark,
tableIdentifier)
+ case _ => Nil
+ }
+}
diff --git
a/extensions/spark/kyuubi-spark-authz/src/test/scala/org/apache/kyuubi/plugin/spark/authz/gen/DeltaCommands.scala
b/extensions/spark/kyuubi-spark-authz/src/test/scala/org/apache/kyuubi/plugin/spark/authz/gen/DeltaCommands.scala
index db72cde83..ed7c89bed 100644
---
a/extensions/spark/kyuubi-spark-authz/src/test/scala/org/apache/kyuubi/plugin/spark/authz/gen/DeltaCommands.scala
+++
b/extensions/spark/kyuubi-spark-authz/src/test/scala/org/apache/kyuubi/plugin/spark/authz/gen/DeltaCommands.scala
@@ -49,7 +49,11 @@ object DeltaCommands extends CommandSpecs[TableCommandSpec] {
val cmd = "org.apache.spark.sql.delta.commands.OptimizeTableCommand"
val childDesc = TableDesc("child", classOf[ResolvedTableTableExtractor])
val tableDesc = TableDesc("tableId",
classOf[TableIdentifierOptionTableExtractor])
- TableCommandSpec(cmd, Seq(childDesc, tableDesc), ALTERTABLE_COMPACT)
+ val uriDescs = Seq(
+ UriDesc("child", classOf[ResolvedTableURIExtractor]),
+ UriDesc("tableId", classOf[TableIdentifierOptionURIExtractor]),
+ UriDesc("path", classOf[StringURIExtractor]))
+ TableCommandSpec(cmd, Seq(childDesc, tableDesc), ALTERTABLE_COMPACT,
uriDescs = uriDescs)
}
val VacuumTableCommand = {
diff --git
a/extensions/spark/kyuubi-spark-authz/src/test/scala/org/apache/kyuubi/plugin/spark/authz/ranger/DeltaCatalogRangerSparkExtensionSuite.scala
b/extensions/spark/kyuubi-spark-authz/src/test/scala/org/apache/kyuubi/plugin/spark/authz/ranger/DeltaCatalogRangerSparkExtensionSuite.scala
index 801f9f745..f4243835f 100644
---
a/extensions/spark/kyuubi-spark-authz/src/test/scala/org/apache/kyuubi/plugin/spark/authz/ranger/DeltaCatalogRangerSparkExtensionSuite.scala
+++
b/extensions/spark/kyuubi-spark-authz/src/test/scala/org/apache/kyuubi/plugin/spark/authz/ranger/DeltaCatalogRangerSparkExtensionSuite.scala
@@ -283,7 +283,7 @@ class DeltaCatalogRangerSparkExtensionSuite extends
RangerSparkExtensionSuite {
}
test("optimize table") {
- assume(isSparkV32OrGreater)
+ assume(isSparkV32OrGreater, "optimize table is available in Delta Lake
1.2.0 and above")
withCleanTmpResources(Seq((s"$namespace1.$table1", "table"),
(s"$namespace1", "database"))) {
doAs(admin, sql(s"CREATE DATABASE IF NOT EXISTS $namespace1"))
@@ -432,6 +432,25 @@ class DeltaCatalogRangerSparkExtensionSuite extends
RangerSparkExtensionSuite {
}
}
}
+
+ test("optimize path-based table") {
+ assume(isSparkV32OrGreater, "optimize table is available in Delta Lake
1.2.0 and above")
+
+ withTempDir(path => {
+ doAs(admin, sql(createPathBasedTableSql(path)))
+ val optimizeTableSql1 = s"OPTIMIZE delta.`$path`"
+ interceptContains[AccessControlException](
+ doAs(someone, sql(optimizeTableSql1)))(
+ s"does not have [write] privilege on [[$path, $path/]]")
+ doAs(admin, sql(optimizeTableSql1))
+
+ val optimizeTableSql2 = s"OPTIMIZE '$path'"
+ interceptContains[AccessControlException](
+ doAs(someone, sql(optimizeTableSql2)))(
+ s"does not have [write] privilege on [[$path, $path/]]")
+ doAs(admin, sql(optimizeTableSql2))
+ })
+ }
}
object DeltaCatalogRangerSparkExtensionSuite {