yuqi1129 commented on code in PR #6100:
URL: https://github.com/apache/gravitino/pull/6100#discussion_r1903533940


##########
catalogs/catalog-hive/src/test/java/org/apache/gravitino/catalog/hive/integration/test/CatalogHiveIT.java:
##########
@@ -229,7 +229,14 @@ public void stop() throws IOException {
                 catalog.asSchemas().dropSchema(schema, true);
               }));
       Arrays.stream(metalake.listCatalogs())
-          .forEach((catalogName -> metalake.dropCatalog(catalogName, true)));
+          .forEach(
+              catalogName -> {
+                try {
+                  metalake.dropCatalog(catalogName, true);
+                } catch (Exception e) {
+                  // Ignore exception

Review Comment:
   Why are we making these changes?  



##########
core/src/main/java/org/apache/gravitino/authorization/AuthorizationUtils.java:
##########
@@ -364,4 +378,129 @@ private static void checkCatalogType(
           catalogIdent, catalog.type(), privilege);
     }
   }
+
+  public static List<String> getMetadataObjectLocation(
+      NameIdentifier ident, Entity.EntityType type) {
+    List<String> locations = new ArrayList<>();
+    MetadataObject metadataObject;
+    try {
+      metadataObject = NameIdentifierUtil.toMetadataObject(ident, type);
+    } catch (IllegalArgumentException e) {
+      LOG.warn("Illegal argument exception for metadata object %s type %s", 
ident, type, e);
+      return locations;
+    }
+
+    String metalake =
+        (type == Entity.EntityType.METALAKE ? ident.name() : 
ident.namespace().level(0));
+    try {
+      switch (metadataObject.type()) {
+        case METALAKE:
+          {
+            NameIdentifier[] identifiers =
+                
GravitinoEnv.getInstance().catalogDispatcher().listCatalogs(Namespace.of(metalake));
+            List<String> finalLocationPath = locations;

Review Comment:
   What's the `finalLocationPath` used for?



##########
authorizations/authorization-ranger/src/main/java/org/apache/gravitino/authorization/ranger/RangerAuthorizationHadoopSQLPlugin.java:
##########
@@ -76,6 +85,328 @@ public Map<Privilege.Name, Set<AuthorizationPrivilege>> 
privilegesMappingRule()
         ImmutableSet.of(RangerHadoopSQLPrivilege.READ, 
RangerHadoopSQLPrivilege.SELECT));
   }
 
+  /**
+   * Find the managed policy for the ranger securable object.
+   *
+   * @param authzMetadataObject The ranger securable object to find the 
managed policy.
+   * @return The managed policy for the metadata object.
+   */
+  public RangerPolicy findManagedPolicy(AuthorizationMetadataObject 
authzMetadataObject)
+      throws AuthorizationPluginException {
+    List<RangerPolicy> policies = wildcardSearchPolies(authzMetadataObject);
+    if (!policies.isEmpty()) {
+      /**
+       * Because Ranger doesn't support the precise search, Ranger will return 
the policy meets the
+       * wildcard(*,?) conditions, If you use `db.table` condition to search 
policy, the Ranger will
+       * match `db1.table1`, `db1.table2`, `db*.table*`, So we need to 
manually precisely filter
+       * this research results.
+       */
+      List<String> nsMetadataObj = authzMetadataObject.names();
+      Map<String, String> preciseFilters = new HashMap<>();
+      for (int i = 0; i < nsMetadataObj.size() && i < 
policyResourceDefinesRule().size(); i++) {
+        preciseFilters.put(policyResourceDefinesRule().get(i), 
nsMetadataObj.get(i));
+      }
+      policies =
+          policies.stream()
+              .filter(
+                  policy ->
+                      policy.getResources().entrySet().stream()
+                          .allMatch(
+                              entry ->
+                                  preciseFilters.containsKey(entry.getKey())
+                                      && entry.getValue().getValues().size() 
== 1
+                                      && entry
+                                          .getValue()
+                                          .getValues()
+                                          
.contains(preciseFilters.get(entry.getKey()))))
+              .collect(Collectors.toList());
+    }
+    // Only return the policies that are managed by Gravitino.
+    if (policies.size() > 1) {
+      throw new AuthorizationPluginException("Each metadata object can have at 
most one policy.");
+    }
+
+    if (policies.isEmpty()) {
+      return null;
+    }
+
+    RangerPolicy policy = policies.get(0);
+    // Delegating Gravitino management policies cannot contain duplicate 
privilege
+    policy.getPolicyItems().forEach(RangerHelper::checkPolicyItemAccess);
+    policy.getDenyPolicyItems().forEach(RangerHelper::checkPolicyItemAccess);
+    
policy.getRowFilterPolicyItems().forEach(RangerHelper::checkPolicyItemAccess);
+    
policy.getDataMaskPolicyItems().forEach(RangerHelper::checkPolicyItemAccess);
+
+    return policy;
+  }
+
+  /** Wildcard search the Ranger policies in the different Ranger service. */
+  @Override
+  protected List<RangerPolicy> wildcardSearchPolies(
+      AuthorizationMetadataObject authzMetadataObject) {
+    List<String> resourceDefines = policyResourceDefinesRule();
+    Map<String, String> searchFilters = new HashMap<>();
+    searchFilters.put(SearchFilter.SERVICE_NAME, rangerServiceName);
+    for (int i = 0; i < authzMetadataObject.names().size() && i < 
resourceDefines.size(); i++) {
+      searchFilters.put(
+          SearchFilter.RESOURCE_PREFIX + resourceDefines.get(i),
+          authzMetadataObject.names().get(i));
+    }
+
+    try {
+      List<RangerPolicy> policies = rangerClient.findPolicies(searchFilters);
+      return policies;
+    } catch (RangerServiceException e) {
+      throw new AuthorizationPluginException(e, "Failed to find the policies 
in the Ranger");
+    }
+  }
+
+  /**
+   * IF rename the SCHEMA, Need to rename these the relevant policies, 
`{schema}`, `{schema}.*`,
+   * `{schema}.*.*` <br>
+   * IF rename the TABLE, Need to rename these the relevant policies, 
`{schema}.*`, `{schema}.*.*`
+   * <br>
+   * IF rename the COLUMN, Only need to rename `{schema}.*.*` <br>
+   */
+  @Override
+  protected void doRenameMetadataObject(
+      AuthorizationMetadataObject authzMetadataObject,
+      AuthorizationMetadataObject newAuthzMetadataObject) {
+    List<Map<String, String>> loop = new ArrayList<>();

Review Comment:
   I see the code logic is almost the same as that in 
`RangerAuthorizationHDFSPlugin`



##########
core/src/main/java/org/apache/gravitino/authorization/AuthorizationUtils.java:
##########
@@ -364,4 +378,129 @@ private static void checkCatalogType(
           catalogIdent, catalog.type(), privilege);
     }
   }
+
+  public static List<String> getMetadataObjectLocation(
+      NameIdentifier ident, Entity.EntityType type) {
+    List<String> locations = new ArrayList<>();
+    MetadataObject metadataObject;
+    try {
+      metadataObject = NameIdentifierUtil.toMetadataObject(ident, type);
+    } catch (IllegalArgumentException e) {
+      LOG.warn("Illegal argument exception for metadata object %s type %s", 
ident, type, e);
+      return locations;
+    }
+
+    String metalake =
+        (type == Entity.EntityType.METALAKE ? ident.name() : 
ident.namespace().level(0));
+    try {
+      switch (metadataObject.type()) {
+        case METALAKE:
+          {
+            NameIdentifier[] identifiers =
+                
GravitinoEnv.getInstance().catalogDispatcher().listCatalogs(Namespace.of(metalake));
+            List<String> finalLocationPath = locations;
+            Arrays.stream(identifiers)
+                .collect(Collectors.toList())
+                .forEach(
+                    identifier -> {
+                      Catalog catalogObj =
+                          
GravitinoEnv.getInstance().catalogDispatcher().loadCatalog(identifier);
+                      if (catalogObj.provider().equals("hive")) {
+                        Schema schema =
+                            GravitinoEnv.getInstance()
+                                .schemaDispatcher()
+                                .loadSchema(
+                                    NameIdentifier.of(
+                                        metalake,
+                                        catalogObj.name(),
+                                        "default" /*Hive default schema*/));

Review Comment:
   If I'm not mistaken, for metadata `metalake` and `catalog`, you will get the 
location of schema(databases) `default` ?



##########
authorizations/authorization-ranger/src/main/java/org/apache/gravitino/authorization/ranger/RangerAuthorizationHDFSPlugin.java:
##########
@@ -278,88 +696,91 @@ public List<AuthorizationSecurableObject> 
translateOwner(MetadataObject gravitin
   }
 
   @Override
-  public AuthorizationMetadataObject translateMetadataObject(MetadataObject 
metadataObject) {
-    Preconditions.checkArgument(
-        allowMetadataObjectTypesRule().contains(metadataObject.type()),
-        String.format(
-            "The metadata object type %s is not supported in the 
RangerAuthorizationHDFSPlugin",
-            metadataObject.type()));
-    List<String> nsMetadataObject =
-        
Lists.newArrayList(SecurableObjects.DOT_SPLITTER.splitToList(metadataObject.fullName()));
-    Preconditions.checkArgument(
-        nsMetadataObject.size() > 0, "The metadata object must have at least 
one name.");
-
-    PathBasedMetadataObject rangerPathBaseMetadataObject;
-    switch (metadataObject.type()) {
-      case METALAKE:
-      case CATALOG:
-        rangerPathBaseMetadataObject =
-            new PathBasedMetadataObject("", PathBasedMetadataObject.Type.PATH);
-        break;
-      case SCHEMA:
-        rangerPathBaseMetadataObject =
-            new PathBasedMetadataObject(
-                metadataObject.fullName(), PathBasedMetadataObject.Type.PATH);
-        break;
-      case FILESET:
-        rangerPathBaseMetadataObject =
-            new PathBasedMetadataObject(
-                getLocationPath(metadataObject), 
PathBasedMetadataObject.Type.PATH);
-        break;
-      default:
-        throw new AuthorizationPluginException(
-            "The metadata object type %s is not supported in the 
RangerAuthorizationHDFSPlugin",
-            metadataObject.type());
-    }
-    rangerPathBaseMetadataObject.validateAuthorizationMetadataObject();
-    return rangerPathBaseMetadataObject;
-  }
-
-  private NameIdentifier getObjectNameIdentifier(MetadataObject 
metadataObject) {
-    return NameIdentifier.parse(String.format("%s.%s", metalake, 
metadataObject.fullName()));
+  public List<AuthorizationMetadataObject> 
translateMetadataObject(MetadataObject metadataObject) {
+    List<AuthorizationMetadataObject> authzMetadataObjects = new ArrayList<>();
+    Entity.EntityType entityType = 
MetadataObjectUtil.toEntityType(metadataObject);
+    NameIdentifier identifier =
+        metadataObject.type().equals(MetadataObject.Type.METALAKE)
+            ? NameIdentifier.of(metadataObject.fullName())
+            : NameIdentifier.parse(String.join(".", metalake, 
metadataObject.fullName()));
+    List<String> locations = 
AuthorizationUtils.getMetadataObjectLocation(identifier, entityType);
+    locations.stream()
+        .forEach(
+            locationPath -> {
+              PathBasedMetadataObject pathBaseMetadataObject =
+                  new PathBasedMetadataObject(
+                      metadataObject.parent(),
+                      metadataObject.name(),
+                      locationPath,
+                      PathBasedMetadataObject.Type.PATH);
+              pathBaseMetadataObject.validateAuthorizationMetadataObject();
+              authzMetadataObjects.add(pathBaseMetadataObject);
+            });
+    return authzMetadataObjects;
   }
 
-  @VisibleForTesting
-  public String getLocationPath(MetadataObject metadataObject) throws 
NoSuchEntityException {
-    String locationPath = null;
-    switch (metadataObject.type()) {
-      case METALAKE:
-      case SCHEMA:
-      case TABLE:
-        break;
-      case CATALOG:
-        {
-          Namespace nsMetadataObj = 
Namespace.fromString(metadataObject.fullName());
-          NameIdentifier ident = NameIdentifier.of(metalake, 
nsMetadataObj.level(0));
-          Catalog catalog = 
GravitinoEnv.getInstance().catalogDispatcher().loadCatalog(ident);
-          if (catalog.provider().equals("hive")) {
-            Schema schema =
-                GravitinoEnv.getInstance()
-                    .schemaDispatcher()
-                    .loadSchema(
-                        NameIdentifier.of(
-                            metalake, nsMetadataObj.level(0), "default" /*Hive 
default schema*/));
-            String defaultSchemaLocation = 
schema.properties().get(HiveConstants.LOCATION);
-            locationPath = 
pattern.matcher(defaultSchemaLocation).replaceAll("");
-          }
-        }
-        break;
-      case FILESET:
-        FilesetDispatcher filesetDispatcher = 
GravitinoEnv.getInstance().filesetDispatcher();
-        NameIdentifier identifier = getObjectNameIdentifier(metadataObject);
-        Fileset fileset = filesetDispatcher.loadFileset(identifier);
+  @Override
+  public Boolean onMetadataUpdated(MetadataObjectChange... changes) throws 
RuntimeException {
+    for (MetadataObjectChange change : changes) {
+      if (change instanceof MetadataObjectChange.RenameMetadataObject) {
+        MetadataObject metadataObject =
+            ((MetadataObjectChange.RenameMetadataObject) 
change).metadataObject();
+        MetadataObject newMetadataObject =
+            ((MetadataObjectChange.RenameMetadataObject) 
change).newMetadataObject();
         Preconditions.checkArgument(
-            fileset != null, String.format("Fileset %s is not found", 
identifier));
-        String filesetLocation = fileset.storageLocation();
+            metadataObject.type() == newMetadataObject.type(),
+            "The old and new metadata object type must be equal!");
+        if (metadataObject.type() == MetadataObject.Type.METALAKE) {
+          // Rename the metalake name
+          this.metalake = newMetadataObject.name();
+          // Did not need to update the Ranger policy
+          continue;
+        } else if (metadataObject.type() == MetadataObject.Type.CATALOG) {
+          // Did not need to update the Ranger policy
+          continue;
+        }
+        List<AuthorizationMetadataObject> oldAuthzMetadataObjects =
+            translateMetadataObject(metadataObject);
+        List<AuthorizationMetadataObject> newAuthzMetadataObjects =
+            translateMetadataObject(newMetadataObject);
         Preconditions.checkArgument(
-            filesetLocation != null, String.format("Fileset %s location is not 
found", identifier));
-        locationPath = pattern.matcher(filesetLocation).replaceAll("");
-        break;
-      default:
-        throw new AuthorizationPluginException(
-            "The metadata object type %s is not supported in the 
RangerAuthorizationHDFSPlugin",
-            metadataObject.type());
+            oldAuthzMetadataObjects.size() == newAuthzMetadataObjects.size(),
+            "The old and new metadata objects size must be equal!");
+        for (int i = 0; i < oldAuthzMetadataObjects.size(); i++) {
+          AuthorizationMetadataObject oldAuthMetadataObject = 
oldAuthzMetadataObjects.get(i);
+          AuthorizationMetadataObject newAuthzMetadataObject = 
newAuthzMetadataObjects.get(i);
+          if (oldAuthMetadataObject.equals(newAuthzMetadataObject)) {

Review Comment:
   PathBasedMetadataObject has not overwrites `equals`, so there may be 
problems to use equals
   
   ```java
       PathBasedMetadataObject pathBasedMetadataObject1 =
           new PathBasedMetadataObject("parent", "name", "path", 
PathBasedMetadataObject.Type.PATH);
       pathBasedMetadataObject1.validateAuthorizationMetadataObject();
   
   
       PathBasedMetadataObject pathBasedMetadataObject2 =
           new PathBasedMetadataObject("parent", "name", "path", 
PathBasedMetadataObject.Type.PATH);
       pathBasedMetadataObject2.validateAuthorizationMetadataObject();
   
       boolean result = 
pathBasedMetadataObject1.equals(pathBasedMetadataObject2); // alreays false
       System.out.println(result);
   ```



##########
authorizations/authorization-ranger/src/main/java/org/apache/gravitino/authorization/ranger/RangerAuthorizationHDFSPlugin.java:
##########
@@ -118,27 +127,372 @@ public List<String> policyResourceDefinesRule() {
     return ImmutableList.of(RangerDefines.PolicyResource.PATH.getName());
   }
 
+  /**
+   * Find the managed policy for the ranger securable object.
+   *
+   * @param authzMetadataObject The ranger securable object to find the 
managed policy.
+   * @return The managed policy for the metadata object.
+   */
+  public RangerPolicy findManagedPolicy(AuthorizationMetadataObject 
authzMetadataObject)
+      throws AuthorizationPluginException {
+    List<RangerPolicy> policies = wildcardSearchPolies(authzMetadataObject);
+    if (!policies.isEmpty()) {
+      /**
+       * Because Ranger doesn't support the precise search, Ranger will return 
the policy meets the
+       * wildcard(*,?) conditions, If you use `/a/b` condition to search 
policy, the Ranger will
+       * match `/a/b1`, `/a/b2`, `/a/b*`, So we need to manually precisely 
filter this research
+       * results.
+       */
+      List<String> nsMetadataObj = authzMetadataObject.names();
+      PathBasedMetadataObject pathAuthzMetadataObject =
+          (PathBasedMetadataObject) authzMetadataObject;
+      Map<String, String> preciseFilters = new HashMap<>();
+      for (int i = 0; i < nsMetadataObj.size() && i < 
policyResourceDefinesRule().size(); i++) {
+        preciseFilters.put(policyResourceDefinesRule().get(i), 
pathAuthzMetadataObject.path());
+      }
+      policies =
+          policies.stream()
+              .filter(
+                  policy ->
+                      policy.getResources().entrySet().stream()
+                          .allMatch(
+                              entry ->
+                                  preciseFilters.containsKey(entry.getKey())
+                                      && entry.getValue().getValues().size() 
== 1
+                                      && entry
+                                          .getValue()
+                                          .getValues()
+                                          
.contains(preciseFilters.get(entry.getKey()))))
+              .collect(Collectors.toList());
+    }
+    // Only return the policies that are managed by Gravitino.
+    if (policies.size() > 1) {
+      throw new AuthorizationPluginException("Each metadata object can have at 
most one policy.");
+    }
+
+    if (policies.isEmpty()) {
+      return null;
+    }
+
+    RangerPolicy policy = policies.get(0);
+    // Delegating Gravitino management policies cannot contain duplicate 
privilege
+    policy.getPolicyItems().forEach(RangerHelper::checkPolicyItemAccess);
+    policy.getDenyPolicyItems().forEach(RangerHelper::checkPolicyItemAccess);
+    
policy.getRowFilterPolicyItems().forEach(RangerHelper::checkPolicyItemAccess);
+    
policy.getDataMaskPolicyItems().forEach(RangerHelper::checkPolicyItemAccess);
+
+    return policy;
+  }
+
+  @Override
+  /** Wildcard search the Ranger policies in the different Ranger service. */
+  protected List<RangerPolicy> wildcardSearchPolies(
+      AuthorizationMetadataObject authzMetadataObject) {
+    Preconditions.checkArgument(authzMetadataObject instanceof 
PathBasedMetadataObject);
+    PathBasedMetadataObject pathBasedMetadataObject = 
(PathBasedMetadataObject) authzMetadataObject;
+    List<String> resourceDefines = policyResourceDefinesRule();
+    Map<String, String> searchFilters = new HashMap<>();
+    searchFilters.put(SearchFilter.SERVICE_NAME, rangerServiceName);
+    resourceDefines.stream()
+        .forEach(
+            resourceDefine -> {
+              searchFilters.put(
+                  SearchFilter.RESOURCE_PREFIX + resourceDefine, 
pathBasedMetadataObject.path());
+            });
+    try {
+      List<RangerPolicy> policies = rangerClient.findPolicies(searchFilters);
+      return policies;
+    } catch (RangerServiceException e) {
+      throw new AuthorizationPluginException(e, "Failed to find the policies 
in the Ranger");
+    }
+  }
+
+  /**
+   * IF rename the SCHEMA, Need to rename these the relevant policies, 
`{schema}`, `{schema}.*`,
+   * `{schema}.*.*` <br>
+   * IF rename the TABLE, Need to rename these the relevant policies, 
`{schema}.*`, `{schema}.*.*`
+   * <br>
+   */
+  @Override
+  protected void doRenameMetadataObject(
+      AuthorizationMetadataObject authzMetadataObject,
+      AuthorizationMetadataObject newAuthzMetadataObject) {
+    List<Map<String, String>> loop;
+    if (newAuthzMetadataObject.type().equals(SCHEMA)) {
+      loop =
+          ImmutableList.of(
+              ImmutableMap.of(
+                  authzMetadataObject.names().get(0), 
newAuthzMetadataObject.names().get(0)),
+              ImmutableMap.of(RangerHelper.RESOURCE_ALL, 
RangerHelper.RESOURCE_ALL),
+              ImmutableMap.of(RangerHelper.RESOURCE_ALL, 
RangerHelper.RESOURCE_ALL));
+    } else if (newAuthzMetadataObject.type().equals(TABLE)) {
+      loop =
+          ImmutableList.of(
+              ImmutableMap.of(
+                  authzMetadataObject.names().get(0), 
newAuthzMetadataObject.names().get(0)),
+              ImmutableMap.of(
+                  authzMetadataObject.names().get(1), 
newAuthzMetadataObject.names().get(1)),
+              ImmutableMap.of(RangerHelper.RESOURCE_ALL, 
RangerHelper.RESOURCE_ALL));
+    } else if (newAuthzMetadataObject.type().equals(COLUMN)) {
+      loop =
+          ImmutableList.of(
+              ImmutableMap.of(
+                  authzMetadataObject.names().get(0), 
newAuthzMetadataObject.names().get(0)),
+              ImmutableMap.of(
+                  authzMetadataObject.names().get(1), 
newAuthzMetadataObject.names().get(1)),
+              ImmutableMap.of(
+                  authzMetadataObject.names().get(2), 
newAuthzMetadataObject.names().get(2)));
+    } else if (newAuthzMetadataObject.type().equals(PATH)) {
+      // do nothing when fileset is renamed
+      return;
+    } else {
+      throw new IllegalArgumentException(
+          "Unsupported metadata object type: " + authzMetadataObject.type());
+    }
+
+    List<String> oldMetadataNames = new ArrayList<>();
+    List<String> newMetadataNames = new ArrayList<>();
+    for (int index = 0; index < loop.size(); index++) {
+      
oldMetadataNames.add(loop.get(index).keySet().stream().findFirst().get());
+      
newMetadataNames.add(loop.get(index).values().stream().findFirst().get());
+
+      AuthorizationMetadataObject.Type type =
+          (index == 0
+              ? RangerHadoopSQLMetadataObject.Type.SCHEMA
+              : (index == 1
+                  ? RangerHadoopSQLMetadataObject.Type.TABLE
+                  : RangerHadoopSQLMetadataObject.Type.COLUMN));
+      AuthorizationMetadataObject oldHadoopSQLMetadataObject =
+          new RangerHadoopSQLMetadataObject(
+              AuthorizationMetadataObject.getParentFullName(oldMetadataNames),
+              AuthorizationMetadataObject.getLastName(oldMetadataNames),
+              type);
+      AuthorizationMetadataObject newHadoopSQLMetadataObject =
+          new RangerHadoopSQLMetadataObject(
+              AuthorizationMetadataObject.getParentFullName(newMetadataNames),
+              AuthorizationMetadataObject.getLastName(newMetadataNames),
+              type);
+      updatePolicyByMetadataObject(
+          MetadataObject.Type.SCHEMA, oldHadoopSQLMetadataObject, 
newHadoopSQLMetadataObject);
+    }
+  }
+
+  @Override
+  protected void updatePolicyByMetadataObject(
+      MetadataObject.Type operationType,
+      AuthorizationMetadataObject oldAuthzMetaobject,
+      AuthorizationMetadataObject newAuthzMetaobject) {
+    List<RangerPolicy> oldPolicies = wildcardSearchPolies(oldAuthzMetaobject);
+    List<RangerPolicy> existNewPolicies = 
wildcardSearchPolies(newAuthzMetaobject);
+    if (oldPolicies.isEmpty()) {
+      LOG.warn("Cannot find the Ranger policy for the metadata object({})!", 
oldAuthzMetaobject);
+    }
+    if (!existNewPolicies.isEmpty()) {
+      LOG.warn("The Ranger policy for the metadata object({}) already 
exists!", newAuthzMetaobject);
+    }
+    Map<MetadataObject.Type, Integer> operationTypeIndex =
+        ImmutableMap.of(
+            MetadataObject.Type.SCHEMA, 0,
+            MetadataObject.Type.TABLE, 1,
+            MetadataObject.Type.COLUMN, 2);
+    oldPolicies.stream()
+        .forEach(
+            policy -> {
+              try {
+                String policyName = policy.getName();
+                int index = operationTypeIndex.get(operationType);
+
+                // Update the policy name is following Gravitino's spec
+                if (policy
+                    .getName()
+                    .equals(
+                        
AuthorizationSecurableObject.DOT_JOINER.join(oldAuthzMetaobject.names()))) {
+                  List<String> policyNames =
+                      Lists.newArrayList(
+                          
AuthorizationSecurableObject.DOT_SPLITTER.splitToList(policyName));
+                  Preconditions.checkArgument(
+                      policyNames.size() >= oldAuthzMetaobject.names().size(),
+                      String.format("The policy name(%s) is invalid!", 
policyName));
+                  if 
(policyNames.get(index).equals(RangerHelper.RESOURCE_ALL)) {
+                    // Doesn't need to rename the policy `*`
+                    return;
+                  }
+                  policyNames.set(index, 
newAuthzMetaobject.names().get(index));
+                  
policy.setName(AuthorizationSecurableObject.DOT_JOINER.join(policyNames));
+                }
+                // Update the policy resource name to new name
+                policy
+                    .getResources()
+                    .put(
+                        rangerHelper.policyResourceDefines.get(index),
+                        new RangerPolicy.RangerPolicyResource(
+                            newAuthzMetaobject.names().get(index)));
+
+                boolean alreadyExist =
+                    existNewPolicies.stream()
+                        .anyMatch(
+                            existNewPolicy ->
+                                
existNewPolicy.getName().equals(policy.getName())
+                                    || 
existNewPolicy.getResources().equals(policy.getResources()));
+                if (alreadyExist) {
+                  LOG.warn(
+                      "The Ranger policy for the metadata object({}) already 
exists!",
+                      newAuthzMetaobject);
+                  return;
+                }
+
+                // Update the policy
+                rangerClient.updatePolicy(policy.getId(), policy);
+              } catch (RangerServiceException e) {
+                LOG.error("Failed to rename the policy {}!", policy);
+                throw new RuntimeException(e);
+              }
+            });
+  }
+
+  /**
+   * IF remove the SCHEMA, need to remove these the relevant policies, 
`{schema}`, `{schema}.*`,
+   * `{schema}.*.*` <br>
+   * IF remove the TABLE, need to remove these the relevant policies, 
`{schema}.*`, `{schema}.*.*`
+   * <br>
+   * IF remove the COLUMN, Only need to remove `{schema}.*.*` <br>
+   */
+  @Override
+  protected void doRemoveMetadataObject(AuthorizationMetadataObject 
authzMetadataObject) {
+    if (authzMetadataObject.type().equals(SCHEMA)) {
+      doRemoveSchemaMetadataObject(authzMetadataObject);
+    } else if (authzMetadataObject.type().equals(TABLE)) {
+      doRemoveTableMetadataObject(authzMetadataObject);
+    } else if (authzMetadataObject.type().equals(PATH)) {
+      removePolicyByMetadataObject(authzMetadataObject);
+    } else {
+      throw new IllegalArgumentException(
+          "Unsupported authorization metadata object type: " + 
authzMetadataObject.type());
+    }
+  }
+
+  /**
+   * Remove the SCHEMA, Need to remove these the relevant policies, 
`{schema}`, `{schema}.*`,
+   * `{schema}.*.*` permissions.
+   */
+  private void doRemoveSchemaMetadataObject(AuthorizationMetadataObject 
authzMetadataObject) {
+    Preconditions.checkArgument(
+        authzMetadataObject instanceof PathBasedMetadataObject,
+        "The metadata object must be a PathBasedMetadataObject");
+    Preconditions.checkArgument(
+        authzMetadataObject.type() == SCHEMA, "The metadata object type must 
be SCHEMA");
+    Preconditions.checkArgument(
+        authzMetadataObject.names().size() == 1, "The metadata object names 
must be 1");
+    if (RangerHelper.RESOURCE_ALL.equals(authzMetadataObject.name())) {
+      // Remove all schema in this catalog
+      NameIdentifier[] catalogs =
+          
GravitinoEnv.getInstance().catalogDispatcher().listCatalogs(Namespace.of(metalake));
+      Arrays.asList(catalogs).stream()
+          .forEach(
+              catalog -> {
+                List<String> catalogLocations =
+                    AuthorizationUtils.getMetadataObjectLocation(
+                        NameIdentifier.of(catalog.name()), 
Entity.EntityType.CATALOG);
+                catalogLocations.stream()
+                    .forEach(
+                        locationPath -> {
+                          AuthorizationMetadataObject catalogMetadataObject =
+                              new PathBasedMetadataObject(
+                                  metalake, catalog.name(), locationPath, 
PATH);
+                          doRemoveSchemaMetadataObject(catalogMetadataObject);
+                        });
+              });
+    } else {
+      // Remove all table in this schema
+      NameIdentifier[] tables =
+          GravitinoEnv.getInstance()
+              .tableDispatcher()
+              .listTables(Namespace.of(authzMetadataObject.name()));
+      Arrays.asList(tables).stream()
+          .forEach(
+              table -> {
+                NameIdentifier identifier =
+                    NameIdentifier.of(authzMetadataObject.name(), 
table.name());
+                List<String> tabLocations =
+                    AuthorizationUtils.getMetadataObjectLocation(
+                        identifier, Entity.EntityType.TABLE);
+                tabLocations.stream()
+                    .forEach(
+                        locationPath -> {
+                          AuthorizationMetadataObject tableMetadataObject =
+                              new PathBasedMetadataObject(
+                                  authzMetadataObject.name(), table.name(), 
locationPath, PATH);
+                          doRemoveTableMetadataObject(tableMetadataObject);
+                        });
+                // Remove schema
+                Schema schema =
+                    GravitinoEnv.getInstance()
+                        .schemaDispatcher()
+                        
.loadSchema(NameIdentifier.of(authzMetadataObject.name()));
+                List<String> schemaLocations =
+                    AuthorizationUtils.getMetadataObjectLocation(
+                        identifier, Entity.EntityType.SCHEMA);
+                schemaLocations.stream()
+                    .forEach(
+                        locationPath -> {
+                          AuthorizationMetadataObject schemaMetadataObject =
+                              new PathBasedMetadataObject(
+                                  authzMetadataObject.name(), schema.name(), 
locationPath, PATH);
+                          removePolicyByMetadataObject(schemaMetadataObject);
+                        });
+              });

Review Comment:
   I'm not entirely clear on those codes, don't we need to remove all tables 
and then the schema location?



##########
core/src/main/java/org/apache/gravitino/authorization/AuthorizationUtils.java:
##########
@@ -364,4 +378,129 @@ private static void checkCatalogType(
           catalogIdent, catalog.type(), privilege);
     }
   }
+
+  public static List<String> getMetadataObjectLocation(
+      NameIdentifier ident, Entity.EntityType type) {
+    List<String> locations = new ArrayList<>();
+    MetadataObject metadataObject;
+    try {
+      metadataObject = NameIdentifierUtil.toMetadataObject(ident, type);
+    } catch (IllegalArgumentException e) {
+      LOG.warn("Illegal argument exception for metadata object %s type %s", 
ident, type, e);
+      return locations;
+    }
+
+    String metalake =
+        (type == Entity.EntityType.METALAKE ? ident.name() : 
ident.namespace().level(0));
+    try {
+      switch (metadataObject.type()) {
+        case METALAKE:
+          {
+            NameIdentifier[] identifiers =
+                
GravitinoEnv.getInstance().catalogDispatcher().listCatalogs(Namespace.of(metalake));
+            List<String> finalLocationPath = locations;
+            Arrays.stream(identifiers)
+                .collect(Collectors.toList())
+                .forEach(
+                    identifier -> {
+                      Catalog catalogObj =
+                          
GravitinoEnv.getInstance().catalogDispatcher().loadCatalog(identifier);
+                      if (catalogObj.provider().equals("hive")) {
+                        Schema schema =
+                            GravitinoEnv.getInstance()
+                                .schemaDispatcher()
+                                .loadSchema(
+                                    NameIdentifier.of(
+                                        metalake,
+                                        catalogObj.name(),
+                                        "default" /*Hive default schema*/));
+                        if 
(schema.properties().containsKey(HiveConstants.LOCATION)) {
+                          String defaultSchemaLocation =
+                              schema.properties().get(HiveConstants.LOCATION);
+                          Preconditions.checkArgument(
+                              defaultSchemaLocation != null,
+                              String.format("Catalog %s location is not 
found", ident));
+                          String location =
+                              
HDFS_PATTERN.matcher(defaultSchemaLocation).replaceAll("");

Review Comment:
   Since this is common tools for all `PathBaseObjectMetadata`, it's not so 
property to use HDFS



##########
core/src/main/java/org/apache/gravitino/authorization/AuthorizationUtils.java:
##########
@@ -364,4 +378,129 @@ private static void checkCatalogType(
           catalogIdent, catalog.type(), privilege);
     }
   }
+
+  public static List<String> getMetadataObjectLocation(
+      NameIdentifier ident, Entity.EntityType type) {
+    List<String> locations = new ArrayList<>();
+    MetadataObject metadataObject;
+    try {
+      metadataObject = NameIdentifierUtil.toMetadataObject(ident, type);
+    } catch (IllegalArgumentException e) {
+      LOG.warn("Illegal argument exception for metadata object %s type %s", 
ident, type, e);
+      return locations;
+    }
+
+    String metalake =
+        (type == Entity.EntityType.METALAKE ? ident.name() : 
ident.namespace().level(0));
+    try {
+      switch (metadataObject.type()) {
+        case METALAKE:
+          {
+            NameIdentifier[] identifiers =
+                
GravitinoEnv.getInstance().catalogDispatcher().listCatalogs(Namespace.of(metalake));
+            List<String> finalLocationPath = locations;
+            Arrays.stream(identifiers)
+                .collect(Collectors.toList())
+                .forEach(
+                    identifier -> {
+                      Catalog catalogObj =
+                          
GravitinoEnv.getInstance().catalogDispatcher().loadCatalog(identifier);
+                      if (catalogObj.provider().equals("hive")) {
+                        Schema schema =
+                            GravitinoEnv.getInstance()
+                                .schemaDispatcher()
+                                .loadSchema(
+                                    NameIdentifier.of(
+                                        metalake,
+                                        catalogObj.name(),
+                                        "default" /*Hive default schema*/));
+                        if 
(schema.properties().containsKey(HiveConstants.LOCATION)) {

Review Comment:
   Just curious, has the current code base implemented the logic for fetching 
the schema location?



##########
authorizations/authorization-ranger/src/main/java/org/apache/gravitino/authorization/ranger/RangerAuthorizationHDFSPlugin.java:
##########
@@ -278,88 +696,91 @@ public List<AuthorizationSecurableObject> 
translateOwner(MetadataObject gravitin
   }
 
   @Override
-  public AuthorizationMetadataObject translateMetadataObject(MetadataObject 
metadataObject) {
-    Preconditions.checkArgument(
-        allowMetadataObjectTypesRule().contains(metadataObject.type()),
-        String.format(
-            "The metadata object type %s is not supported in the 
RangerAuthorizationHDFSPlugin",
-            metadataObject.type()));
-    List<String> nsMetadataObject =
-        
Lists.newArrayList(SecurableObjects.DOT_SPLITTER.splitToList(metadataObject.fullName()));
-    Preconditions.checkArgument(
-        nsMetadataObject.size() > 0, "The metadata object must have at least 
one name.");
-
-    PathBasedMetadataObject rangerPathBaseMetadataObject;
-    switch (metadataObject.type()) {
-      case METALAKE:
-      case CATALOG:
-        rangerPathBaseMetadataObject =
-            new PathBasedMetadataObject("", PathBasedMetadataObject.Type.PATH);
-        break;
-      case SCHEMA:
-        rangerPathBaseMetadataObject =
-            new PathBasedMetadataObject(
-                metadataObject.fullName(), PathBasedMetadataObject.Type.PATH);
-        break;
-      case FILESET:
-        rangerPathBaseMetadataObject =
-            new PathBasedMetadataObject(
-                getLocationPath(metadataObject), 
PathBasedMetadataObject.Type.PATH);
-        break;
-      default:
-        throw new AuthorizationPluginException(
-            "The metadata object type %s is not supported in the 
RangerAuthorizationHDFSPlugin",
-            metadataObject.type());
-    }
-    rangerPathBaseMetadataObject.validateAuthorizationMetadataObject();
-    return rangerPathBaseMetadataObject;
-  }
-
-  private NameIdentifier getObjectNameIdentifier(MetadataObject 
metadataObject) {
-    return NameIdentifier.parse(String.format("%s.%s", metalake, 
metadataObject.fullName()));
+  public List<AuthorizationMetadataObject> 
translateMetadataObject(MetadataObject metadataObject) {
+    List<AuthorizationMetadataObject> authzMetadataObjects = new ArrayList<>();
+    Entity.EntityType entityType = 
MetadataObjectUtil.toEntityType(metadataObject);
+    NameIdentifier identifier =
+        metadataObject.type().equals(MetadataObject.Type.METALAKE)
+            ? NameIdentifier.of(metadataObject.fullName())
+            : NameIdentifier.parse(String.join(".", metalake, 
metadataObject.fullName()));

Review Comment:
   I wonder if metadata object of `metalake` and `catalog` can have locations.



##########
core/src/main/java/org/apache/gravitino/hook/FilesetHookDispatcher.java:
##########
@@ -103,8 +104,11 @@ public Fileset alterFileset(NameIdentifier ident, 
FilesetChange... changes)
 
   @Override
   public boolean dropFileset(NameIdentifier ident) {
+    List<String> locations =
+        AuthorizationUtils.getMetadataObjectLocation(ident, 
Entity.EntityType.FILESET);
     boolean dropped = dispatcher.dropFileset(ident);
-    AuthorizationUtils.authorizationPluginRemovePrivileges(ident, 
Entity.EntityType.FILESET);
+    AuthorizationUtils.authorizationPluginRemovePrivileges(
+        ident, Entity.EntityType.FILESET, locations);
     return dropped;

Review Comment:
   Why do we perform the authorization logic here, can we move them to the 
plugin?



##########
authorizations/authorization-ranger/src/main/java/org/apache/gravitino/authorization/ranger/RangerAuthorizationHDFSPlugin.java:
##########
@@ -118,27 +127,372 @@ public List<String> policyResourceDefinesRule() {
     return ImmutableList.of(RangerDefines.PolicyResource.PATH.getName());
   }
 
+  /**
+   * Find the managed policy for the ranger securable object.
+   *
+   * @param authzMetadataObject The ranger securable object to find the 
managed policy.
+   * @return The managed policy for the metadata object.
+   */
+  public RangerPolicy findManagedPolicy(AuthorizationMetadataObject 
authzMetadataObject)
+      throws AuthorizationPluginException {
+    List<RangerPolicy> policies = wildcardSearchPolies(authzMetadataObject);
+    if (!policies.isEmpty()) {
+      /**
+       * Because Ranger doesn't support the precise search, Ranger will return 
the policy meets the
+       * wildcard(*,?) conditions, If you use `/a/b` condition to search 
policy, the Ranger will
+       * match `/a/b1`, `/a/b2`, `/a/b*`, So we need to manually precisely 
filter this research
+       * results.
+       */
+      List<String> nsMetadataObj = authzMetadataObject.names();
+      PathBasedMetadataObject pathAuthzMetadataObject =
+          (PathBasedMetadataObject) authzMetadataObject;
+      Map<String, String> preciseFilters = new HashMap<>();
+      for (int i = 0; i < nsMetadataObj.size() && i < 
policyResourceDefinesRule().size(); i++) {
+        preciseFilters.put(policyResourceDefinesRule().get(i), 
pathAuthzMetadataObject.path());
+      }
+      policies =
+          policies.stream()
+              .filter(
+                  policy ->
+                      policy.getResources().entrySet().stream()
+                          .allMatch(
+                              entry ->
+                                  preciseFilters.containsKey(entry.getKey())
+                                      && entry.getValue().getValues().size() 
== 1
+                                      && entry
+                                          .getValue()
+                                          .getValues()
+                                          
.contains(preciseFilters.get(entry.getKey()))))
+              .collect(Collectors.toList());
+    }
+    // Only return the policies that are managed by Gravitino.
+    if (policies.size() > 1) {
+      throw new AuthorizationPluginException("Each metadata object can have at 
most one policy.");
+    }
+
+    if (policies.isEmpty()) {
+      return null;
+    }
+
+    RangerPolicy policy = policies.get(0);
+    // Delegating Gravitino management policies cannot contain duplicate 
privilege
+    policy.getPolicyItems().forEach(RangerHelper::checkPolicyItemAccess);
+    policy.getDenyPolicyItems().forEach(RangerHelper::checkPolicyItemAccess);
+    
policy.getRowFilterPolicyItems().forEach(RangerHelper::checkPolicyItemAccess);
+    
policy.getDataMaskPolicyItems().forEach(RangerHelper::checkPolicyItemAccess);
+
+    return policy;
+  }
+
+  @Override
+  /** Wildcard search the Ranger policies in the different Ranger service. */
+  protected List<RangerPolicy> wildcardSearchPolies(
+      AuthorizationMetadataObject authzMetadataObject) {
+    Preconditions.checkArgument(authzMetadataObject instanceof 
PathBasedMetadataObject);
+    PathBasedMetadataObject pathBasedMetadataObject = 
(PathBasedMetadataObject) authzMetadataObject;
+    List<String> resourceDefines = policyResourceDefinesRule();
+    Map<String, String> searchFilters = new HashMap<>();
+    searchFilters.put(SearchFilter.SERVICE_NAME, rangerServiceName);
+    resourceDefines.stream()
+        .forEach(
+            resourceDefine -> {
+              searchFilters.put(
+                  SearchFilter.RESOURCE_PREFIX + resourceDefine, 
pathBasedMetadataObject.path());
+            });
+    try {
+      List<RangerPolicy> policies = rangerClient.findPolicies(searchFilters);
+      return policies;
+    } catch (RangerServiceException e) {
+      throw new AuthorizationPluginException(e, "Failed to find the policies 
in the Ranger");
+    }
+  }
+
+  /**
+   * IF rename the SCHEMA, Need to rename these the relevant policies, 
`{schema}`, `{schema}.*`,
+   * `{schema}.*.*` <br>
+   * IF rename the TABLE, Need to rename these the relevant policies, 
`{schema}.*`, `{schema}.*.*`
+   * <br>
+   */
+  @Override
+  protected void doRenameMetadataObject(
+      AuthorizationMetadataObject authzMetadataObject,
+      AuthorizationMetadataObject newAuthzMetadataObject) {
+    List<Map<String, String>> loop;
+    if (newAuthzMetadataObject.type().equals(SCHEMA)) {
+      loop =
+          ImmutableList.of(
+              ImmutableMap.of(
+                  authzMetadataObject.names().get(0), 
newAuthzMetadataObject.names().get(0)),
+              ImmutableMap.of(RangerHelper.RESOURCE_ALL, 
RangerHelper.RESOURCE_ALL),
+              ImmutableMap.of(RangerHelper.RESOURCE_ALL, 
RangerHelper.RESOURCE_ALL));
+    } else if (newAuthzMetadataObject.type().equals(TABLE)) {
+      loop =
+          ImmutableList.of(
+              ImmutableMap.of(
+                  authzMetadataObject.names().get(0), 
newAuthzMetadataObject.names().get(0)),
+              ImmutableMap.of(
+                  authzMetadataObject.names().get(1), 
newAuthzMetadataObject.names().get(1)),
+              ImmutableMap.of(RangerHelper.RESOURCE_ALL, 
RangerHelper.RESOURCE_ALL));
+    } else if (newAuthzMetadataObject.type().equals(COLUMN)) {
+      loop =
+          ImmutableList.of(
+              ImmutableMap.of(
+                  authzMetadataObject.names().get(0), 
newAuthzMetadataObject.names().get(0)),
+              ImmutableMap.of(
+                  authzMetadataObject.names().get(1), 
newAuthzMetadataObject.names().get(1)),
+              ImmutableMap.of(
+                  authzMetadataObject.names().get(2), 
newAuthzMetadataObject.names().get(2)));
+    } else if (newAuthzMetadataObject.type().equals(PATH)) {
+      // do nothing when fileset is renamed
+      return;
+    } else {
+      throw new IllegalArgumentException(
+          "Unsupported metadata object type: " + authzMetadataObject.type());
+    }
+
+    List<String> oldMetadataNames = new ArrayList<>();
+    List<String> newMetadataNames = new ArrayList<>();
+    for (int index = 0; index < loop.size(); index++) {
+      
oldMetadataNames.add(loop.get(index).keySet().stream().findFirst().get());
+      
newMetadataNames.add(loop.get(index).values().stream().findFirst().get());
+
+      AuthorizationMetadataObject.Type type =
+          (index == 0
+              ? RangerHadoopSQLMetadataObject.Type.SCHEMA
+              : (index == 1
+                  ? RangerHadoopSQLMetadataObject.Type.TABLE
+                  : RangerHadoopSQLMetadataObject.Type.COLUMN));
+      AuthorizationMetadataObject oldHadoopSQLMetadataObject =
+          new RangerHadoopSQLMetadataObject(
+              AuthorizationMetadataObject.getParentFullName(oldMetadataNames),
+              AuthorizationMetadataObject.getLastName(oldMetadataNames),
+              type);
+      AuthorizationMetadataObject newHadoopSQLMetadataObject =
+          new RangerHadoopSQLMetadataObject(
+              AuthorizationMetadataObject.getParentFullName(newMetadataNames),
+              AuthorizationMetadataObject.getLastName(newMetadataNames),
+              type);
+      updatePolicyByMetadataObject(
+          MetadataObject.Type.SCHEMA, oldHadoopSQLMetadataObject, 
newHadoopSQLMetadataObject);
+    }
+  }
+
+  @Override
+  protected void updatePolicyByMetadataObject(
+      MetadataObject.Type operationType,
+      AuthorizationMetadataObject oldAuthzMetaobject,
+      AuthorizationMetadataObject newAuthzMetaobject) {
+    List<RangerPolicy> oldPolicies = wildcardSearchPolies(oldAuthzMetaobject);
+    List<RangerPolicy> existNewPolicies = 
wildcardSearchPolies(newAuthzMetaobject);
+    if (oldPolicies.isEmpty()) {
+      LOG.warn("Cannot find the Ranger policy for the metadata object({})!", 
oldAuthzMetaobject);
+    }
+    if (!existNewPolicies.isEmpty()) {
+      LOG.warn("The Ranger policy for the metadata object({}) already 
exists!", newAuthzMetaobject);
+    }
+    Map<MetadataObject.Type, Integer> operationTypeIndex =
+        ImmutableMap.of(
+            MetadataObject.Type.SCHEMA, 0,
+            MetadataObject.Type.TABLE, 1,
+            MetadataObject.Type.COLUMN, 2);
+    oldPolicies.stream()
+        .forEach(
+            policy -> {
+              try {
+                String policyName = policy.getName();
+                int index = operationTypeIndex.get(operationType);
+
+                // Update the policy name is following Gravitino's spec
+                if (policy
+                    .getName()
+                    .equals(
+                        
AuthorizationSecurableObject.DOT_JOINER.join(oldAuthzMetaobject.names()))) {
+                  List<String> policyNames =
+                      Lists.newArrayList(
+                          
AuthorizationSecurableObject.DOT_SPLITTER.splitToList(policyName));
+                  Preconditions.checkArgument(
+                      policyNames.size() >= oldAuthzMetaobject.names().size(),
+                      String.format("The policy name(%s) is invalid!", 
policyName));
+                  if 
(policyNames.get(index).equals(RangerHelper.RESOURCE_ALL)) {
+                    // Doesn't need to rename the policy `*`
+                    return;
+                  }
+                  policyNames.set(index, 
newAuthzMetaobject.names().get(index));
+                  
policy.setName(AuthorizationSecurableObject.DOT_JOINER.join(policyNames));
+                }
+                // Update the policy resource name to new name
+                policy
+                    .getResources()
+                    .put(
+                        rangerHelper.policyResourceDefines.get(index),
+                        new RangerPolicy.RangerPolicyResource(
+                            newAuthzMetaobject.names().get(index)));
+
+                boolean alreadyExist =
+                    existNewPolicies.stream()
+                        .anyMatch(
+                            existNewPolicy ->
+                                
existNewPolicy.getName().equals(policy.getName())
+                                    || 
existNewPolicy.getResources().equals(policy.getResources()));
+                if (alreadyExist) {
+                  LOG.warn(
+                      "The Ranger policy for the metadata object({}) already 
exists!",
+                      newAuthzMetaobject);
+                  return;
+                }
+
+                // Update the policy
+                rangerClient.updatePolicy(policy.getId(), policy);
+              } catch (RangerServiceException e) {
+                LOG.error("Failed to rename the policy {}!", policy);
+                throw new RuntimeException(e);
+              }
+            });
+  }
+
+  /**
+   * IF remove the SCHEMA, need to remove these the relevant policies, 
`{schema}`, `{schema}.*`,
+   * `{schema}.*.*` <br>
+   * IF remove the TABLE, need to remove these the relevant policies, 
`{schema}.*`, `{schema}.*.*`
+   * <br>
+   * IF remove the COLUMN, Only need to remove `{schema}.*.*` <br>
+   */
+  @Override
+  protected void doRemoveMetadataObject(AuthorizationMetadataObject 
authzMetadataObject) {
+    if (authzMetadataObject.type().equals(SCHEMA)) {
+      doRemoveSchemaMetadataObject(authzMetadataObject);
+    } else if (authzMetadataObject.type().equals(TABLE)) {
+      doRemoveTableMetadataObject(authzMetadataObject);
+    } else if (authzMetadataObject.type().equals(PATH)) {
+      removePolicyByMetadataObject(authzMetadataObject);
+    } else {
+      throw new IllegalArgumentException(
+          "Unsupported authorization metadata object type: " + 
authzMetadataObject.type());
+    }
+  }
+
+  /**
+   * Remove the SCHEMA, Need to remove these the relevant policies, 
`{schema}`, `{schema}.*`,
+   * `{schema}.*.*` permissions.
+   */
+  private void doRemoveSchemaMetadataObject(AuthorizationMetadataObject 
authzMetadataObject) {
+    Preconditions.checkArgument(
+        authzMetadataObject instanceof PathBasedMetadataObject,
+        "The metadata object must be a PathBasedMetadataObject");
+    Preconditions.checkArgument(
+        authzMetadataObject.type() == SCHEMA, "The metadata object type must 
be SCHEMA");
+    Preconditions.checkArgument(
+        authzMetadataObject.names().size() == 1, "The metadata object names 
must be 1");
+    if (RangerHelper.RESOURCE_ALL.equals(authzMetadataObject.name())) {
+      // Remove all schema in this catalog
+      NameIdentifier[] catalogs =
+          
GravitinoEnv.getInstance().catalogDispatcher().listCatalogs(Namespace.of(metalake));
+      Arrays.asList(catalogs).stream()
+          .forEach(
+              catalog -> {
+                List<String> catalogLocations =
+                    AuthorizationUtils.getMetadataObjectLocation(
+                        NameIdentifier.of(catalog.name()), 
Entity.EntityType.CATALOG);
+                catalogLocations.stream()
+                    .forEach(
+                        locationPath -> {
+                          AuthorizationMetadataObject catalogMetadataObject =
+                              new PathBasedMetadataObject(
+                                  metalake, catalog.name(), locationPath, 
PATH);
+                          doRemoveSchemaMetadataObject(catalogMetadataObject);
+                        });
+              });

Review Comment:
   If the metadata object type is `schema`, will we remove all schemas in all 
catalogs?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to