clp007 commented on issue #11446:
URL: https://github.com/apache/hudi/issues/11446#issuecomment-2522254520

   I had the same problem when I deleted the data of some partitions ,The data 
uses secondary partition, event_date=2024-12-03/event_name=active ,about 200 
partitions every day
   Environment Description
   
   Hudi version : 0.15.0
   
   Spark version : 3.5.1
   
   Hive version : NA
   
   Hadoop version : 3.3.6
   
   Storage (HDFS/S3/GCS..) : GCS
   
   24/12/06 06:22:39 WARN HttpParser: URI is too large >8192                    
   
   24/12/06 06:22:39 ERROR PriorityBasedFileSystemView: Got error running 
preferred function. Trying secondary
   org.apache.hudi.exception.HoodieRemoteException: URI Too Long
        at 
org.apache.hudi.common.table.view.RemoteHoodieTableFileSystemView.loadPartitions(RemoteHoodieTableFileSystemView.java:447)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.common.table.view.RemoteHoodieTableFileSystemView.loadPartitions(RemoteHoodieTableFileSystemView.java:465)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.common.table.view.PriorityBasedFileSystemView.lambda$loadPartitions$6e5c444d$1(PriorityBasedFileSystemView.java:187)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.common.table.view.PriorityBasedFileSystemView.execute(PriorityBasedFileSystemView.java:69)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.common.table.view.PriorityBasedFileSystemView.loadPartitions(PriorityBasedFileSystemView.java:185)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.table.action.clean.CleanPlanActionExecutor.requestClean(CleanPlanActionExecutor.java:133)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.table.action.clean.CleanPlanActionExecutor.requestClean(CleanPlanActionExecutor.java:174)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.table.action.clean.CleanPlanActionExecutor.execute(CleanPlanActionExecutor.java:200)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.table.HoodieSparkCopyOnWriteTable.scheduleCleaning(HoodieSparkCopyOnWriteTable.java:212)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.client.BaseHoodieTableServiceClient.scheduleTableServiceInternal(BaseHoodieTableServiceClient.java:647)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.client.BaseHoodieTableServiceClient.clean(BaseHoodieTableServiceClient.java:746)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.client.BaseHoodieWriteClient.clean(BaseHoodieWriteClient.java:843)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.client.BaseHoodieWriteClient.clean(BaseHoodieWriteClient.java:816)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.client.BaseHoodieWriteClient.clean(BaseHoodieWriteClient.java:847)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.client.BaseHoodieWriteClient.autoCleanOnCommit(BaseHoodieWriteClient.java:581)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.client.BaseHoodieWriteClient.mayBeCleanAndArchive(BaseHoodieWriteClient.java:560)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.client.BaseHoodieWriteClient.commitStats(BaseHoodieWriteClient.java:251)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.client.SparkRDDWriteClient.commit(SparkRDDWriteClient.java:108) 
~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.commitAndPerformPostOperations(HoodieSparkSqlWriter.scala:1082)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.writeInternal(HoodieSparkSqlWriter.scala:508)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.write(HoodieSparkSqlWriter.scala:187)
 ~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:125) 
~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.hudi.DefaultSource.createRelation(DefaultSource.scala:168) 
~[hudi-spark3.5-bundle_2.12-0.15.0.jar:0.15.0]
        at 
org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:48)
 ~[spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:75)
 ~[spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:73)
 ~[spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:84)
 ~[spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.$anonfun$applyOrElse$1(QueryExecution.scala:107)
 ~[spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125)
 ~[spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201)
 ~[spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108)
 ~[spark-sql_2.12-3.5.1.jar:3.5.1]
        at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) 
~[spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66)
 ~[spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:107)
 ~[spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:98)
 ~[spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDownWithPruning$1(TreeNode.scala:473)
 ~[spark-catalyst_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(origin.scala:76) 
[spark-sql-api_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDownWithPruning(TreeNode.scala:473)
 [spark-catalyst_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.org$apache$spark$sql$catalyst$plans$logical$AnalysisHelper$$super$transformDownWithPruning(LogicalPlan.scala:32)
 [spark-catalyst_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning(AnalysisHelper.scala:267)
 [spark-catalyst_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning$(AnalysisHelper.scala:263)
 [spark-catalyst_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:32)
 [spark-catalyst_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:32)
 [spark-catalyst_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:449) 
[spark-catalyst_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.execution.QueryExecution.eagerlyExecuteCommands(QueryExecution.scala:98)
 [spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted$lzycompute(QueryExecution.scala:85)
 [spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted(QueryExecution.scala:83)
 [spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.execution.QueryExecution.assertCommandExecuted(QueryExecution.scala:142)
 [spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:859) 
[spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:388) 
[spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
org.apache.spark.sql.DataFrameWriter.saveInternal(DataFrameWriter.scala:361) 
[spark-sql_2.12-3.5.1.jar:3.5.1]
        at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:240) 
[spark-sql_2.12-3.5.1.jar:3.5.1]
        at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
~[?:?]
        at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
 ~[?:?]
        at 
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 ~[?:?]
        at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[?:?]
        at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244) 
[py4j-0.10.9.7.jar:?]
        at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:374) 
[py4j-0.10.9.7.jar:?]
        at py4j.Gateway.invoke(Gateway.java:282) [py4j-0.10.9.7.jar:?]
        at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132) 
[py4j-0.10.9.7.jar:?]
        at py4j.commands.CallCommand.execute(CallCommand.java:79) 
[py4j-0.10.9.7.jar:?]
        at 
py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182) 
[py4j-0.10.9.7.jar:?]
        at py4j.ClientServerConnection.run(ClientServerConnection.java:106) 
[py4j-0.10.9.7.jar:?]
        at java.base/java.lang.Thread.run(Thread.java:829) [?:?] 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to