LuciferYang commented on PR #42668: URL: https://github.com/apache/spark/pull/42668#issuecomment-1692675393
The daily tests for Java 17 have failed for two consecutive days, including the HiveExternalCatalogVersionsSuite. The test was ABORTED for the same reasons. Let's run it again to see if the failure is consistent: - https://github.com/apache/spark/actions/runs/5956925790/job/16158714165 - https://github.com/apache/spark/actions/runs/5969348559/job/16195073478 ``` 2023-08-23T23:00:49.6547573Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> : java.lang.RuntimeException: problem during retrieve of org.apache.spark#spark-submit-parent-4c061f04-b951-4d06-8909-cde5452988d9: java.lang.RuntimeException: Multiple artifacts of the module log4j#log4j;1.2.17 are retrieved to the same file! Update the retrieve pattern to fix this error.[0m[0m 2023-08-23T23:00:49.6548745Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.ivy.core.retrieve.RetrieveEngine.retrieve(RetrieveEngine.java:238)[0m[0m 2023-08-23T23:00:49.6549572Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.ivy.core.retrieve.RetrieveEngine.retrieve(RetrieveEngine.java:89)[0m[0m 2023-08-23T23:00:49.6550334Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.ivy.Ivy.retrieve(Ivy.java:551)[0m[0m 2023-08-23T23:00:49.6551079Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.deploy.SparkSubmitUtils$.resolveMavenCoordinates(SparkSubmit.scala:1464)[0m[0m 2023-08-23T23:00:49.6552024Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.hive.client.IsolatedClientLoader$.$anonfun$downloadVersion$2(IsolatedClientLoader.scala:138)[0m[0m 2023-08-23T23:00:49.6552884Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.catalyst.util.package$.quietly(package.scala:42)[0m[0m 2023-08-23T23:00:49.6553755Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.hive.client.IsolatedClientLoader$.downloadVersion(IsolatedClientLoader.scala:138)[0m[0m 2023-08-23T23:00:49.6554705Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.hive.client.IsolatedClientLoader$.liftedTree1$1(IsolatedClientLoader.scala:65)[0m[0m 2023-08-23T23:00:49.6555637Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.hive.client.IsolatedClientLoader$.forVersion(IsolatedClientLoader.scala:64)[0m[0m 2023-08-23T23:00:49.6556554Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:443)[0m[0m 2023-08-23T23:00:49.6557340Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:356)[0m[0m 2023-08-23T23:00:49.6558187Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.hive.HiveExternalCatalog.client$lzycompute(HiveExternalCatalog.scala:71)[0m[0m 2023-08-23T23:00:49.6559061Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.hive.HiveExternalCatalog.client(HiveExternalCatalog.scala:70)[0m[0m 2023-08-23T23:00:49.6559962Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:224)[0m[0m 2023-08-23T23:00:49.6560766Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23)[0m[0m 2023-08-23T23:00:49.6561584Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:102)[0m[0m 2023-08-23T23:00:49.6562510Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:224)[0m[0m 2023-08-23T23:00:49.6563435Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:150)[0m[0m 2023-08-23T23:00:49.6564323Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:140)[0m[0m 2023-08-23T23:00:49.6565340Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.hive.HiveSessionStateBuilder.externalCatalog(HiveSessionStateBuilder.scala:45)[0m[0m 2023-08-23T23:00:49.6566321Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$1(HiveSessionStateBuilder.scala:60)[0m[0m 2023-08-23T23:00:49.6567363Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.catalyst.catalog.SessionCatalog.externalCatalog$lzycompute(SessionCatalog.scala:118)[0m[0m 2023-08-23T23:00:49.6568372Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.catalyst.catalog.SessionCatalog.externalCatalog(SessionCatalog.scala:118)[0m[0m 2023-08-23T23:00:49.6569393Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.catalyst.catalog.SessionCatalog.tableExists(SessionCatalog.scala:490)[0m[0m 2023-08-23T23:00:49.6570685Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.execution.command.CreateDataSourceTableAsSelectCommand.run(createDataSourceTables.scala:155)[0m[0m 2023-08-23T23:00:49.6571842Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult$lzycompute(commands.scala:113)[0m[0m 2023-08-23T23:00:49.6572932Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult(commands.scala:111)[0m[0m 2023-08-23T23:00:49.6573996Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.execution.command.DataWritingCommandExec.executeCollect(commands.scala:125)[0m[0m 2023-08-23T23:00:49.6575045Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.$anonfun$applyOrElse$1(QueryExecution.scala:97)[0m[0m 2023-08-23T23:00:49.6576066Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103)[0m[0m 2023-08-23T23:00:49.6576937Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163)[0m[0m 2023-08-23T23:00:49.6577807Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90)[0m[0m 2023-08-23T23:00:49.6578620Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775)[0m[0m 2023-08-23T23:00:49.6579432Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)[0m[0m 2023-08-23T23:00:49.6580357Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:97)[0m[0m 2023-08-23T23:00:49.6581331Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:93)[0m[0m 2023-08-23T23:00:49.6582239Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDownWithPruning$1(TreeNode.scala:481)[0m[0m 2023-08-23T23:00:49.6583101Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:82)[0m[0m 2023-08-23T23:00:49.6584088Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.catalyst.trees.TreeNode.transformDownWithPruning(TreeNode.scala:481)[0m[0m 2023-08-23T23:00:49.6585236Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.org$apache$spark$sql$catalyst$plans$logical$AnalysisHelper$$super$transformDownWithPruning(LogicalPlan.scala:30)[0m[0m 2023-08-23T23:00:49.6586519Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning(AnalysisHelper.scala:267)[0m[0m 2023-08-23T23:00:49.6587686Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning$(AnalysisHelper.scala:263)[0m[0m 2023-08-23T23:00:49.6588898Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:30)[0m[0m 2023-08-23T23:00:49.6590014Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:30)[0m[0m 2023-08-23T23:00:49.6590993Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:457)[0m[0m 2023-08-23T23:00:49.6591930Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.execution.QueryExecution.eagerlyExecuteCommands(QueryExecution.scala:93)[0m[0m 2023-08-23T23:00:49.6592914Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.execution.QueryExecution.commandExecuted$lzycompute(QueryExecution.scala:80)[0m[0m 2023-08-23T23:00:49.6593856Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.execution.QueryExecution.commandExecuted(QueryExecution.scala:78)[0m[0m 2023-08-23T23:00:49.6594687Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.Dataset.<init>(Dataset.scala:219)[0m[0m 2023-08-23T23:00:49.6595379Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:99)[0m[0m 2023-08-23T23:00:49.6596103Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775)[0m[0m 2023-08-23T23:00:49.6596807Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:96)[0m[0m 2023-08-23T23:00:49.6597520Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:618)[0m[0m 2023-08-23T23:00:49.6598276Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775)[0m[0m 2023-08-23T23:00:49.6599022Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:613)[0m[0m 2023-08-23T23:00:49.6599819Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)[0m[0m 2023-08-23T23:00:49.6600723Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)[0m[0m 2023-08-23T23:00:49.6601707Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)[0m[0m 2023-08-23T23:00:49.6602513Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at java.base/java.lang.reflect.Method.invoke(Method.java:568)[0m[0m 2023-08-23T23:00:49.6603272Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)[0m[0m 2023-08-23T23:00:49.6604007Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)[0m[0m 2023-08-23T23:00:49.6604724Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at py4j.Gateway.invoke(Gateway.java:282)[0m[0m 2023-08-23T23:00:49.6605416Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)[0m[0m 2023-08-23T23:00:49.6606209Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at py4j.commands.CallCommand.execute(CallCommand.java:79)[0m[0m 2023-08-23T23:00:49.6606969Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)[0m[0m 2023-08-23T23:00:49.6607743Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at py4j.ClientServerConnection.run(ClientServerConnection.java:106)[0m[0m 2023-08-23T23:00:49.6608415Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at java.base/java.lang.Thread.run(Thread.java:833)[0m[0m 2023-08-23T23:00:49.6609288Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> Caused by: java.lang.RuntimeException: Multiple artifacts of the module log4j#log4j;1.2.17 are retrieved to the same file! Update the retrieve pattern to fix this error.[0m[0m 2023-08-23T23:00:49.6610288Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.ivy.core.retrieve.RetrieveEngine.determineArtifactsToCopy(RetrieveEngine.java:426)[0m[0m 2023-08-23T23:00:49.6611332Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> at org.apache.ivy.core.retrieve.RetrieveEngine.retrieve(RetrieveEngine.java:122)[0m[0m 2023-08-23T23:00:49.6612046Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> ... 66 more[0m[0m 2023-08-23T23:00:49.6612498Z [0m[[0m[0minfo[0m] [0m[0m[31m 2023-08-23 16:00:48.209 - stdout> [0m[0m ``` -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected] --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
