[ 
https://issues.apache.org/jira/browse/SPARK-52612?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dongjoon Hyun updated SPARK-52612:
----------------------------------
        Parent: SPARK-54359
    Issue Type: Sub-task  (was: Bug)

> When executing `./build/sbt clean Test/package`, `spark-avro-*.jar` should 
> not be collected into the `assembly/target/scala-2.13/jars` directory
> ------------------------------------------------------------------------------------------------------------------------------------------------
>
>                 Key: SPARK-52612
>                 URL: https://issues.apache.org/jira/browse/SPARK-52612
>             Project: Spark
>          Issue Type: Sub-task
>          Components: Build
>    Affects Versions: 4.1.0, 4.0.0
>            Reporter: Yang Jie
>            Assignee: Yang Jie
>            Priority: Minor
>              Labels: pull-request-available
>             Fix For: 4.1.0, 4.0.1
>
>
> After the merge of SPARK-48763, when we execute the command `./build/sbt 
> clean Test/package`, we will find that the `spark-avro-*.jar` file is being 
> collected into the `assembly/target/scala-2.13/jars` directory (which did not 
> happen before). This will lead to potential errors like the following when we 
> run benchmarks using Github Actions:
>  
> For example, running the StateStoreBasicOperationsBenchmark using GitHub 
> Actions.
> {code:java}
> 25/06/28 07:03:45 ERROR SparkContext: Failed to add 
> file:///home/runner/work/spark/spark/assembly/target/scala-2.13/jars/spark-avro_2.13-4.1.0-SNAPSHOT.jar
>  to Spark environment
> java.lang.IllegalArgumentException: requirement failed: File 
> spark-avro_2.13-4.1.0-SNAPSHOT.jar was already registered with a different 
> path (old path = 
> /home/runner/work/spark/spark/connector/avro/target/scala-2.13/spark-avro_2.13-4.1.0-SNAPSHOT.jar,
>  new path = 
> /home/runner/work/spark/spark/assembly/target/scala-2.13/jars/spark-avro_2.13-4.1.0-SNAPSHOT.jar
>     at scala.Predef$.require(Predef.scala:337)
>     at 
> org.apache.spark.rpc.netty.NettyStreamManager.addJar(NettyStreamManager.scala:85)
>     at 
> org.apache.spark.SparkContext.addLocalJarFile$1(SparkContext.scala:2184)
>     at org.apache.spark.SparkContext.addJar(SparkContext.scala:2233)
>     at org.apache.spark.SparkContext.$anonfun$new$15(SparkContext.scala:538)
>     at 
> org.apache.spark.SparkContext.$anonfun$new$15$adapted(SparkContext.scala:538)
>     at scala.collection.IterableOnceOps.foreach(IterableOnce.scala:619)
>     at scala.collection.IterableOnceOps.foreach$(IterableOnce.scala:617)
>     at scala.collection.AbstractIterable.foreach(Iterable.scala:935)
>     at org.apache.spark.SparkContext.<init>(SparkContext.scala:538)
>     at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:3062)
>     at 
> org.apache.spark.sql.classic.SparkSession$Builder.$anonfun$build$2(SparkSession.scala:839)
>     at scala.Option.getOrElse(Option.scala:201)
>     at 
> org.apache.spark.sql.classic.SparkSession$Builder.build(SparkSession.scala:830)
>     at 
> org.apache.spark.sql.classic.SparkSession$Builder.getOrCreate(SparkSession.scala:859)
>     at 
> org.apache.spark.sql.classic.SparkSession$Builder.getOrCreate(SparkSession.scala:732)
>     at 
> org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:923)
>     at 
> org.apache.spark.sql.execution.benchmark.SqlBasedBenchmark.getSparkSession(SqlBasedBenchmark.scala:45)
>     at 
> org.apache.spark.sql.execution.benchmark.SqlBasedBenchmark.getSparkSession$(SqlBasedBenchmark.scala:38)
>     at 
> org.apache.spark.sql.execution.benchmark.StateStoreBasicOperationsBenchmark$.getSparkSession(StateStoreBasicOperationsBenchmark.scala:48)
>     at 
> org.apache.spark.sql.execution.benchmark.SqlBasedBenchmark.$init$(SqlBasedBenchmark.scala:35)
>     at 
> org.apache.spark.sql.execution.benchmark.StateStoreBasicOperationsBenchmark$.<clinit>(StateStoreBasicOperationsBenchmark.scala:48)
>     at 
> org.apache.spark.sql.execution.benchmark.StateStoreBasicOperationsBenchmark.main(StateStoreBasicOperationsBenchmark.scala)
>     at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
>     at 
> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:75)
>     at 
> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:52)
>     at java.base/java.lang.reflect.Method.invoke(Method.java:580)
>     at 
> org.apache.spark.benchmark.Benchmarks$.$anonfun$main$7(Benchmarks.scala:128)
>     at scala.collection.ArrayOps$.foreach$extension(ArrayOps.scala:1324)
>     at org.apache.spark.benchmark.Benchmarks$.main(Benchmarks.scala:91)
>     at org.apache.spark.benchmark.Benchmarks.main(Benchmarks.scala)
>     at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
>     at 
> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:75)
>     at 
> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:52)
>     at java.base/java.lang.reflect.Method.invoke(Method.java:580)
>     at 
> org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
>     at 
> org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:1027)
>     at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:204)
>     at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:227)
>     at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:96)
>     at 
> org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1132)
>     at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1141)
>     at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> 25/06/28 07:03:45 WARN SparkContext: The JAR 
> file:/home/runner/work/spark/spark/core/target/scala-2.13/spark-core_2.13-4.1.0-SNAPSHOT-tests.jar
>  at spark://localhost:39213/jars/spark-core_2.13-4.1.0-SNAPSHOT-tests.jar has 
> been added already. Overwriting of added jar is not supported in the current 
> version. {code}
>  
> This won't affect the final execution result, but it seems we should fix it.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to