HyukjinKwon commented on a change in pull request #31849:
URL: https://github.com/apache/spark/pull/31849#discussion_r596549816



##########
File path: core/src/main/scala/org/apache/spark/deploy/SparkSubmit.scala
##########
@@ -1153,6 +1153,8 @@ private[spark] object SparkSubmitUtils extends Logging {
     // We need a chain resolver if we want to check multiple repositories
     val cr = new ChainResolver
     cr.setName("spark-list")
+    cr.setChangingMatcher(PatternMatcher.REGEXP)
+    cr.setChangingPattern(".*-SNAPSHOT")

Review comment:
       While I agree with the intention (I faced the same problem and had to 
manually remove cache too), I concern that it will actually make users more 
confused in a way because it's Maven or Ivy's standard behaviour, and now we're 
changing how they work by default in Spark, which probably users wouldn't know.
   
   I know it's very unlikely but some users might want to use one cached 
snapshot (presumably as they know the behaviours of Maven or Ivy resolvers 
work). Let's say, one CI regularly publishes snapshot, and users want to test 
one specific version created at the specific time. After this PR, they are 
forced to use the newest snapshot always.
   




----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to