HeartSaVioR commented on issue #25670: [SPARK-28869][CORE] Roll over event log 
files
URL: https://github.com/apache/spark/pull/25670#issuecomment-527653940
 
 
   1. Given we've added layer in this patch, would we still want to convert 
existing tests (EventLoggingListenerSuite/FsHistoryProviderSuite/etc.) to test 
with both cases (single/rolling)?
   
   2. I'm seeing some weird exception while testing manually:
   
   ```
   19/09/04 06:00:11 ERROR AsyncEventQueue: Listener EventLoggingListener threw 
an exception
   java.util.ConcurrentModificationException
        at java.util.Hashtable$Enumerator.next(Hashtable.java:1387)
        at 
scala.collection.convert.Wrappers$JPropertiesWrapper$$anon$6.next(Wrappers.scala:424)
        at 
scala.collection.convert.Wrappers$JPropertiesWrapper$$anon$6.next(Wrappers.scala:420)
        at scala.collection.Iterator.foreach(Iterator.scala:941)
        at scala.collection.Iterator.foreach$(Iterator.scala:941)
        at scala.collection.AbstractIterator.foreach(Iterator.scala:1429)
        at scala.collection.IterableLike.foreach(IterableLike.scala:74)
        at scala.collection.IterableLike.foreach$(IterableLike.scala:73)
        at scala.collection.AbstractIterable.foreach(Iterable.scala:56)
        at scala.collection.TraversableLike.map(TraversableLike.scala:237)
        at scala.collection.TraversableLike.map$(TraversableLike.scala:230)
        at scala.collection.AbstractTraversable.map(Traversable.scala:108)
        at org.apache.spark.util.JsonProtocol$.mapToJson(JsonProtocol.scala:514)
        at 
org.apache.spark.util.JsonProtocol$.$anonfun$propertiesToJson$1(JsonProtocol.scala:520)
        at scala.Option.map(Option.scala:163)
        at 
org.apache.spark.util.JsonProtocol$.propertiesToJson(JsonProtocol.scala:519)
        at 
org.apache.spark.util.JsonProtocol$.jobStartToJson(JsonProtocol.scala:155)
        at 
org.apache.spark.util.JsonProtocol$.sparkEventToJson(JsonProtocol.scala:79)
        at 
org.apache.spark.scheduler.EventLoggingListener.logEvent(EventLoggingListener.scala:96)
        at 
org.apache.spark.scheduler.EventLoggingListener.onJobStart(EventLoggingListener.scala:158)
        at 
org.apache.spark.scheduler.SparkListenerBus.doPostEvent(SparkListenerBus.scala:37)
        at 
org.apache.spark.scheduler.SparkListenerBus.doPostEvent$(SparkListenerBus.scala:28)
        at 
org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
        at 
org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
        at org.apache.spark.util.ListenerBus.postToAll(ListenerBus.scala:99)
        at org.apache.spark.util.ListenerBus.postToAll$(ListenerBus.scala:84)
        at 
org.apache.spark.scheduler.AsyncEventQueue.super$postToAll(AsyncEventQueue.scala:102)
        at 
org.apache.spark.scheduler.AsyncEventQueue.$anonfun$dispatch$1(AsyncEventQueue.scala:102)
        at 
scala.runtime.java8.JFunction0$mcJ$sp.apply(JFunction0$mcJ$sp.java:23)
        at scala.util.DynamicVariable.withValue(DynamicVariable.scala:62)
        at 
org.apache.spark.scheduler.AsyncEventQueue.org$apache$spark$scheduler$AsyncEventQueue$$dispatch(AsyncEventQueue.scala:97)
        at 
org.apache.spark.scheduler.AsyncEventQueue$$anon$2.$anonfun$run$1(AsyncEventQueue.scala:93)
        at org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1319)
        at 
org.apache.spark.scheduler.AsyncEventQueue$$anon$2.run(AsyncEventQueue.scala:93)
   ```
   
   Fixing it would be simple - clone the properties and include into 
`SparkListenerJobStart`, but just want to know it didn't occur before this 
patch.

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to