[
https://issues.apache.org/jira/browse/SPARK-2228?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14045564#comment-14045564
]
Patrick Wendell commented on SPARK-2228:
----------------------------------------
I ran your reproduction locally. What I found was that it just generates events
more quickly than the listener can process, so that was triggering all of the
subsequent errors:
{code}
$ cat job-log.txt |grep ERROR | head -n 10
14/06/26 22:41:02 ERROR scheduler.LiveListenerBus: Dropping SparkListenerEvent
because no remaining room in event queue. This likely means one of the
SparkListeners is too slow and cannot keep up withthe rate at which tasks are
being started by the scheduler.
14/06/26 22:42:01 ERROR scheduler.LiveListenerBus: Listener JobProgressListener
threw an exception
14/06/26 22:42:01 ERROR scheduler.LiveListenerBus: Listener JobProgressListener
threw an exception
14/06/26 22:42:01 ERROR scheduler.LiveListenerBus: Listener JobProgressListener
threw an exception
14/06/26 22:42:01 ERROR scheduler.LiveListenerBus: Listener JobProgressListener
threw an exception
14/06/26 22:42:01 ERROR scheduler.LiveListenerBus: Listener JobProgressListener
threw an exception
14/06/26 22:42:01 ERROR scheduler.LiveListenerBus: Listener JobProgressListener
threw an exception
{code}
If someone submits a job that creates thousands of stages in a few seconds this
can happen. But I haven't seen it happen in a real production job that does
actual nontrivial work inside of the stage.
We could consider an alternative design that applies back pressure instead of
dropping events.
> onStageSubmitted does not properly called so NoSuchElement will be thrown in
> onStageCompleted
> ---------------------------------------------------------------------------------------------
>
> Key: SPARK-2228
> URL: https://issues.apache.org/jira/browse/SPARK-2228
> Project: Spark
> Issue Type: Bug
> Components: Spark Core
> Affects Versions: 1.0.0
> Reporter: Baoxu Shi
>
> We are using `SaveAsObjectFile` and `objectFile` to cut off lineage during
> iterative computing, but after several hundreds of iterations, there will be
> `NoSuchElementsError`. We check the code and locate the problem at
> `org.apache.spark.ui.jobs.JobProgressListener`. When `onStageCompleted` is
> called, such `stageId` can not be found in `stageIdToPool`, but it does exist
> in other HashMaps. So we think `onStageSubmitted` is not properly called.
> `Spark` did add a stage but failed to send the message to listeners. When
> sending `finish` message to listeners, the error occurs.
> This problem will cause a huge number of `active stages` showing in
> `SparkUI`, which is really annoying. But it may not affect the final result,
> according to the result of my testing code.
> I'm willing to help solve this problem, any idea about which part should I
> change? I assume `org.apache.spark.scheduler.SparkListenerBus` have something
> to do with it but it looks fine to me.
> FYI, here is the test code that could reproduce the problem. I do not know
> who to put code here with highlight, so I put the code on gist to make the
> issue looks clean.
> https://gist.github.com/bxshi/b5c0fe0ae089c75a39bd
--
This message was sent by Atlassian JIRA
(v6.2#6252)