[
https://issues.apache.org/jira/browse/SPARK-1235?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Matei Zaharia resolved SPARK-1235.
----------------------------------
Resolution: Fixed
Fix Version/s: 1.0.0
Resolved in https://github.com/apache/spark/pull/186.
> DAGScheduler ignores exceptions thrown in handleTaskCompletion
> --------------------------------------------------------------
>
> Key: SPARK-1235
> URL: https://issues.apache.org/jira/browse/SPARK-1235
> Project: Spark
> Issue Type: Bug
> Affects Versions: 0.9.0, 0.9.1
> Reporter: Kay Ousterhout
> Assignee: Nan Zhu
> Priority: Blocker
> Fix For: 1.0.0
>
>
> If an exception gets thrown in the handleTaskCompletion method, the method
> exits, but the exception is caught somewhere (not clear where) and the
> DAGScheduler keeps running. Jobs hang as a result -- because not all of the
> task completion code gets run.
> This was first reported by Brad Miller on the mailing list:
> http://apache-spark-user-list.1001560.n3.nabble.com/Fwd-pyspark-crash-on-mesos-td2256.html
> and this behavior seems to have changed since 0.8 (when, based on Brad's
> description, it sounds like an exception in handleTaskCompletion would cause
> the DAGScheduler to crash), suggesting that this may be related to the Scala
> 2.10.3.
> To reproduce this problem, add "throw new Exception("foo")" anywhere in
> handleTaskCompletion and run any job locally. The job will hang and you can
> see the exception get printed in the logs.
--
This message was sent by Atlassian JIRA
(v6.2#6252)