[
https://issues.apache.org/jira/browse/SPARK-19293?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16057127#comment-16057127
]
coneyliu edited comment on SPARK-19293 at 6/21/17 8:01 AM:
-----------------------------------------------------------
Have you tried the latest code? The exceptions you give are all about
`InterruptedException` and `RuntimeException`, those seems have been fixed in
recently patch, such as SPARK-20358 and more.
was (Author: coneyliu):
Have you tried the latest code? The exceptions you give are all about
`InterruptedException` and `RuntimeException`, those seems have been fixed in
recently path, such as #SPARK-20358 and more.
> Spark 2.1.x unstable with spark.speculation=true
> ------------------------------------------------
>
> Key: SPARK-19293
> URL: https://issues.apache.org/jira/browse/SPARK-19293
> Project: Spark
> Issue Type: Bug
> Components: Spark Core
> Affects Versions: 2.1.0, 2.1.1
> Reporter: Damian Momot
> Priority: Critical
>
> After upgrading from Spark 2.0.2 to 2.1.0 we've observed that jobs are often
> failing when speculative mode is enabled.
> In 2.0.2 speculative tasks were simply skipped if they were not used for
> result (i.e. other instance finished earlier) - and it was clearly visible in
> UI that those tasks were not counted as failures.
> In 2.1.0 many tasks are being marked failed/killed when speculative tasks
> start to run (that is at the end of stage when there are spare executors to
> use) which also leads to entire stage/job failures.
> Disabling spark.speculation solves failing problem - but speculative mode is
> very useful especially when different executors run on machines with varying
> load (for example in YARN)
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]