[ 
https://issues.apache.org/jira/browse/SPARK-4723?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14557630#comment-14557630
 ] 

Josh Rosen commented on SPARK-4723:
-----------------------------------

I think the original PR here was closed because reviewers felt that "for some 
reason" was too vague.  I think that I've identified some issues where we'll 
retry a stage an infinite number of times if there's a bug in the shuffle write 
path on the sending side which always leads to fetch failures.  In a nutshell, 
it's possible for bugs in Spark to trigger infinite retry behavior; you can try 
manually introducing such bugs in the block serving code in order to reproduce 
the infinite retry symptom.  Therefore, this might still be worth fixing; it 
_would_ mask any such bugs, but would contribute to higher overall resiliency 
in the system.

> To abort the stages which have attempted some times
> ---------------------------------------------------
>
>                 Key: SPARK-4723
>                 URL: https://issues.apache.org/jira/browse/SPARK-4723
>             Project: Spark
>          Issue Type: Improvement
>          Components: Scheduler
>            Reporter: YanTang Zhai
>            Priority: Minor
>
> For some reason, some stages may attempt many times. A threshold could be 
> added and the stages which have attempted more than the threshold could be 
> aborted.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to