[
https://issues.apache.org/jira/browse/SPARK-37580?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
wuyi resolved SPARK-37580.
--------------------------
Assignee: wangshengjie
Resolution: Fixed
Issue resolved by https://github.com/apache/spark/pull/34834
> Optimize current TaskSetManager abort logic when task failed count reach the
> threshold
> --------------------------------------------------------------------------------------
>
> Key: SPARK-37580
> URL: https://issues.apache.org/jira/browse/SPARK-37580
> Project: Spark
> Issue Type: Improvement
> Components: Spark Core
> Affects Versions: 3.2.0
> Reporter: wangshengjie
> Assignee: wangshengjie
> Priority: Major
>
> In production environment, we found some logic leak about TaskSetManager
> abort. For example:
> If one task has failed 3 times(max failed threshold is 4 in default), and
> there is a retry task and speculative task both in running state, then one of
> these 2 task attempts succeed and to cancel another. But executor which task
> need to be cancelled lost(oom in our situcation), this task marked as failed,
> and TaskSetManager handle this failed task attempt, it has failed 4 times so
> abort this stage and cause job failed.
> I created the patch for this bug and will soon be sent the pull request.
>
--
This message was sent by Atlassian Jira
(v8.20.1#820001)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]