Github user squito commented on the issue:
https://github.com/apache/spark/pull/17208
hmm I think you're right @zsxwing that we should be updating `isZombie`
before `sched.dagScheduler.taskEnded` and `sched.dagScheduler.taskSetFailed` is
called, just to keep state consistent. I don't think you'll actually hit the
bug described here, as (a) if it was from a fetch failure, `isZombie` is
already set first or if (b) its just a regular task failure, and it leads to
the stage getting aborted, then there aren't any more retries of the stage
anyway.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]