Github user kayousterhout commented on the issue:
https://github.com/apache/spark/pull/15249
Re: executor blacklisting, one more reason I've heard for this (I think
from Imran) is that tasks can fail on an executor because of memory pressure --
in which case the task may succeed on other executors that have fewer RDDs
in-memory. I'm inclined to keep it since it seems useful, and if someone
doesn't care about it, they can always set the max executor failures = max host
failures.
@mridulm re: (a), it sounds like for this case, task-set-level
blacklisting, as Imran suggested, is actually preferable to task-level
blacklisting (since the issues you mentioned all seem like issues that would
affect all tasks). Is that correct? Trying to understand the issue before
figuring out how to fix it.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]