Hi Warren,
it's often an exception stemming from an OOM at the executor level.
If you are caching data make sure you spill to disk, if needed.
You could also try to increase off-heap memory to alleviate the issue.
Of course also handing more memory to the executor helps.
Best regards,
Hi All,
I have seen this exception many times in my production environment for long
running batch job. Is there some stigmatization of all root causes of this
exception? Below is my analysis:
1. This happens when executor try to fetch MapStatus of some shuffle.
2. Each executor maintains a local
Hi,
I have set up a spark job and it keeps failing even though I tried a lot of
different configurations regarding memory parameters (as suggested in other
threads I read).
My configuration:
Cluster of 4 machines: 4vCPU, 16Go RAM.
YARN version: 2.7.1
Spark version: 1.5.2
I tried a lot of