[
https://issues.apache.org/jira/browse/SPARK-9111?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14630648#comment-14630648
]
Zhang, Liye commented on SPARK-9111:
------------------------------------
Hi [~srowen], the memory dump mentioned here is about the spark memory usage,
which is related with umbrella
[SPARK-9103|https://issues.apache.org/jira/browse/SPARK-9103], not the
HeapDump, since we want to know what is the memory status for different spark
component. It's not easy to get how much memory used for a specific spark
component directly from the HeapDump, right?
> Dumping the memory info when an executor dies abnormally
> --------------------------------------------------------
>
> Key: SPARK-9111
> URL: https://issues.apache.org/jira/browse/SPARK-9111
> Project: Spark
> Issue Type: Sub-task
> Components: Spark Core
> Reporter: Zhang, Liye
> Priority: Minor
>
> When an executor is not normally finished, we shall give out it's memory dump
> info right before the JVM shutting down. So that if the executor is killed
> because of OOM, we can easily checkout how is the memory used and which part
> cause the OOM.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]