I looked into this. But I found it is possible like this
https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/status/AppStatusListener.scala#L229

Line no 230. This is for executors.

Just wanna cross verify is that right?



On Mon, 21 Oct 2019, 17:24 Alonso Isidoro Roman, <alons...@gmail.com> wrote:

> Take a look in this thread
> <https://stackoverflow.com/questions/48768188/spark-execution-memory-monitoring#_=_>
>
> El lun., 21 oct. 2019 a las 13:45, Sriram Ganesh (<srigns...@gmail.com>)
> escribió:
>
>> Hi,
>>
>> I wanna monitor how much memory executor and task used for a given job.
>> Is there any direct method available for it which can be used to track this
>> metric?
>>
>> --
>> *Sriram G*
>> *Tech*
>>
>>
>
> --
> Alonso Isidoro Roman
> [image: https://]about.me/alonso.isidoro.roman
>
> <https://about.me/alonso.isidoro.roman?promo=email_sig&utm_source=email_sig&utm_medium=email_sig&utm_campaign=external_links>
>

Reply via email to