[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3971?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14570953#comment-14570953
 ] 

David Morel commented on MAPREDUCE-3971:
----------------------------------------

Wondering as well, or wether I should open a separate ticket. What I'm noticing 
when accessing  /ws/v1/history/mapreduce/jobs/{jobid} for jobs having over 100 
Ktasks is a sharp rise in MemHeapUsedM and a server that becomes unresponsive, 
even though the heap is set pretty high. I suspect this is due to parsing the 
whole jhist file, which I've seen instances of over 900MB. As a result, to 
extract stats that I need, I had to write my own external script to parse the 
jhist files retrieved from HDFS. Moreover, i noticed some bad interactions when 
using oozie (tasks seen as KILLED even though the log says they're SUCCEEDED), 
making the history server a SPOF in my architecture...

> Job History web services need to have limits on the number of items they can 
> return.
> ------------------------------------------------------------------------------------
>
>                 Key: MAPREDUCE-3971
>                 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3971
>             Project: Hadoop Map/Reduce
>          Issue Type: Sub-task
>          Components: mrv2
>    Affects Versions: 0.23.2
>            Reporter: Robert Joseph Evans
>
> The Job History web services canput a very large load on the job history 
> server.  We should put in a limit on the number of entries that can be 
> returned by the web service, and also add in the ability to modify the 
> starting location in the list, so that all entries can still be downlaoded.  
> Just not all at once.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to