[ 
https://issues.apache.org/jira/browse/KNOX-989?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16115331#comment-16115331
 ] 

Mohammad Kamrul Islam commented on KNOX-989:
--------------------------------------------

I would prefer to do it in service-level (such as /gateway/sandbox/webhdfs).
I can work on a patch if there is no objection.

> Revisit JMX Metrics to fix the Out of Memory issue
> --------------------------------------------------
>
>                 Key: KNOX-989
>                 URL: https://issues.apache.org/jira/browse/KNOX-989
>             Project: Apache Knox
>          Issue Type: Bug
>          Components: Server
>            Reporter: Sandeep More
>            Assignee: Mohammad Kamrul Islam
>             Fix For: 0.14.0
>
>
> Bug  [KNOX-986|https://issues.apache.org/jira/browse/KNOX-986] uncovers 
> problem with Metrics when large number of unique URLs are accessed via Knox. 
> The problem here is that Knox creates metrics objects per unique URL, the 
> metrics objects are not flushed out (for obvious reason - to maintain the 
> metric state). 
> We need to come up with a proper fix to mitigate this while being able to use 
> the JMX Metrics. 
> One way of doing this would be to have Metrics objects at service level ( 
> e.g. /gateway/sandbox/webhdfs/* ) the other way would be to have a reaper 
> process that clears out the unused objects. Other suggestions are welcomed !



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Reply via email to