[
https://issues.apache.org/jira/browse/KNOX-989?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16141819#comment-16141819
]
ASF subversion and git services commented on KNOX-989:
------------------------------------------------------
Commit ac532bd73c5da8997f865e048b36c76fd3c11301 in knox's branch
refs/heads/master from [~moresandeep]
[ https://git-wip-us.apache.org/repos/asf?p=knox.git;h=ac532bd ]
KNOX-989 - Report metrics at service level (/webhdfs/v1) instead of url with
args (/webhdfs/v1/?op=LISTSTATUS) (Mohammad Kamrul Islam via Sandeep More)
> Revisit JMX Metrics to fix the Out of Memory issue
> --------------------------------------------------
>
> Key: KNOX-989
> URL: https://issues.apache.org/jira/browse/KNOX-989
> Project: Apache Knox
> Issue Type: Bug
> Components: Server
> Reporter: Sandeep More
> Assignee: Mohammad Kamrul Islam
> Fix For: 0.14.0
>
> Attachments: KNOX-989.1.patch, KNOX-989.2.patch, Screen Shot
> 2017-08-16 at 1.56.16 PM.png
>
>
> Bug [KNOX-986|https://issues.apache.org/jira/browse/KNOX-986] uncovers
> problem with Metrics when large number of unique URLs are accessed via Knox.
> The problem here is that Knox creates metrics objects per unique URL, the
> metrics objects are not flushed out (for obvious reason - to maintain the
> metric state).
> We need to come up with a proper fix to mitigate this while being able to use
> the JMX Metrics.
> One way of doing this would be to have Metrics objects at service level (
> e.g. /gateway/sandbox/webhdfs/* ) the other way would be to have a reaper
> process that clears out the unused objects. Other suggestions are welcomed !
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)