[ https://issues.apache.org/jira/browse/HDFS-94?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12776711#action_12776711 ]
Tsz Wo (Nicholas), SZE commented on HDFS-94: -------------------------------------------- > I wonder why it shows a total of 17.78GB instead of 20GB Would it be the case that you have hit some limit? The following is quoted from [java man page|http://java.sun.com/javase/6/docs/technotes/tools/solaris/java.html]: {quote} On Solaris 7 and Solaris 8 SPARC platforms, the upper limit for this value is approximately 4000m minus overhead amounts. On Solaris 2.6 and x86 platforms, the upper limit is approximately 2000m minus overhead amounts. On Linux platforms, the upper limit is approximately 2000m minus overhead amounts. {quote} > The "Heap Size" in HDFS web ui may not be accurate > -------------------------------------------------- > > Key: HDFS-94 > URL: https://issues.apache.org/jira/browse/HDFS-94 > Project: Hadoop HDFS > Issue Type: Bug > Reporter: Tsz Wo (Nicholas), SZE > > It seems that the Heap Size shown in HDFS web UI is not accurate. It keeps > showing 100% of usage. e.g. > {noformat} > Heap Size is 10.01 GB / 10.01 GB (100%) > {noformat} -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.