[ https://issues.apache.org/jira/browse/HBASE-7391?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13733419#comment-13733419 ]
Hadoop QA commented on HBASE-7391: ---------------------------------- {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12596819/HBASE-7391_1.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 12 new or modified tests. {color:green}+1 hadoop1.0{color}. The patch compiles against the hadoop 1.0 profile. {color:green}+1 hadoop2.0{color}. The patch compiles against the hadoop 2.0 profile. {color:red}-1 javadoc{color}. The javadoc tool appears to have generated 18 warning messages. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 lineLengths{color}. The patch does not introduce lines longer than 100 {color:green}+1 site{color}. The mvn site goal succeeds with this patch. {color:red}-1 core tests{color}. The patch failed these unit tests: Test results: https://builds.apache.org/job/PreCommit-HBASE-Build/6653//testReport/ Findbugs warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/6653//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html Findbugs warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/6653//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html Findbugs warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/6653//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html Findbugs warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/6653//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html Findbugs warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/6653//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html Findbugs warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/6653//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html Findbugs warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/6653//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html Findbugs warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/6653//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html Console output: https://builds.apache.org/job/PreCommit-HBASE-Build/6653//console This message is automatically generated. > Review/improve HLog compression's memory consumption > ---------------------------------------------------- > > Key: HBASE-7391 > URL: https://issues.apache.org/jira/browse/HBASE-7391 > Project: HBase > Issue Type: Bug > Reporter: Jean-Daniel Cryans > Assignee: ramkrishna.s.vasudevan > Fix For: 0.95.2 > > Attachments: HBASE-7391_1.patch > > > From Ram in > http://mail-archives.apache.org/mod_mbox/hbase-dev/201205.mbox/%3C00bc01cd31e6$7caf1320$760d3960$%25vasude...@huawei.com%3E: > {quote} > One small observation after giving +1 on the RC. > The WAL compression feature causes OOME and causes Full GC. > The problem is, if we have 1500 regions and I need to create recovered.edits > for each of the region (I don’t have much data in the regions (~300MB)). > Now when I try to build the dictionary there is a Node object getting > created. > Each node object occupies 32 bytes. > We have 5 such dictionaries. > Initially we create indexToNodes array and its size is 32767. > So now we have 32*5*32767 = ~5MB. > Now I have 1500 regions. > So 5MB*1500 = ~7GB.(Excluding actual data). This seems to a very high > initial memory foot print and this never allows me to split the logs and I > am not able to make the cluster up at all. > Our configured heap size was 8GB, tested in 3 node cluster with 5000 > regions, very less data( 1GB in hdfs cluster including replication), some > small data is spread evenly across all regions. > The formula is 32(Node object size)*5(No of dictionary)*32767(no of node > objects)*noofregions. > {quote} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira