[ 
https://issues.apache.org/jira/browse/HBASE-5458?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13556739#comment-13556739
 ] 

Hadoop QA commented on HBASE-5458:
----------------------------------

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12565392/HBASE-5458-trunk-2.patch
  against trunk revision .

    {color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

    {color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
                        Please justify why no new tests are needed for this 
patch.
                        Also please list what manual steps were performed to 
verify this patch.

    {color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

    {color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

    {color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

    {color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

    {color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

    {color:green}+1 lineLengths{color}.  The patch does not introduce lines 
longer than 100

     {color:red}-1 core tests{color}.  The patch failed these unit tests:
                       
org.apache.hadoop.hbase.regionserver.TestScanWithBloomError
                  org.apache.hadoop.hbase.io.hfile.TestHFile
                  org.apache.hadoop.hbase.io.hfile.TestHFileBlockCompatibility
                  org.apache.hadoop.hbase.io.hfile.TestChecksum
                  org.apache.hadoop.hbase.io.hfile.TestHFileWriterV2

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/4077//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/4077//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/4077//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop2-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/4077//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/4077//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/4077//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/4077//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/4077//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/4077//console

This message is automatically generated.
                
> Thread safety issues with Compression.Algorithm.GZ and CompressionTest
> ----------------------------------------------------------------------
>
>                 Key: HBASE-5458
>                 URL: https://issues.apache.org/jira/browse/HBASE-5458
>             Project: HBase
>          Issue Type: Bug
>          Components: io
>    Affects Versions: 0.90.5, 0.92.2, 0.96.0, 0.94.4
>            Reporter: David McIntosh
>            Assignee: Elliott Clark
>            Priority: Minor
>         Attachments: HBASE-5458-090-0.patch, HBASE-5458-090-1.patch, 
> HBASE-5458-090-2.patch, HBASE-5458-092-2.patch, HBASE-5458-094-2.patch, 
> HBASE-5458-trunk-2.patch
>
>
> I've seen some occasional NullPointerExceptions in 
> ZlibFactory.isNativeZlibLoaded(conf) during region server startups and the 
> completebulkload process.  This is being caused by a null configuration 
> getting passed to the isNativeZlibLoaded method.  I think this happens when 2 
> or more threads call the CompressionTest.testCompression method at once.  If 
> the GZ algorithm has not been tested yet both threads could continue on and 
> attempt to load the compressor.  For GZ the getCodec method is not thread 
> safe which could lead to one thread getting a reference to a GzipCodec that 
> has a null configuration.
> {code}
> current:
>       DefaultCodec getCodec(Configuration conf) {
>         if (codec == null) {
>           codec = new GzipCodec();
>           codec.setConf(new Configuration(conf));
>         }
>         return codec;
>       }
> {code}
> one possible fix would be something like this:
> {code}
>       DefaultCodec getCodec(Configuration conf) {
>         if (codec == null) {
>           GzipCodec gzip = new GzipCodec();
>           gzip.setConf(new Configuration(conf));
>           codec = gzip;
>         }
>         return codec;
>       }
> {code}
> But that may not be totally safe without some synchronization.  An upstream 
> fix in CompressionTest could also prevent multi thread access to 
> GZ.getCodec(conf)
> exceptions:
> 12/02/21 16:11:56 ERROR handler.OpenRegionHandler: Failed open of 
> region=all-monthly,,1326263896983.bf574519a95263ec23a2bad9f5b8cbf4.
> java.io.IOException: java.lang.NullPointerException
>         at 
> org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:89)
>         at 
> org.apache.hadoop.hbase.regionserver.HRegion.checkCompressionCodecs(HRegion.java:2670)
>         at 
> org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:2659)
>         at 
> org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:2647)
>         at 
> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.openRegion(OpenRegionHandler.java:312)
>         at 
> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.process(OpenRegionHandler.java:99)
>         at 
> org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:158)
>         at 
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>         at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>         at java.lang.Thread.run(Thread.java:662)
> Caused by: java.lang.NullPointerException
>         at 
> org.apache.hadoop.io.compress.zlib.ZlibFactory.isNativeZlibLoaded(ZlibFactory.java:63)
>         at 
> org.apache.hadoop.io.compress.GzipCodec.getCompressorType(GzipCodec.java:166)
>         at 
> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:100)
>         at 
> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:112)
>         at 
> org.apache.hadoop.hbase.io.hfile.Compression$Algorithm.getCompressor(Compression.java:236)
>         at 
> org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:84)
>         ... 9 more
> Caused by: java.io.IOException: java.lang.NullPointerException
>         at 
> org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:89)
>         at 
> org.apache.hadoop.hbase.io.hfile.HFile$Reader.readTrailer(HFile.java:890)
>         at 
> org.apache.hadoop.hbase.io.hfile.HFile$Reader.loadFileInfo(HFile.java:819)
>         at 
> org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.groupOrSplit(LoadIncrementalHFiles.java:405)
>         at 
> org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles$2.call(LoadIncrementalHFiles.java:323)
>         at 
> org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles$2.call(LoadIncrementalHFiles.java:321)
>         at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
>         at java.util.concurrent.FutureTask.run(FutureTask.java:138)
>         at 
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>         at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>         at java.lang.Thread.run(Thread.java:662)
> Caused by: java.lang.NullPointerException
>         at 
> org.apache.hadoop.io.compress.zlib.ZlibFactory.isNativeZlibLoaded(ZlibFactory.java:63)
>         at 
> org.apache.hadoop.io.compress.GzipCodec.getCompressorType(GzipCodec.java:166)
>         at 
> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:100)
>         at 
> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:112)
>         at 
> org.apache.hadoop.hbase.io.hfile.Compression$Algorithm.getCompressor(Compression.java:236)
>         at 
> org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:84)
>         ... 10 more

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to