[ 
https://issues.apache.org/jira/browse/HBASE-6040?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13286300#comment-13286300
 ] 

Hadoop QA commented on HBASE-6040:
----------------------------------

-1 overall.  Here are the results of testing the latest attachment 
  
http://issues.apache.org/jira/secure/attachment/12530320/HBASE-6040_Trunk.patch
  against trunk revision .

    +1 @author.  The patch does not contain any @author tags.

    -1 tests included.  The patch doesn't appear to include any new or modified 
tests.
                        Please justify why no new tests are needed for this 
patch.
                        Also please list what manual steps were performed to 
verify this patch.

    +1 hadoop2.0.  The patch compiles against the hadoop 2.0 profile.

    +1 javadoc.  The javadoc tool did not generate any warning messages.

    +1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

    -1 findbugs.  The patch appears to cause Findbugs (version 1.3.9) to fail.

    +1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

    +1 core tests.  The patch passed unit tests in .

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/2069//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/2069//console

This message is automatically generated.
                
> Use block encoding and HBase handled checksum verification in bulk loading 
> using HFileOutputFormat
> --------------------------------------------------------------------------------------------------
>
>                 Key: HBASE-6040
>                 URL: https://issues.apache.org/jira/browse/HBASE-6040
>             Project: HBase
>          Issue Type: Improvement
>          Components: mapreduce
>    Affects Versions: 0.94.0, 0.96.0
>            Reporter: Anoop Sam John
>            Assignee: Anoop Sam John
>             Fix For: 0.96.0, 0.94.1
>
>         Attachments: HBASE-6040_94.patch, HBASE-6040_Trunk.patch
>
>
> When the data is bulk loaded using HFileOutputFormat, we are not using the 
> block encoding and the HBase handled checksum features..  When the writer is 
> created for making the HFile, I am not seeing any such info passing to the 
> WriterBuilder.
> In HFileOutputFormat.getNewWriter(byte[] family, Configuration conf), we dont 
> have these info and do not pass also to the writer... So those HFiles will 
> not have these optimizations..
> Later in LoadIncrementalHFiles.copyHFileHalf(), where we physically divide 
> one HFile(created by the MR) iff it can not belong to just one region, I can 
> see we pass the datablock encoding details and checksum details to the new 
> HFile writer. But this step wont happen normally I think..

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira

        

Reply via email to