[ 
https://issues.apache.org/jira/browse/HBASE-6040?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Hofhansl closed HBASE-6040.
--------------------------------

    
> Use block encoding and HBase handled checksum verification in bulk loading 
> using HFileOutputFormat
> --------------------------------------------------------------------------------------------------
>
>                 Key: HBASE-6040
>                 URL: https://issues.apache.org/jira/browse/HBASE-6040
>             Project: HBase
>          Issue Type: Improvement
>          Components: mapreduce
>    Affects Versions: 0.94.0, 0.96.0
>            Reporter: Anoop Sam John
>            Assignee: Anoop Sam John
>             Fix For: 0.94.1
>
>         Attachments: HBASE-6040_94.patch, HBASE-6040_Trunk.patch
>
>
> When the data is bulk loaded using HFileOutputFormat, we are not using the 
> block encoding and the HBase handled checksum features..  When the writer is 
> created for making the HFile, I am not seeing any such info passing to the 
> WriterBuilder.
> In HFileOutputFormat.getNewWriter(byte[] family, Configuration conf), we dont 
> have these info and do not pass also to the writer... So those HFiles will 
> not have these optimizations..
> Later in LoadIncrementalHFiles.copyHFileHalf(), where we physically divide 
> one HFile(created by the MR) iff it can not belong to just one region, I can 
> see we pass the datablock encoding details and checksum details to the new 
> HFile writer. But this step wont happen normally I think..

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to