[ 
https://issues.apache.org/jira/browse/HBASE-15669?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15249291#comment-15249291
 ] 

Anoop Sam John commented on HBASE-15669:
----------------------------------------

bq.stores.get(j).getStoreFileSize();
Do we need a check like hasStoreFileSize()? getStoreFileSize(): 0?

int totalEdits = edit.size();
752           for (int i = 0; i < totalEdits; i++) {
totalEdits? This is cells in this edit right?  In loop condition part u can 
have i < cells.size()?  Other places also similar way.  Will it add more burden 
on other normal edits size calc?  Like we have qualifier check on each and 
every cell.
There can be one WALEdit with a mix of bulk load cells + normal cells?  I dont 
think so.  So we can early out when 1st cell in WALEdit is not a bulk load 
cell?  May be this optimization can come in some other places also?

Can we mark the WALEdit itself as bulk load related? Am not sure.. May be see 
that later.


> HFile size is not considered correctly in a replication request
> ---------------------------------------------------------------
>
>                 Key: HBASE-15669
>                 URL: https://issues.apache.org/jira/browse/HBASE-15669
>             Project: HBase
>          Issue Type: Bug
>          Components: Replication
>    Affects Versions: 1.3.0
>            Reporter: Ashish Singhi
>            Assignee: Ashish Singhi
>             Fix For: 2.0.0, 1.3.0, 1.4.0
>
>         Attachments: HBASE-15669.patch
>
>
> In a single replication request from source cluster a RS can send either at 
> most {{replication.source.size.capacity}} size of data or 
> {{replication.source.nb.capacity}} entries. 
> The size is calculated by considering the cells size in each entry which will 
> get calculated wrongly in case of bulk loaded data replication, in this case 
> we need to consider the size of hfiles not cell.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to