[ 
https://issues.apache.org/jira/browse/HBASE-15669?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15248289#comment-15248289
 ] 

Ted Yu commented on HBASE-15669:
--------------------------------

{code}
3063       * @param storeFilesSize  Map of store files and its length
{code}
'its length' -> 'their lengths'
{code}
3083            builder.setStoreFileSize(storeFilesSize.get(name));
{code}
What if there is no size for this file ?
I see LOG.warn() below. Is that enough ?
{code}
5314        Map<String, Long> storeFilesSize = new HashMap<String, Long>();
{code}
Rename the variable storeFilesSizes
{code}
762                 LOG.error("Failed to deserialize bulk load entry from wal 
edit. "
763                     + "This its hfiles count will not be added into 
metric.");
{code}
Can more information be added to the above log so that user has more 
information ?

> HFile size is not considered correctly in a replication request
> ---------------------------------------------------------------
>
>                 Key: HBASE-15669
>                 URL: https://issues.apache.org/jira/browse/HBASE-15669
>             Project: HBase
>          Issue Type: Bug
>          Components: Replication
>    Affects Versions: 1.3.0
>            Reporter: Ashish Singhi
>            Assignee: Ashish Singhi
>             Fix For: 2.0.0, 1.3.0, 1.4.0
>
>         Attachments: HBASE-15669.patch
>
>
> In a single replication request from source cluster a RS can send either at 
> most {{replication.source.size.capacity}} size of data or 
> {{replication.source.nb.capacity}} entries. 
> The size is calculated by considering the cells size in each entry which will 
> get calculated wrongly in case of bulk loaded data replication, in this case 
> we need to consider the size of hfiles not cell.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to