[ https://issues.apache.org/jira/browse/HBASE-15669?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15248289#comment-15248289 ]
Ted Yu commented on HBASE-15669: -------------------------------- {code} 3063 * @param storeFilesSize Map of store files and its length {code} 'its length' -> 'their lengths' {code} 3083 builder.setStoreFileSize(storeFilesSize.get(name)); {code} What if there is no size for this file ? I see LOG.warn() below. Is that enough ? {code} 5314 Map<String, Long> storeFilesSize = new HashMap<String, Long>(); {code} Rename the variable storeFilesSizes {code} 762 LOG.error("Failed to deserialize bulk load entry from wal edit. " 763 + "This its hfiles count will not be added into metric."); {code} Can more information be added to the above log so that user has more information ? > HFile size is not considered correctly in a replication request > --------------------------------------------------------------- > > Key: HBASE-15669 > URL: https://issues.apache.org/jira/browse/HBASE-15669 > Project: HBase > Issue Type: Bug > Components: Replication > Affects Versions: 1.3.0 > Reporter: Ashish Singhi > Assignee: Ashish Singhi > Fix For: 2.0.0, 1.3.0, 1.4.0 > > Attachments: HBASE-15669.patch > > > In a single replication request from source cluster a RS can send either at > most {{replication.source.size.capacity}} size of data or > {{replication.source.nb.capacity}} entries. > The size is calculated by considering the cells size in each entry which will > get calculated wrongly in case of bulk loaded data replication, in this case > we need to consider the size of hfiles not cell. -- This message was sent by Atlassian JIRA (v6.3.4#6332)