[
https://issues.apache.org/jira/browse/HBASE-834?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Billy Pearson updated HBASE-834:
--------------------------------
Fix Version/s: 0.3.0
0.2.1
Changing this to assign to 2.1 and 3.0
Just noticed we now have a problem of never removing data from
(deletes,ttl,max_version) from mapfiles If we never compact all the mapfiles at
some point.
Currently the only way we do is after a split or if the mapfile sizes are just
right to include all the mapfile in the incremental compaction.
> Upper bound on files we compact at any one time
> -----------------------------------------------
>
> Key: HBASE-834
> URL: https://issues.apache.org/jira/browse/HBASE-834
> Project: Hadoop HBase
> Issue Type: Improvement
> Reporter: stack
> Priority: Minor
> Fix For: 0.2.1, 0.3.0
>
>
> From Billy in HBASE-64, which we closed because it got pulled all over the
> place:
> {code}
> Currently we do compaction on a region when the
> hbase.hstore.compactionThreshold is reached - default 3
> I thank we should configure a max number of mapfiles to compact at one time
> simulator to doing a minor compaction in bigtable. This keep compaction's
> form getting tied up in one region to long letting other regions get way to
> many memcache flushes making compaction take longer and longer for each region
> If we did that when a regions updates start to slack off the max number will
> eventuly include all mapfiles causeing a major compaction on that region.
> Unlike big table this would leave the master out of the process and letting
> the region server handle the major compaction when it has time.
> When doing a minor compaction on a few files I thank we should compact the
> newest mapfiles first leave the larger/older ones for when we have low
> updates to a region.
> {code}
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.