Downside of too many HFiles

2013-06-12 Thread Rahul Ravindran
Hello,
I am trying to understand the downsides of having a large number of hfiles by 
having a large hbase.hstore.compactionThreshold

  This delays major compaction. However, the amount of data that needs to be 
read and re-written as a single hfile during major compaction will remain the 
same unless we have large number of deletes or expired rows

I understand the random reads will be affected since each hfile may be a 
candidate for the row, but is there any other downside I am missing?


~Rahul.

Re: Downside of too many HFiles

2013-06-12 Thread Thanh Do
you may run into OOM when doing compaction.


On Wed, Jun 12, 2013 at 10:14 AM, Rahul Ravindran rahu...@yahoo.com wrote:

 Hello,
 I am trying to understand the downsides of having a large number of hfiles
 by having a large hbase.hstore.compactionThreshold

   This delays major compaction. However, the amount of data that needs to
 be read and re-written as a single hfile during major compaction will
 remain the same unless we have large number of deletes or expired rows

 I understand the random reads will be affected since each hfile may be a
 candidate for the row, but is there any other downside I am missing?


 ~Rahul.