Shorten the retry pause in your clients too as was previously suggested.  Does 
that help?



On Jan 14, 2011, at 9:29, Jonathan Gray <jg...@fb.com> wrote:

> These are a different kind of pause (those caused by blockingStoreFiles).
> 
> This is HBase stepping in and actually blocking updates to a region because 
> compactions have not been able to keep up with the write load.  It could 
> manifest itself in the same way but this is different than shorter pauses 
> caused by periodic offlining of regions during balancing and splits.
> 
> Wayne, have you confirmed in your RegionServer logs that the pauses are 
> associated with splits or region movement, and that you are not seeing the 
> blocking store files issue?
> 
> JG
> 
>> -----Original Message-----
>> From: c...@tarnas.org [mailto:c...@tarnas.org] On Behalf Of Christopher
>> Tarnas
>> Sent: Friday, January 14, 2011 7:29 AM
>> To: user@hbase.apache.org
>> Subject: Re: Cluster Wide Pauses
>> 
>> I have been seeing similar problems and found by raising the
>> hbase.hregion.memstore.block.multiplier
>> to above 12 (default is two) and the hbase.hstore.blockingStoreFiles to 16 I
>> managed to reduce the frequency of the pauses during loads.  My nodes are
>> pretty beefy (48 GB of ram) so I had room to experiment.
>> 
>> From what I understand that gave the regionservers more buffer before
>> they had to halt the world to catch up. The pauses still happen but their
>> impact is less now.
>> 
>> -chris
>> 
>> On Fri, Jan 14, 2011 at 8:34 AM, Wayne <wav...@gmail.com> wrote:
>> 
>>> We have not found any smoking gun here. Most likely these are region
>>> splits on a quickly growing/hot region that all clients get caught waiting 
>>> for.
>>> 
>>> 
>>> On Thu, Jan 13, 2011 at 7:49 AM, Wayne <wav...@gmail.com> wrote:
>>> 
>>>> Thank you for the lead! We will definitely look closer at the OS logs.
>>>> 
>>>> 
>>>> On Thu, Jan 13, 2011 at 6:59 AM, Tatsuya Kawano
>>>> <tatsuya6...@gmail.com
>>>> wrote:
>>>> 
>>>>> 
>>>>> Hi Wayne,
>>>>> 
>>>>>> We are seeing some TCP Resets on all nodes at the same time, and
>>>>> sometimes
>>>>>> quite a lot of them.
>>>>> 
>>>>> 
>>>>> Have you checked this article from Andrei and Cosmin? They had a
>>>>> busy firewall to cause network blackout.
>>>>> 
>>>>> http://hstack.org/hbase-performance-testing/
>>>>> 
>>>>> Maybe it's not your case but just for sure.
>>>>> 
>>>>> Thanks,
>>>>> 
>>>>> --
>>>>> Tatsuya Kawano (Mr.)
>>>>> Tokyo, Japan
>>>>> 
>>>>> 
>>>>> On Jan 13, 2011, at 4:52 AM, Wayne <wav...@gmail.com> wrote:
>>>>> 
>>>>>> We are seeing some TCP Resets on all nodes at the same time, and
>>>>> sometimes
>>>>>> quite a lot of them. We have yet to correlate the pauses to the
>>>>>> TCP
>>>>> resets
>>>>>> but I am starting to wonder if this is partly a network problem.
>>>>>> Does Gigabit Ethernet break down on high volume nodes? Do high
>>>>>> volume nodes
>>>>> use
>>>>>> 10G or Infiniband?
>>>>>> 
>>>>>> 
>>>>>> On Wed, Jan 12, 2011 at 1:52 PM, Stack <st...@duboce.net> wrote:
>>>>>> 
>>>>>>> Jon asks that you describe your loading in the issue.  Would you
>>>>>>> mind doing so.  Ted, stick up in the issue the workload and
>>>>>>> configs. you are running if you don't mind.  I'd like to try it over 
>>>>>>> here.
>>>>>>> Thanks lads,
>>>>>>> St.Ack
>>>>>>> 
>>>>>>> 
>>>>>>> On Wed, Jan 12, 2011 at 9:03 AM, Wayne <wav...@gmail.com>
>> wrote:
>>>>>>>> Added: https://issues.apache.org/jira/browse/HBASE-3438.
>>>>>>>> 
>>>>>>>> On Wed, Jan 12, 2011 at 11:40 AM, Wayne <wav...@gmail.com>
>> wrote:
>>>>>>>> 
>>>>>>>>> We are using 0.89.20100924, r1001068
>>>>>>>>> 
>>>>>>>>> We are seeing see it during heavy write load (which is all the
>>> time),
>>>>>>> but
>>>>>>>>> yesterday we had read load as well as write load and saw both
>>>>>>>>> reads
>>>>> and
>>>>>>>>> writes stop for 10+ seconds. The region size is the biggest
>>>>>>>>> clue we
>>>>> have
>>>>>>>>> found from our tests as setting up a new cluster with a 1GB
>>>>>>>>> max
>>>>> region
>>>>>>> size
>>>>>>>>> and starting to load heavily we will see this a lot for long
>>>>>>>>> long
>>>>> time
>>>>>>>>> frames. Maybe the bigger file gets hung up more easily with a
>>> split?
>>>>>>> Your
>>>>>>>>> description below also fits in that early on the load is not
>>> balanced
>>>>> so
>>>>>>> it
>>>>>>>>> is easier to stop everything on one node as the balance is not
>>> great
>>>>>>> early
>>>>>>>>> on. I will file a JIRA. I will also try to dig deeper into the
>>>>>>>>> logs
>>>>>>> during
>>>>>>>>> the pauses to find a node that might be stuck in a split.
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> On Wed, Jan 12, 2011 at 11:17 AM, Stack <st...@duboce.net>
>> wrote:
>>>>>>>>> 
>>>>>>>>>> On Tue, Jan 11, 2011 at 2:34 PM, Wayne <wav...@gmail.com>
>> wrote:
>>>>>>>>>>> We have very frequent cluster wide pauses that stop all
>>>>>>>>>>> reads and
>>>>>>>>>> writes
>>>>>>>>>>> for seconds.
>>>>>>>>>> 
>>>>>>>>>> All reads and all writes?
>>>>>>>>>> 
>>>>>>>>>> I've seen the pause too for writes.  Its something I've
>>>>>>>>>> always
>>> meant
>>>>>>>>>> to look into.  Friso postulates one cause.  Another that
>>>>>>>>>> we've
>>>>> talked
>>>>>>>>>> of is a region taking a while to come back on line after a
>>>>>>>>>> split
>>> or
>>>>> a
>>>>>>>>>> rebalance for whatever reason.  Client loading might be 'random'
>>>>>>>>>> spraying over lots of random regions but they all get stuck
>>> waiting
>>>>> on
>>>>>>>>>> one particular region to come back online.
>>>>>>>>>> 
>>>>>>>>>> I suppose reads could be blocked for same reason if all are
>>>>>>>>>> trying
>>>>> to
>>>>>>>>>> read from the offlined region.
>>>>>>>>>> 
>>>>>>>>>> What version of hbase are you using?  Splits should be faster
>>>>>>>>>> in
>>>>> 0.90
>>>>>>>>>> now that the split daughters come up on the same region.
>>>>>>>>>> 
>>>>>>>>>> Sorry I don't have a better answer for you.  Need to dig in.
>>>>>>>>>> 
>>>>>>>>>> File a JIRA.  If you want to help out some, stick some data
>>>>>>>>>> up in
>>>>> it.
>>>>>>>>>> Some suggestions would be to enable logging of when we
>> lookup
>>> region
>>>>>>>>>> locations in client and then note when requests go to zero.
>>>>>>>>>> Can
>>> you
>>>>>>>>>> figure what region the clients are waiting on (if they are
>>>>>>>>>> waiting
>>>>> on
>>>>>>>>>> any).  If you can pull out a particular one, try and elicit
>>>>>>>>>> its history at time of blockage.  Is it being moved or
>>>>>>>>>> mid-split?  I suppose it makes sense that bigger regions
>>>>>>>>>> would make the
>>> situation
>>>>>>>>>> 'worse'.  I can take a look at it too.
>>>>>>>>>> 
>>>>>>>>>> St.Ack
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> We are constantly loading data to this cluster of 10 nodes.
>>>>>>>>>>> These pauses can happen as frequently as every minute but
>>> sometimes
>>>>>>> are
>>>>>>>>>> not
>>>>>>>>>>> seen for 15+ minutes. Basically watching the Region server
>>>>>>>>>>> list
>>>>> with
>>>>>>>>>> request
>>>>>>>>>>> counts is the only evidence of what is going on. All reads
>>>>>>>>>>> and
>>>>> writes
>>>>>>>>>>> totally stop and if there is ever any activity it is on the
>>>>>>>>>>> node
>>>>>>> hosting
>>>>>>>>>> the
>>>>>>>>>>> .META. table with a request count of region count + 1. This
>>> problem
>>>>>>>>>> seems to
>>>>>>>>>>> be worse with a larger region size. We tried a 1GB region
>>>>>>>>>>> size
>>> and
>>>>>>> saw
>>>>>>>>>> this
>>>>>>>>>>> more than we saw actual activity (and stopped using a larger
>>> region
>>>>>>> size
>>>>>>>>>>> because of it). We went back to the default region size and
>>>>>>>>>>> it
>>> was
>>>>>>>>>> better,
>>>>>>>>>>> but we had too many regions so now we are up to 512M for a
>>>>>>>>>>> region
>>>>>>> size
>>>>>>>>>> and
>>>>>>>>>>> we are seeing it more again.
>>>>>>>>>>> 
>>>>>>>>>>> Does anyone know what this is? We have dug into all of the
>>>>>>>>>>> logs
>>> to
>>>>>>> find
>>>>>>>>>> some
>>>>>>>>>>> sort of pause but are not able to find anything. Is this an
>>>>>>>>>>> wal
>>>>> hlog
>>>>>>>>>> roll?
>>>>>>>>>>> Is this a region split or compaction? Of course our biggest
>>>>>>>>>>> fear
>>> is
>>>>> a
>>>>>>> GC
>>>>>>>>>>> pause on the master but we do not have java logging turned
>>>>>>>>>>> on
>>> with
>>>>>>> the
>>>>>>>>>>> master to tell. What could possibly stop the entire cluster
>>>>>>>>>>> from
>>>>>>> working
>>>>>>>>>> for
>>>>>>>>>>> seconds at a time very frequently?
>>>>>>>>>>> 
>>>>>>>>>>> Thanks in advance for any ideas of what could be causing this.
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>>> 
>>>>> 
>>>>> 
>>>> 
>>> 

Reply via email to