Yu Li,
        What is the global memstore percent that u have?  40 or 50%?
G1GC or CMS?   MSLAB on or off?  MSLAB pool?  What percentage?  All
these will help us in write path off heaping also.

-Anoop-

On Mon, Nov 21, 2016 at 10:58 AM, Yu Li <car...@gmail.com> wrote:
> Thanks @Nick for the feedback, and sorry for the lagging response.
>
> Regarding the backport, I've opened HBASE-17138 for further discussion, FYI.
>
> Regarding the fullGC part, Anoop is right, we still don't have write-path
> offheap so there'll still be fullGC happening. And since the block cache
> normally occupies 30% of the overall (hbase.block.cache.size is 0.3 by
> default), we reduce the heap size by 30% after making blockcache offheap,
> and the fullGC time is reduced at a relative ratio
>
> Best Regards,
> Yu
>
> On 21 November 2016 at 00:29, Anoop John <anoop.hb...@gmail.com> wrote:
>
>> Regarding ur Q on max GC pause Nick, the use case is R+ W workload. This is
>> what I know.. thr write path is still having heavy on heap usage and
>> garbage. Might be a reason I guess.. once we can have write path off heap
>> also, the huge % of our memory need can then be handled by off heap and I
>> believe that can make this still down. The point is then we can work with
>> much reduced xmx. The off heap memory what we use is fixed buffers which
>> live for the life of the RS.  ie. DBBs in Bucket cache/ DBBs in
>> MemstoreChunkPool.
>>
>> -Anoop-
>>
>> On Sunday, November 20, 2016, Nick Dimiduk <ndimi...@apache.org> wrote:
>> > Very encouraging indeed! Thank you so much for sharing your results with
>> > the community!!! This is excellent to see and we really appreciate your
>> > openness. I have a couple comments/questions.
>> >
>> > (1) from the DISCUSS thread re: EOL of 1.1, it seems we'll continue to
>> > support 1.x releases for some time, including with overlap of the 2.0
>> line.
>> > Based on the energy of the community, I would guess these will be
>> > maintained until 2.2 at least. Therefore, offheap patches that have seen
>> > production exposure seem like a reasonable candidate for backport,
>> perhaps
>> > in a 1.4 or 1.5 release timeframe.
>> >
>> > (2) I'm surprised to see your max GC pause go from 11s -> 7s. Do I
>> > understand you correctly? This is an improvement to be sure, but I would
>> > have expected more gain. Can you elaborate on your GC and heapsize
>> > settings? Have you profiled the heaps at all to see where the pressure
>> lies
>> > once the bulk of the data pathway is moved to direct memory?
>> >
>> > Thanks a lot!
>> > Nick
>> >
>> > On Fri, Nov 18, 2016 at 12:11 AM Yu Li <car...@gmail.com> wrote:
>> >
>> >> Dear all,
>> >>
>> >> We have backported read path offheap (HBASE-11425) to our customized
>> >> hbase-1.1.2 (thanks @Anoop for the help/support) and run it online for
>> more
>> >> than a month, and would like to share our experience, for what it's
>> worth
>> >> (smile).
>> >>
>> >> Generally speaking, we gained a better and more stable
>> >> throughput/performance with offheap, and below are some details:
>> >>
>> >> 1. QPS become more stable with offheap
>> >>
>> >> Performance w/o offheap:
>> >>
>> >> Performance w/ offheap:
>> >>
>> >> These data come from our online A/B test cluster (with 450 physical
>> >> machines, and each with 256G memory + 64 core) with real world
>> workloads,
>> >> it shows using offheap we could gain a more stable throughput as well as
>> >> better performance
>> >>
>> >> Not showing fully online data here because for online we published the
>> >> version with both offheap and NettyRpcServer together, so no standalone
>> >> comparison data for offheap
>> >>
>> >> 2. Full GC frequency and cost
>> >>
>> >> Average Full GC STW time reduce from 11s to 7s with offheap.
>> >>
>> >> 3. Young GC frequency and cost
>> >>
>> >> No performance degradation observed with offheap.
>> >>
>> >> 4. Peak throughput of one single RS
>> >>
>> >> On Singles Day (11/11), peak throughput of one single RS reached 100K,
>> >> among which 90K from Get. Plus internet in/out data we could know the
>> >> average result size of get request is ~1KB
>> >>
>> >> Offheap are used on all online machines (more than 1600 nodes) instead
>> of
>> >> LruCache, so the above QPS is gained from offheap bucketcache, along
>> with
>> >> NettyRpcServer(HBASE-15756).
>> >> Just let us know if any comments. Thanks.
>> >>
>> >> Best Regards,
>> >> Yu
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >
>>

Reply via email to