Very encouraging indeed! Thank you so much for sharing your results with the community!!! This is excellent to see and we really appreciate your openness. I have a couple comments/questions.
(1) from the DISCUSS thread re: EOL of 1.1, it seems we'll continue to support 1.x releases for some time, including with overlap of the 2.0 line. Based on the energy of the community, I would guess these will be maintained until 2.2 at least. Therefore, offheap patches that have seen production exposure seem like a reasonable candidate for backport, perhaps in a 1.4 or 1.5 release timeframe. (2) I'm surprised to see your max GC pause go from 11s -> 7s. Do I understand you correctly? This is an improvement to be sure, but I would have expected more gain. Can you elaborate on your GC and heapsize settings? Have you profiled the heaps at all to see where the pressure lies once the bulk of the data pathway is moved to direct memory? Thanks a lot! Nick On Fri, Nov 18, 2016 at 12:11 AM Yu Li <car...@gmail.com> wrote: > Dear all, > > We have backported read path offheap (HBASE-11425) to our customized > hbase-1.1.2 (thanks @Anoop for the help/support) and run it online for more > than a month, and would like to share our experience, for what it's worth > (smile). > > Generally speaking, we gained a better and more stable > throughput/performance with offheap, and below are some details: > > 1. QPS become more stable with offheap > > Performance w/o offheap: > > Performance w/ offheap: > > These data come from our online A/B test cluster (with 450 physical > machines, and each with 256G memory + 64 core) with real world workloads, > it shows using offheap we could gain a more stable throughput as well as > better performance > > Not showing fully online data here because for online we published the > version with both offheap and NettyRpcServer together, so no standalone > comparison data for offheap > > 2. Full GC frequency and cost > > Average Full GC STW time reduce from 11s to 7s with offheap. > > 3. Young GC frequency and cost > > No performance degradation observed with offheap. > > 4. Peak throughput of one single RS > > On Singles Day (11/11), peak throughput of one single RS reached 100K, > among which 90K from Get. Plus internet in/out data we could know the > average result size of get request is ~1KB > > Offheap are used on all online machines (more than 1600 nodes) instead of > LruCache, so the above QPS is gained from offheap bucketcache, along with > NettyRpcServer(HBASE-15756). > Just let us know if any comments. Thanks. > > Best Regards, > Yu > > > > > > > >