Hi Mark, >>Yep! At least from what I've seen so far, jemalloc is still a little >>faster for 4k random writes even compared to tcmalloc with the patch + >>128MB thread cache. Should have some data soon (mostly just a >>reproduction of Sandisk and Intel's work).
I definitively switch to jemmaloc from my production ceph cluster, I was too tired of this tcmalloc problem (I have hit the bug once or twice, even with TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES) >>Should have some data soon (mostly just a >>>>reproduction of Sandisk and Intel's work). Client side,it could be great to run fio or rados bench with jemalloc too, I have see around 20% improvement vs glibc. "LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libjemalloc.so.1 fio ...." (For my production, I'm running qemu with jemalloc too now) Regards, Alexandre ----- Mail original ----- De: "Mark Nelson" <mnel...@redhat.com> À: "ceph-users" <ceph-users@lists.ceph.com> Envoyé: Lundi 17 Août 2015 16:24:16 Objet: Re: [ceph-users] tcmalloc use a lot of CPU On 08/17/2015 07:03 AM, Alexandre DERUMIER wrote: > Hi, > >>> Is this phenomenon normal?Is there any idea about this problem? > > It's a known problem with tcmalloc (search on the ceph mailing). > > starting osd with "TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=128M" environnement > variable should help. Note that this only works if you use a version of gperftools/tcmalloc newer than 2.1. > > > Another way, is to compile ceph with jemalloc instead tcmalloc (./configure > --with-jemalloc ...) Yep! At least from what I've seen so far, jemalloc is still a little faster for 4k random writes even compared to tcmalloc with the patch + 128MB thread cache. Should have some data soon (mostly just a reproduction of Sandisk and Intel's work). > > > > ----- Mail original ----- > De: "YeYin" <ey...@qq.com> > À: "ceph-users" <ceph-users@lists.ceph.com> > Envoyé: Lundi 17 Août 2015 11:58:26 > Objet: [ceph-users] tcmalloc use a lot of CPU > > Hi, all, > When I do performance test with rados bench, I found tcmalloc consumed a lot > of CPU: > > Samples: 265K of event 'cycles', Event count (approx.): 104385445900 > + 27.58% libtcmalloc.so.4.1.0 [.] tcmalloc::CentralFreeList::FetchFromSpans() > + 15.25% libtcmalloc.so.4.1.0 [.] > tcmalloc::ThreadCache::ReleaseToCentralCache(tcmalloc::ThreadCache::FreeList*, > unsigned long, > + 12.20% libtcmalloc.so.4.1.0 [.] > tcmalloc::CentralFreeList::ReleaseToSpans(void*) > + 1.63% perf [.] append_chain > + 1.39% libtcmalloc.so.4.1.0 [.] > tcmalloc::CentralFreeList::ReleaseListToSpans(void*) > + 1.02% libtcmalloc.so.4.1.0 [.] > tcmalloc::CentralFreeList::RemoveRange(void**, void**, int) > + 0.85% libtcmalloc.so.4.1.0 [.] 0x0000000000017e6f > + 0.75% libtcmalloc.so.4.1.0 [.] > tcmalloc::ThreadCache::IncreaseCacheLimitLocked() > + 0.67% libc-2.12.so [.] memcpy > + 0.53% libtcmalloc.so.4.1.0 [.] operator delete(void*) > > Ceph version: > # ceph --version > ceph version 0.87.2 (87a7cec9ab11c677de2ab23a7668a77d2f5b955e) > > Kernel version: > 3.10.83 > > Is this phenomenon normal? Is there any idea about this problem? > > Thanks. > Ye > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com