It'll depend on your hardware/test/etc. https://memcached.org/blog/persistent-memory/ - a thorough performance test with some higher end numbers on both throughput and latency along with 50th/90th/95/99/etc percentiles and latency point clouds for each sub test. That was a big machine though.
...and no I'm not going to ignore your 50/50 ratio :) the ratio changes the results too much. people will have to test with what they expect to see. if you do 100% get it scales linearly with the number of worker threads/cores. Anything below 100% gets will slowly scale down to 500-1m ops/s depending on the hardware and size of the objects. On Mon, 7 Oct 2019, Pradeep Fernando wrote: > Hi, > Thanks for the help!. > After a couple of trial and error configs, I figured out 'concurrency > parameter' used in memaslap as the culprit. > In my configs I was using 16 (constant) as the concurrency input. Scaling the > value along with thread count gave me sane numbers. > > The average latency is 120 us for get/set workload ( please ignore my 50/50 > ratio) and throughput max out around 500K ops/second. > graph attached. > > I know that the benchmarking numbers are heavily dependent on the setup and > other things. But Is my numbers are faithful enough to be quoted for > memcached single server numbers? In other words, are these numbers are way > off from a typical memcached performance numbers? > > thanks > --Pradeep > > > On Mon, Oct 7, 2019 at 2:42 PM dormando <dorma...@rydia.net> wrote: > Hey, > > Sorry; I'm not going to have any other major insights :) I'd have to sit > here playing 20 questions to figure out your test setup. If you're > running > memaslap from another box, that one needs to be cpu pinned as well. If > it's a VM, the governor/etc might not even matter. > > Also I don't use memaslap at all, so I can't attest to it. I use > https://github.com/memcached/mc-crusher with the external latency > sampling > util it comes with. it's not as easy to use though. > > On Mon, 7 Oct 2019, Pradeep Fernando wrote: > > > Hi Dormando, > > That is great insight.!. > > However, it did not solve the problem. I disabled turbo, as per your > instructions. > > I even, set the CPU to operate with maximum performance, with > > > cpupower frequency-set --governor performance ( i verified this by > monitoring cpu freq) > > > > Still the same unexplained behavior. :(. Do you have any other > suggestions? > > > > thanks > > --Pradeep > > > > On Mon, Oct 7, 2019 at 1:08 PM dormando <dorma...@rydia.net> wrote: > > Hi, > > > > First as an aside; 1/1 get/set ratio is unusual for mc. The > gets scale a > > lot better than sets. If you get into testing more "realistic" > perf > > numbers make sure to increase the get rate. > > > > You're probably just running into CPU scaling. OS's come with a > "battery > > saver" or "ondemand" performance scheduler by default. They > also have > > turbo. Once you start loading it up more the CPU will stay in > the higher > > frequency states or begin to issue turbo, which will lower the > latency. > > > > /usr/bin/echo 1 > /sys/devices/system/cpu/intel_pstate/no_turbo > > cpupower frequency-set -g performance > > > > ... or whatever works for your platform. > > > > On Mon, 7 Oct 2019, Pradeep Fernando wrote: > > > > > Hi Devs, > > > I run memaslap to understand the performance characteristics > of memcached, > > > > > > My setup : both memcached and memaslap running on a single > machine with NUMA. memcached is bound to NUMA 1. Gave 3GB of memory to > memcached. > > > workload : get/set 0.5/0.5 > > > > > > I increase number of thread from memaslap and observe > throughput latency numbers. > > > > > > I see increase in throughput (expected) but latency drops as > I crease the load. > > > The initial average latency is 83 us and it drops to 30us > with number of threads = 8, this is an unexpected number.-- I expected the > latency to go up. > > > Am I reading the output wrong? > > > > > > Apologies, if this question does not qualify for this mailing > list. If so, please direct me to correct list I can get help. :) > > > > > > --Pradeep > > > > > > > > > > > > Thread count = 1 > > > > > > > > > Total Statistics (11447336 events) > > > Min: 11 > > > Max: 1663 > > > Avg: 83 > > > Geo: 79.83 > > > Std: 36.39 > > > Log2 Dist: > > > 4: 42 594 351733 9527982 > > > 8: 1551101 7451 7103 1330 > > > > > > cmd_get: 5723671 > > > cmd_set: 5723681 > > > get_misses: 0 > > > written_bytes: 394933167 > > > read_bytes: 343419948 > > > object_bytes: 183157792 > > > > > > Run time: 60.0s Ops: 11447352 TPS: 190765 Net_rate: 11.7M/s > > > > > > Thread count = 2 > > > > > > Total Statistics (30888799 events) > > > Min: 12 > > > Max: 2011 > > > Avg: 30 > > > Geo: 29.68 > > > Std: 15.32 > > > Log2 Dist: > > > 4: 170225 25862674 4766668 66398 > > > 8: 154 5017 17493 170 > > > > > > cmd_get: 15444404 > > > cmd_set: 15444411 > > > get_misses: 0 > > > written_bytes: 1065663678 > > > read_bytes: 926663772 > > > object_bytes: 494221152 > > > > > > Run time: 60.0s Ops: 30888815 TPS: 514751 Net_rate: 31.7M/s > > > > > > -- > > > > > > --- > > > You received this message because you are subscribed to the > Google Groups "memcached" group. > > > To unsubscribe from this group and stop receiving emails from > it, send an email to memcached+unsubscr...@googlegroups.com. > > > To view this discussion on the web visit > https://groups.google.com/d/msgid/memcached/cd593815-c27b-4995-bb7f-21859d9a3187%40googlegroups.com. > > > > > > > > > > -- > > > > --- > > You received this message because you are subscribed to the > Google Groups "memcached" group. > > To unsubscribe from this group and stop receiving emails from > it, send an email to memcached+unsubscr...@googlegroups.com. > > To view this discussion on the web visit > https://groups.google.com/d/msgid/memcached/alpine.DEB.2.21.1910071005340.21578%40dskull. > > > > > > > > -- > > Pradeep Fernando. > > > > -- > > > > --- > > You received this message because you are subscribed to the Google > Groups "memcached" group. > > To unsubscribe from this group and stop receiving emails from it, > send an email to memcached+unsubscr...@googlegroups.com. > > To view this discussion on the web visit > https://groups.google.com/d/msgid/memcached/CAPSEm%3Dbas4jg_8ToMP3VBVE3GMfzk9vHaQ7TMXdmqUww67dFsg%40mail.gmail.com. > > > > > > -- > > --- > You received this message because you are subscribed to the Google > Groups "memcached" group. > To unsubscribe from this group and stop receiving emails from it, send > an email to memcached+unsubscr...@googlegroups.com. > To view this discussion on the web visit > https://groups.google.com/d/msgid/memcached/alpine.DEB.2.21.1910071140110.21578%40dskull. > > > > -- > Pradeep Fernando. > > -- > > --- > You received this message because you are subscribed to the Google Groups > "memcached" group. > To unsubscribe from this group and stop receiving emails from it, send an > email to memcached+unsubscr...@googlegroups.com. > To view this discussion on the web visit > https://groups.google.com/d/msgid/memcached/CAPSEm%3DZB_BK1VH_kKgTq576CMbV9MS%2Bap_v9DnO_VPfbhnTtsA%40mail.gmail.com. > > -- --- You received this message because you are subscribed to the Google Groups "memcached" group. To unsubscribe from this group and stop receiving emails from it, send an email to memcached+unsubscr...@googlegroups.com. To view this discussion on the web visit https://groups.google.com/d/msgid/memcached/alpine.DEB.2.21.1910071343070.21578%40dskull.