Thanks for the article link. That is some comprehensive benchmarking.

Compared to article numbers, my latency numbers are sane enough. I hit ~120
us while you get similar/closer numbers at 99 th percentile.

However, my throughput numbers seems to be wrong. I hit a throughput
kneepoint at 500K ops/sec while yours is around ~6000K.
Order of magnitude difference. Can you please comment on it. :)

thanks
--Pradeep

On Mon, Oct 7, 2019 at 4:46 PM dormando <dorma...@rydia.net> wrote:

> It'll depend on your hardware/test/etc.
>
> https://memcached.org/blog/persistent-memory/ - a thorough performance
> test with some higher end numbers on both throughput and latency along
> with 50th/90th/95/99/etc percentiles and latency point clouds for each sub
> test. That was a big machine though.
>
> ...and no I'm not going to ignore your 50/50 ratio :) the ratio changes
> the results too much. people will have to test with what they expect to
> see. if you do 100% get it scales linearly with the number of worker
> threads/cores. Anything below 100% gets will slowly scale down to 500-1m
> ops/s depending on the hardware and size of the objects.
>
> On Mon, 7 Oct 2019, Pradeep Fernando wrote:
>
> > Hi,
> > Thanks for the help!.
> > After a couple of trial and error configs, I figured out 'concurrency
> parameter' used in memaslap as the culprit.
> > In my configs I was using 16 (constant) as the concurrency input.
> Scaling the value along with thread count gave me sane numbers.
> >
> > The average latency is 120 us for get/set workload ( please ignore my
> 50/50 ratio) and throughput max out around 500K ops/second.
> > graph attached.
> >
> > I know that the benchmarking numbers are heavily dependent on the setup
> and other things. But Is my numbers are faithful enough to be quoted for
> > memcached  single server numbers? In other words, are these numbers are
> way off from a typical memcached performance numbers?
> >
> > thanks
> > --Pradeep
> >
> >
> > On Mon, Oct 7, 2019 at 2:42 PM dormando <dorma...@rydia.net> wrote:
> >       Hey,
> >
> >       Sorry; I'm not going to have any other major insights :) I'd have
> to sit
> >       here playing 20 questions to figure out your test setup. If you're
> running
> >       memaslap from another box, that one needs to be cpu pinned as
> well. If
> >       it's a VM, the governor/etc might not even matter.
> >
> >       Also I don't use memaslap at all, so I can't attest to it. I use
> >       https://github.com/memcached/mc-crusher with the external latency
> sampling
> >       util it comes with. it's not as easy to use though.
> >
> >       On Mon, 7 Oct 2019, Pradeep Fernando wrote:
> >
> >       > Hi Dormando,
> >       > That is great insight.!.
> >       > However, it did not solve the problem. I disabled turbo, as per
> your instructions.
> >       > I even, set the CPU to operate with maximum performance, with
> >       > > cpupower frequency-set --governor performance ( i verified
> this by monitoring cpu freq)
> >       >
> >       > Still the same unexplained behavior. :(. Do you have any other
> suggestions?
> >       >
> >       > thanks
> >       > --Pradeep
> >       >
> >       > On Mon, Oct 7, 2019 at 1:08 PM dormando <dorma...@rydia.net>
> wrote:
> >       >       Hi,
> >       >
> >       >       First as an aside; 1/1 get/set ratio is unusual for mc.
> The gets scale a
> >       >       lot better than sets. If you get into testing more
> "realistic" perf
> >       >       numbers make sure to increase the get rate.
> >       >
> >       >       You're probably just running into CPU scaling. OS's come
> with a "battery
> >       >       saver" or "ondemand" performance scheduler by default.
> They also have
> >       >       turbo. Once you start loading it up more the CPU will stay
> in the higher
> >       >       frequency states or begin to issue turbo, which will lower
> the latency.
> >       >
> >       >       /usr/bin/echo 1 >
> /sys/devices/system/cpu/intel_pstate/no_turbo
> >       >       cpupower frequency-set -g performance
> >       >
> >       >       ... or whatever works for your platform.
> >       >
> >       >       On Mon, 7 Oct 2019, Pradeep Fernando wrote:
> >       >
> >       >       > Hi Devs,
> >       >       > I run memaslap to understand the performance
> characteristics of memcached,
> >       >       >
> >       >       > My setup : both memcached and memaslap running on a
> single machine with NUMA. memcached is bound to NUMA 1. Gave 3GB of memory
> to memcached.
> >       >       > workload : get/set 0.5/0.5
> >       >       >
> >       >       > I increase number of thread from memaslap and observe
> throughput latency numbers.
> >       >       >
> >       >       > I see increase in throughput (expected) but latency
> drops as I crease the load.
> >       >       > The initial average latency is 83 us and it drops to
> 30us with number of threads = 8, this is an unexpected number.-- I expected
> the latency to go up.
> >       >       > Am I reading the output wrong?
> >       >       >
> >       >       > Apologies, if this question does not qualify for this
> mailing list. If so, please direct me to correct list I can get help. :)
> >       >       >
> >       >       > --Pradeep
> >       >       >
> >       >       >
> >       >       >
> >       >       > Thread count = 1
> >       >       >
> >       >       >
> >       >       > Total Statistics (11447336 events)
> >       >       >    Min:        11
> >       >       >    Max:      1663
> >       >       >    Avg:        83
> >       >       >    Geo:     79.83
> >       >       >    Std:     36.39
> >       >       >    Log2 Dist:
> >       >       >        4:       42      594   351733   9527982
> >       >       >        8:   1551101     7451     7103     1330
> >       >       >
> >       >       > cmd_get: 5723671
> >       >       > cmd_set: 5723681
> >       >       > get_misses: 0
> >       >       > written_bytes: 394933167
> >       >       > read_bytes: 343419948
> >       >       > object_bytes: 183157792
> >       >       >
> >       >       > Run time: 60.0s Ops: 11447352 TPS: 190765 Net_rate:
> 11.7M/s
> >       >       >
> >       >       > Thread count = 2
> >       >       >
> >       >       > Total Statistics (30888799 events)
> >       >       >    Min:        12
> >       >       >    Max:      2011
> >       >       >    Avg:        30
> >       >       >    Geo:     29.68
> >       >       >    Std:     15.32
> >       >       >    Log2 Dist:
> >       >       >        4:   170225   25862674   4766668    66398
> >       >       >        8:      154     5017    17493      170
> >       >       >
> >       >       > cmd_get: 15444404
> >       >       > cmd_set: 15444411
> >       >       > get_misses: 0
> >       >       > written_bytes: 1065663678
> >       >       > read_bytes: 926663772
> >       >       > object_bytes: 494221152
> >       >       >
> >       >       > Run time: 60.0s Ops: 30888815 TPS: 514751 Net_rate:
> 31.7M/s
> >       >       >
> >       >       > --
> >       >       >
> >       >       > ---
> >       >       > You received this message because you are subscribed to
> the Google Groups "memcached" group.
> >       >       > To unsubscribe from this group and stop receiving emails
> from it, send an email to memcached+unsubscr...@googlegroups.com.
> >       >       > To view this discussion on the web visit
> https://groups.google.com/d/msgid/memcached/cd593815-c27b-4995-bb7f-21859d9a3187%40googlegroups.com
> .
> >       >       >
> >       >       >
> >       >
> >       >       --
> >       >
> >       >       ---
> >       >       You received this message because you are subscribed to
> the Google Groups "memcached" group.
> >       >       To unsubscribe from this group and stop receiving emails
> from it, send an email to memcached+unsubscr...@googlegroups.com.
> >       >       To view this discussion on the web visit
> https://groups.google.com/d/msgid/memcached/alpine.DEB.2.21.1910071005340.21578%40dskull
> .
> >       >
> >       >
> >       >
> >       > --
> >       > Pradeep Fernando.
> >       >
> >       > --
> >       >
> >       > ---
> >       > You received this message because you are subscribed to the
> Google Groups "memcached" group.
> >       > To unsubscribe from this group and stop receiving emails from
> it, send an email to memcached+unsubscr...@googlegroups.com.
> >       > To view this discussion on the web visit
> https://groups.google.com/d/msgid/memcached/CAPSEm%3Dbas4jg_8ToMP3VBVE3GMfzk9vHaQ7TMXdmqUww67dFsg%40mail.gmail.com
> .
> >       >
> >       >
> >
> >       --
> >
> >       ---
> >       You received this message because you are subscribed to the Google
> Groups "memcached" group.
> >       To unsubscribe from this group and stop receiving emails from it,
> send an email to memcached+unsubscr...@googlegroups.com.
> >       To view this discussion on the web visit
> https://groups.google.com/d/msgid/memcached/alpine.DEB.2.21.1910071140110.21578%40dskull
> .
> >
> >
> >
> > --
> > Pradeep Fernando.
> >
> > --
> >
> > ---
> > You received this message because you are subscribed to the Google
> Groups "memcached" group.
> > To unsubscribe from this group and stop receiving emails from it, send
> an email to memcached+unsubscr...@googlegroups.com.
> > To view this discussion on the web visit
> https://groups.google.com/d/msgid/memcached/CAPSEm%3DZB_BK1VH_kKgTq576CMbV9MS%2Bap_v9DnO_VPfbhnTtsA%40mail.gmail.com
> .
> >
> >
>
> --
>
> ---
> You received this message because you are subscribed to the Google Groups
> "memcached" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to memcached+unsubscr...@googlegroups.com.
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/memcached/alpine.DEB.2.21.1910071343070.21578%40dskull
> .
>


-- 
Pradeep Fernando.

-- 

--- 
You received this message because you are subscribed to the Google Groups 
"memcached" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to memcached+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/memcached/CAPSEm%3DYdy4tRDsgSkE0dfD3U9S0gZHB1QFYVfSAdDgP4vSzPgw%40mail.gmail.com.

Reply via email to