Yo, Fwiw, I use mc-crusher as the "official" benchmark: https://github.com/memcached/mc-crusher tho I need to update the README slightly. will do that in a bit.
I also test on hardware uninterrupted by other things with turbo disabled :) testing on laptops can be really tough since you'll get different frequencies minute to minute. You have to interleave test runs A/B/A/B/A/B then average to get through the noise. Also, make sure to test both binprot/ascii. with ascii multiget you can deeply pipeline requests and get more hashing vs syscalls. Also, test short vs long keys. a for loop and some scripting should get you there. :) I don't really want to add a ton of algo's. More or less a "best average" is fine. Given the syscall overhead, hashing is barely measurable for a lot of workloads. When I switched from jenkins to murmur I did a somewhat comprehensive set of tests then swapped default + left old one in just in case. I highly value good defaults! Libmemcached ended up kitchen-sinking hash algo's and I think that didn't work out well in the long run. I did also test hash chain bucket depth a bit. Finally, loading up different counts of keys (1m, 10m, 100m, etc) and re-running uniform random benchmarks since fairness will affect the bucket depth and thus latency. Sorry if that's a pain in the ass, but this thing is quite widely used and there aren't really beta testers :) Have to be thorough these days. -Dormando On Sun, 17 Mar 2019, 'Eamonn Nugent' via memcached wrote: > Reporting back with very preliminary benchmarks. Somehow, xxh64 is actually > faster than xxh3 on my machine. One thing I forgot to > mention before - I also looked at latencies with xxh32/64, and saw the 99th > percentile latency lowered by about half compared to mmh3. > So it could be beneficial in that sense. Latencies with xxh3 are in the 3.6ms > 99% range, xxh64 go down to about 3.0 (I saw 2.5 > yesterday, maybe testing on a laptop with about a billion chrome tabs open > isn't a brilliant idea), and mmh3 were in the 4.xms range. > All of this with modern options, but with non-modern, xxh64 shone quite a > bit. I was doing my testing there yesterday. > I used the following memtier_benchmark command to stress test: > > ./memtier_benchmark -P memcache_binary -p 11211 --key-prefix="" > --key-maximum=9999999999999999999 > > A lot of this seems to be very architecture dependent. Maybe it would make > sense to include a lot of hash algos long term, and let power > users figure out which they feel like using? Not sure, though, and you're the > expert here :P > > Thanks, > > Eamonn > > On Sun, Mar 17, 2019 at 2:59 PM Eamonn Nugent <eamonn.nug...@demilletech.net> > wrote: > Hiya, > Last night, I was running memtier_benchmark on my laptop (mid-2015 15" MBP, > 2.5GHz 4c i7) and achieved about a 10-15% throughput > improvement on both modern and non-modern settings on the 64 bit variant. 32 > bit variant was about equal in performance (the > results showed them to be within about 3% of each other, but most of the > difference was probably just typical entropy). I was able > to solve the 32/64 bit compile time problem by adding in a wrapper and some > compile-time declarations, so I'd say that's about 50% > solved for x86-based systems. But yeah, with ARM, it could turn interesting. > > As a next-ish step, I'm going to attempt to drop in xxh3, but since it's > still in active development, it's probably not good as > anything more than a tech demo. I'm happy, if it would help, just to go nuts > adding a dozen different algos into hash.c, though > (cityhash/farmhash, as you mentioned). In xxhash's implementation, though, I > played with some compile-time flags to make it a bit > faster, and I've been toying with the idea of modifying it so no seed logic > ever occurs, to maybe gain a couple cycles of speed > increase. I'm also looking into seeing if I can find a pure assembly version > to squeeze a bit more speed out of x86 and ARM > versions. I should probably get one of my ARM systems running and test the > difference... > > But hey, thanks for humoring me. Maybe next I'll take a look at the reading & > processing command steps, and see if there's > anything I can do. Or maybe parallelizing rotl... Hm. I'll keep on with > trying it out :) > > Thanks, > > Eamonn > > > On Sun, Mar 17, 2019 at 2:46 PM dormando <dorma...@rydia.net> wrote: > Hey, > > What exact test did you do? > > Well to be honest I've been wanting to swap in xxhash for a long time, > but > in my own profiling other things show up higher than murmur so I keep > deprioritizing it :) > > One big problem with the hash algo is mc keys can be short and are > hashed one at a time. xxhash is more optimized for longer data > (kilobytes > to megabytes). The original author tries to address this with an updated > algorithm: > https://fastcompression.blogspot.com/2019/03/presenting-xxh3.html > > xxhash makes significant use of instruction parallelism, such that if a > key is 8 bytes or less you could end up waiting for the pipeline more > than murmur. Other algos like cityhash/farmhash are better at short keys > IIRC. Also xx's 32bit algo is a bit slower on 64bit machines... so if I > wanted to use it I was going to test both 32bit and 64bit hashes and > then > have to do compile time testing to figure out which to use. It's also > heavily x86 optimized so we might have to default something else for > ARM. > > Sorry, not debated on the list, just in my own head :) It's not quite as > straightforward as just dropping it in. If you're willing to get all the > conditions tested go nuts! :) > > -Dormando > > On Sat, 16 Mar 2019, eamonn.nugent via memcached wrote: > > > Hi there, > > I started using memcached in prod a week or two ago, and am loving > it. I wanted to give back, and took a look > through the issues board, > > but most of them looked solved. So, in my usual "it's never fast > enough" style, I went and profiled its performance, > and had some fun. > > > > After seeing that MurmurHash3 was taking a good amount of the > execution time, I decided to run a test integrating > one of my old favorite > > hash functions, xxhash. My guess is that Memcached could benefit from > using the hash function, as it is faster than > MMH3 and has several > > native variants. I ran some of my own tests, and found roughly equal > performance, but with no tuning performed on > xxhash. For example, > > using an assembly (x86/arm/etc) version could likely speed up > hashing, along with properly implementing it in > memcached. However, I was > > also running this on a much older Nehalem CPU, so there could be > unseen advantages to one or both of the algorithms > by running them on a > > newer CPU. I'm in the process of fighting with my newer systems to > get libevent installed properly, so I'll report > back with more > > up-to-date tests later. > > > > I did a cursory search, but didn't find any mention of the algo in > the mailing list. If this has been discussed, > though, apologies for > > bringing it up again. On the other hand, I would be happy to write a > PR to add it, using the `hash_algorithm` CLI > arg. > > > > Thanks, > > Eamonn > > > > -- > > > > --- > > You received this message because you are subscribed to the Google > Groups "memcached" group. > > To unsubscribe from this group and stop receiving emails from it, > send an email to > memcached+unsubscr...@googlegroups.com. > > For more options, visit https://groups.google.com/d/optout. > > > > > > -- > > --- > You received this message because you are subscribed to a topic in the > Google Groups "memcached" group. > To unsubscribe from this topic, visit > https://groups.google.com/d/topic/memcached/Y02zPF-WTKg/unsubscribe. > To unsubscribe from this group and all its topics, send an email to > memcached+unsubscr...@googlegroups.com. > For more options, visit https://groups.google.com/d/optout. > > -- > > --- > You received this message because you are subscribed to the Google Groups > "memcached" group. > To unsubscribe from this group and stop receiving emails from it, send an > email to memcached+unsubscr...@googlegroups.com. > For more options, visit https://groups.google.com/d/optout. > > -- --- You received this message because you are subscribed to the Google Groups "memcached" group. To unsubscribe from this group and stop receiving emails from it, send an email to memcached+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout.