>From https://cloud.google.com/compute/docs/machine-types, the Maximum 
egress bandwidth is 32 gbps. From our data, our outgoing bandwidth is 
(175k/60)*4096*152  which is 14gbps, so looks like out network outgoing 
bandwidth is covered as well.

On Sunday, October 18, 2020 at 2:44:59 PM UTC-7 Kishore Komya wrote:

> Thanks, I was asking out of curiosity and here are the specific problems I 
> am chasing.
>
> We are running memcache locally on each node in our spark cluster
>
> 1) I saw conn_yields as high as 170065753 and we use SpyMemcachedClient, 
> which has batch size of 4096 by default, so I changed the -R value to 4096, 
> after that I can see  conn_yields is reduced to zero. 
> However, I expected to see improvement in read latency in getAll method 
> for the memcache, however I am seeing spikes in both p50 and p90. As part 
> of the change, I also increased the memcache size fro 15gb to 54gb and 
> after which I could see evictions from  702281306 to zero and improvement 
> in cache hit ratio from 70% to 95%.  Is the spike in read latency of 
> memcache is coming from the increase in the memcache size?
>
> 2) Next we want to move from local memcache to remote memcache, so that 
> all of our nodes can have access to the remote cache.
>
> Here are our stats for the data
>
> Total data = 50GB
>  max # of keys read per minute  = 700M
> max # of requests read per minute = 700M/4096 =175k
> Our TTL = 24 hours and Total # of put requests required per day = 500M, so 
> we get hit ratio > 99.99%
> Max chunk size = 152 based on our existing data
>
> Lets say we are using n1-highmem-16 in GCP with 16 cores and 104 gb ram, 
> it is clear that one machine can hold the data in ram, but can it support 
> our # of read requests ? How many suck kind of machines do we need? Are 
> there are standard benchmarks available so that we can estimate the # of 
> machines for our data?
>
> Thanks,
> Kishor.
>
>
> On Saturday, October 17, 2020 at 12:23:21 PM UTC-7 Dormando wrote:
>
>> Hey, 
>>
>> I'll answer these inline, but up front: Are you having a specific problem 
>> you're tracking down, or is this just out of curiosity? 
>>
>> None of these are things you should waste time thinking about. Memcached 
>> handles it internally. 
>>
>> > 1. 
>> > 
>> > I setup memcache with 55gb, however total_malloced from stats slabs 
>> says only 41657393216, will that grow as the data grows? 
>>
>> Yes. Memory is lazily allocated, one megabyte at a time. 
>>
>> > 2. 
>> > 
>> > How many pages are allocated per slab? Is that dynamic or is there a 
>> limit? 
>>
>> It's dynamic. If you have a new enough version (> 1.5) they are also 
>> rebalanced automatically as necessary. 
>>
>> > 3. 
>> > 
>> > We use not more than 5-6 slabs, and our largest slab is 300 byte, are 
>> there best practices to limit the object size to 301, so that slab 
>> allocation 
>> > logic is simplified 
>>
>> You can probably ignore this, unless you feel like there's a significant 
>> memory waste going on. You can change the slab growth factor (-f) to 
>> create more classes in the lower numbers than higher numbers but again I 
>> wouldn't bother unless you really need to. 
>>
>> It doesn't "simplify" the logic either way. 
>>
>> -Dormando 
>>
>

-- 

--- 
You received this message because you are subscribed to the Google Groups 
"memcached" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to memcached+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/memcached/05936fad-9071-4fa9-8276-7bcabbde2e40n%40googlegroups.com.

Reply via email to