Thank you all for the response. I figured out the root cause.
I thought all my data was in memtable only but the data was actually being
dumped to the disk. That's why I was noticing the drop in throughput.

On Wed, May 24, 2017 at 9:42 AM, daemeon reiydelle <daeme...@gmail.com>
wrote:

> You speak of increase. Please provide your results. Specific examples, Eg
> 25% increase results in n% increase. Also please include number of nodes,
> size of total keyspace, rep factor, etc.
>
> Hopefully this is a 6 node cluster with several hundred gig per keyspace,
> not some single node free tier box.
>
> “All men dream, but not equally. Those who dream by night in the dusty
> recesses of their minds wake up in the day to find it was vanity, but the
> dreamers of the day are dangerous men, for they may act their dreams with
> open eyes, to make it possible.” — T.E. Lawrence
>
> sent from my mobile
> Daemeon Reiydelle
> skype daemeon.c.m.reiydelle
> USA 415.501.0198 <(415)%20501-0198>
>
> On May 24, 2017 9:32 AM, "preetika tyagi" <preetikaty...@gmail.com> wrote:
>
>> Hi,
>>
>> I'm experimenting with memtable/heap size on my Cassandra server to
>> understand how it impacts the latency/throughput for read requests.
>>
>> I vary heap size (Xms and -Xmx) in jvm.options so memtable will be 1/4 of
>> this. When I increase the heap size and hence memtable, I notice the drop
>> in throughput and increase in latency. I'm also creating the database such
>> that its size doesn't exceed the size of memtable. Therefore, all data
>> exist in memtable and I'm not able to reason why bigger size of memtable is
>> resulting into higher latency/low throughput.
>>
>> Since everything is DRAM, shouldn't the throughput/latency remain same in
>> all the cases?
>>
>> Thanks,
>> Preetika
>>
>

Reply via email to