Is the heap larger on the M5.4x instance? Are you sure it's Cassandra generating the read traffic vs just evicting files read by other systems?
In general, I'd call "more RAM means fewer drive reads" a very expected result regardless of the details, especially when it's the difference between fitting the whole data-set in RAM and not, so I'm not sure it's worth doing that much digging. On Wed, Dec 2, 2020 at 8:41 AM Carl Mueller <carl.muel...@smartthings.com.invalid> wrote: > Oh, this is cassandra 2.2.13 (multi tenant delays) and ubuntu 18.04. > > On Wed, Dec 2, 2020 at 10:35 AM Carl Mueller <carl.muel...@smartthings.com> > wrote: > >> We have a cluster that is experiencing very high disk read I/O in the >> 20-40 MB/sec range on m5.2x (gp2 drives). This is verified via VM metrics >> as well as iotop. >> >> When we switch m5.4x it drops to 60 KB/sec. >> >> There is no difference in network send/recv, read/write request counts. >> >> The graph for read kb/sec mirrors the cpu.iowait. >> >> Compaction would have similar writes to go with reads as the sstables >> were written. Flushing would be almost all writes. Swappiness is zero. >> >> I have done inotifywait to compare read volume on the data and log dirs. >> They are roughly equivalent. >> >> File Caching could be a candidate, I used tobert's : >> https://github.com/tobert/pcstat to see what files are in the file >> cache, and that listed all files at 100%, I would think an overloaded file >> cache would have different files swapping into the cache and partials on >> the data files (data density for the node is about 30 GB). >> >> iotop indicates all the read traffic is from cassandra threads. >> >> Anyone have similar experiences? >> >