Hi folks, I'm running a job on an offline node to test how long it takes to run sstablesplit several large sstable.
I'm a bit dismayed to see it took about 22 hours to process a 1.5 gigabyte sstable! I worry about the 32 gigabyte sstable that is my ultimate target to split. This is running on an otherwise unloaded Linux 3.10.0 CentOS 7 server with 4 cpus and 24 gigabytes of ram. Cassandra 3.11.0 and OpenJDK 1.8.0_252 are the installed versions of the software. The machine isn't very busy itself, it looks as though java is only making use of 1 of the 4 processors, and it's not using much of the available 24 gigabytes of memory either, all the memory usage is in the linux buffer cache, which I guess makes sense if it's just working on these large files w/o needing to do a lot of heavy computation on what it reads from them. When you folks run sstablesplit, do you provide specific CASSANDRA_INCLUDE settings to increase its performance? Jim --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org For additional commands, e-mail: user-h...@cassandra.apache.org