Sent from Mail for Windows 10
Thanks for feedback. Just to elaborate more, I am currently writing 600m rows per hour and need to understand if this is about on target or if there are better ways to write or perhaps structure the keyspaces and table structures. And I can use the Cassandra Stress tool to get potential maximum throughput stats. Or use the schema provided(keyspace/table definitions for a stress test) Cassandra, being a scale-out database, can load any arbitrary number of records per hour. The best way to do this is for your given data model, find what your max throughput is on a single node by scaling the number of clients until you start seeing errors (or hit your latency SLA) then pull back by 15-20%. From there, it's a matter of linearly scaling clients and nodes until you hit your desired throughput. I recommend taking a look at TLP-Stress as it's a bit easier to use and understand: https://thelastpickle.com/blog/2018/10/31/tlp-stress-intro.html Best. Marc Selwan | DataStax | PM, Server Team | (925) 413-7079 | Twitter Quick links | DataStax | Training | Documentation | Downloads On Tue, Aug 20, 2019 at 7:16 AM Surbhi Gupta <surbhi.gupt...@gmail.com> wrote: Have you tried ycsa? It is a tool from yahoo for stress testing nosql databases. On Tue, Aug 20, 2019 at 3:34 AM <yanga.zuke...@condorgreen.com> wrote: Hi Everyone, Anyone before who have bused Cassandra-stress. I want to test if it’s possible to load 600 milllions records per hour in Cassandra or Find a better way to optimize Cassandra for this case. Any help will be highly appreciated. Sent from Mail for Window