Hi, Same thing here:
2 nodes, RF = 2. RCL = 1, WCL = 1. Like Tamar I never ran a major compaction and repair once a week each node. 10.59.21.241 eu-west 1b Up Normal 133.02 GB 50.00% 0 10.58.83.109 eu-west 1b Up Normal 98.12 GB 50.00% 85070591730234615865843651857942052864 What phenomena could explain the result above ? By the way, I have copy the data and import it in a one node dev cluster. There I have run a major compaction and the size of my data has been significantly reduced (to about 32 GB instead of 133 GB). How is that possible ? Do you think that if I run major compaction in both nodes it will balance the load evenly ? Should I run major compaction in production ? 2012/10/10 Tamar Fraenkel <ta...@tok-media.com> > Hi! > I am re-posting this, now that I have more data and still *unbalanced ring > *: > > 3 nodes, > RF=3, RCL=WCL=QUORUM > > > Address DC Rack Status State Load > Owns Token > > 113427455640312821154458202477256070485 > x.x.x.x us-east 1c Up Normal 24.02 GB 33.33% 0 > y.y.y.y us-east 1c Up Normal 33.45 GB 33.33% > 56713727820156410577229101238628035242 > z.z.z.z us-east 1c Up Normal 29.85 GB 33.33% > 113427455640312821154458202477256070485 > > repair runs weekly. > I don't run nodetool compact as I read that this may cause the minor > regular compactions not to run and then I will have to run compact > manually. Is that right? > > Any idea if this means something wrong, and if so, how to solve? > > > Thanks, > * > Tamar Fraenkel * > Senior Software Engineer, TOK Media > > [image: Inline image 1] > > ta...@tok-media.com > Tel: +972 2 6409736 > Mob: +972 54 8356490 > Fax: +972 2 5612956 > > > > > > On Tue, Mar 27, 2012 at 9:12 AM, Tamar Fraenkel <ta...@tok-media.com>wrote: > >> Thanks, I will wait and see as data accumulates. >> Thanks, >> >> *Tamar Fraenkel * >> Senior Software Engineer, TOK Media >> >> [image: Inline image 1] >> >> ta...@tok-media.com >> Tel: +972 2 6409736 >> Mob: +972 54 8356490 >> Fax: +972 2 5612956 >> >> >> >> >> >> On Tue, Mar 27, 2012 at 9:00 AM, R. Verlangen <ro...@us2.nl> wrote: >> >>> Cassandra is built to store tons and tons of data. In my opinion roughly >>> ~ 6MB per node is not enough data to allow it to become a fully balanced >>> cluster. >>> >>> >>> 2012/3/27 Tamar Fraenkel <ta...@tok-media.com> >>> >>>> This morning I have >>>> nodetool ring -h localhost >>>> Address DC Rack Status State Load >>>> Owns Token >>>> >>>> 113427455640312821154458202477256070485 >>>> 10.34.158.33 us-east 1c Up Normal 5.78 MB >>>> 33.33% 0 >>>> 10.38.175.131 us-east 1c Up Normal 7.23 MB >>>> 33.33% 56713727820156410577229101238628035242 >>>> 10.116.83.10 us-east 1c Up Normal 5.02 MB >>>> 33.33% 113427455640312821154458202477256070485 >>>> >>>> Version is 1.0.8. >>>> >>>> >>>> *Tamar Fraenkel * >>>> Senior Software Engineer, TOK Media >>>> >>>> [image: Inline image 1] >>>> >>>> ta...@tok-media.com >>>> Tel: +972 2 6409736 >>>> Mob: +972 54 8356490 >>>> Fax: +972 2 5612956 >>>> >>>> >>>> >>>> >>>> >>>> On Tue, Mar 27, 2012 at 4:05 AM, Maki Watanabe <watanabe.m...@gmail.com >>>> > wrote: >>>> >>>>> What version are you using? >>>>> Anyway try nodetool repair & compact. >>>>> >>>>> maki >>>>> >>>>> >>>>> 2012/3/26 Tamar Fraenkel <ta...@tok-media.com> >>>>> >>>>>> Hi! >>>>>> I created Amazon ring using datastax image and started filling the db. >>>>>> The cluster seems un-balanced. >>>>>> >>>>>> nodetool ring returns: >>>>>> Address DC Rack Status State Load >>>>>> Owns Token >>>>>> >>>>>> 113427455640312821154458202477256070485 >>>>>> 10.34.158.33 us-east 1c Up Normal 514.29 KB >>>>>> 33.33% 0 >>>>>> 10.38.175.131 us-east 1c Up Normal 1.5 MB >>>>>> 33.33% 56713727820156410577229101238628035242 >>>>>> 10.116.83.10 us-east 1c Up Normal 1.5 MB >>>>>> 33.33% 113427455640312821154458202477256070485 >>>>>> >>>>>> [default@tok] describe; >>>>>> Keyspace: tok: >>>>>> Replication Strategy: org.apache.cassandra.locator.SimpleStrategy >>>>>> Durable Writes: true >>>>>> Options: [replication_factor:2] >>>>>> >>>>>> [default@tok] describe cluster; >>>>>> Cluster Information: >>>>>> Snitch: org.apache.cassandra.locator.Ec2Snitch >>>>>> Partitioner: org.apache.cassandra.dht.RandomPartitioner >>>>>> Schema versions: >>>>>> 4687d620-7664-11e1-0000-1bcb936807ff: [10.38.175.131, >>>>>> 10.34.158.33, 10.116.83.10] >>>>>> >>>>>> >>>>>> Any idea what is the cause? >>>>>> I am running similar code on local ring and it is balanced. >>>>>> >>>>>> How can I fix this? >>>>>> >>>>>> Thanks, >>>>>> >>>>>> *Tamar Fraenkel * >>>>>> Senior Software Engineer, TOK Media >>>>>> >>>>>> [image: Inline image 1] >>>>>> >>>>>> ta...@tok-media.com >>>>>> Tel: +972 2 6409736 >>>>>> Mob: +972 54 8356490 >>>>>> Fax: +972 2 5612956 >>>>>> >>>>>> >>>>>> >>>>>> >>>>> >>>> >>> >>> >>> -- >>> With kind regards, >>> >>> Robin Verlangen >>> www.robinverlangen.nl >>> >>> >> >
<<tokLogo.png>>