Hello,

We just upgraded from 1.1.2->1.1.9. We utilize the byte ordered partitioner (we generate our own hashes). We have not yet upgraded sstables.

Before the upgrade, we had a balanced ring.

After the upgrade, we see:

10.0.4.22 us-east 1a Up Normal 77.66 GB 0.04% Token(bytes[00000000000000000000000000000001]) 10.0.10.23 us-east 1d Up Normal 82.74 GB 0.04% Token(bytes[15555555555555555555555555555555]) 10.0.8.20 us-east 1c Up Normal 81.79 GB 0.04% Token(bytes[2aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa]) 10.0.4.23 us-east 1a Up Normal 82.66 GB 33.84% Token(bytes[40000000000000000000000000000000]) 10.0.10.20 us-east 1d Up Normal 80.21 GB 67.51% Token(bytes[55555555555555555555555555555554]) 10.0.8.23 us-east 1c Up Normal 77.12 GB 99.89% Token(bytes[6aaaaaaaaaaaaaaaaaaaaaaaaaaaaaac]) 10.0.4.21 us-east 1a Up Normal 81.38 GB 66.09% Token(bytes[80000000000000000000000000000000]) 10.0.10.24 us-east 1d Up Normal 83.43 GB 32.41% Token(bytes[95555555555555555555555555555558]) 10.0.8.21 us-east 1c Up Normal 84.42 GB 0.04% Token(bytes[aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa8]) 10.0.4.25 us-east 1a Up Normal 80.06 GB 0.04% Token(bytes[c0000000000000000000000000000000]) 10.0.10.21 us-east 1d Up Normal 83.57 GB 0.04% Token(bytes[d5555555555555555555555555555558]) 10.0.8.24 us-east 1c Up Normal 90.74 GB 0.04% Token(bytes[eaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa8])


Restarting a node essentially changes who own 99% of the ring.

Given we use an RF of 3, and LOCAL_QUORUM consistency for everything, and we are not seeing errors, something seems to be working correctly. Any idea what is going on above? Should I be alarmed?

-Mike

Reply via email to