Just remove it from the seed list in the cassandra.yaml file and restart the node. Make sure that auto_bootstrap is set to true first though.
On Sun, Apr 2, 2023 at 9:59 PM David Tinker <david.tin...@gmail.com> wrote: > So likely because I made it a seed node when I added it to the cluster it > didn't do the bootstrap process. How can I recover this? > > On Mon, Apr 3, 2023 at 6:41 AM David Tinker <david.tin...@gmail.com> > wrote: > >> Yes replication factor is 3. >> >> I ran nodetool repair -pr on all the nodes (one at a time) and am still >> having issues getting data back from queries. >> >> I did make the new node a seed node. >> >> Re "rack4": I assumed that was just an indication as to the physical >> location of the server for redundancy. This one is separate from the others >> so I used rack4. >> >> On Mon, Apr 3, 2023 at 6:30 AM Carlos Diaz <crdiaz...@gmail.com> wrote: >> >>> I'm assuming that your replication factor is 3. If that's the case, did >>> you intentionally put this node in rack 4? Typically, you want to add >>> nodes in multiples of your replication factor in order to keep the "racks" >>> balanced. In other words, this node should have been added to rack 1, 2 or >>> 3. >>> >>> Having said that, you should be able to easily fix your problem by >>> running a nodetool repair -pr on the new node. >>> >>> On Sun, Apr 2, 2023 at 8:16 PM David Tinker <david.tin...@gmail.com> >>> wrote: >>> >>>> Hi All >>>> >>>> I recently added a node to my 3 node Cassandra 4.0.5 cluster and now >>>> many reads are not returning rows! What do I need to do to fix this? There >>>> weren't any errors in the logs or other problems that I could see. I >>>> expected the cluster to balance itself but this hasn't happened (yet?). The >>>> nodes are similar so I have num_tokens=256 for each. I am using the >>>> Murmur3Partitioner. >>>> >>>> # nodetool status >>>> Datacenter: dc1 >>>> =============== >>>> Status=Up/Down >>>> |/ State=Normal/Leaving/Joining/Moving >>>> -- Address Load Tokens Owns (effective) Host ID >>>> Rack >>>> UN xxx.xxx.xxx.105 2.65 TiB 256 72.9% >>>> afd02287-3f88-4c6f-8b27-06f7a8192402 rack3 >>>> UN xxx.xxx.xxx.253 2.6 TiB 256 73.9% >>>> e1af72be-e5df-4c6b-a124-c7bc48c6602a rack2 >>>> UN xxx.xxx.xxx.24 93.82 KiB 256 80.0% >>>> c4e8b4a0-f014-45e6-afb4-648aad4f8500 rack4 >>>> UN xxx.xxx.xxx.107 2.65 TiB 256 73.2% >>>> ab72f017-be96-41d2-9bef-a551dec2c7b5 rack1 >>>> >>>> # nodetool netstats >>>> Mode: NORMAL >>>> Not sending any streams. >>>> Read Repair Statistics: >>>> Attempted: 0 >>>> Mismatch (Blocking): 0 >>>> Mismatch (Background): 0 >>>> Pool Name Active Pending Completed Dropped >>>> Large messages n/a 0 71754 0 >>>> Small messages n/a 0 8398184 14 >>>> Gossip messages n/a 0 1303634 0 >>>> >>>> # nodetool ring >>>> Datacenter: dc1 >>>> ========== >>>> Address Rack Status State Load Owns >>>> Token >>>> >>>> 9189523899826545641 >>>> xxx.xxx.xxx.24 rack4 Up Normal 93.82 KiB 79.95% >>>> -9194674091837769168 >>>> xxx.xxx.xxx.107 rack1 Up Normal 2.65 TiB 73.25% >>>> -9168781258594813088 >>>> xxx.xxx.xxx.253 rack2 Up Normal 2.6 TiB 73.92% >>>> -9163037340977721917 >>>> xxx.xxx.xxx.105 rack3 Up Normal 2.65 TiB 72.88% >>>> -9148860739730046229 >>>> xxx.xxx.xxx.107 rack1 Up Normal 2.65 TiB 73.25% >>>> -9125240034139323535 >>>> xxx.xxx.xxx.253 rack2 Up Normal 2.6 TiB 73.92% >>>> -9112518853051755414 >>>> xxx.xxx.xxx.105 rack3 Up Normal 2.65 TiB 72.88% >>>> -9100516173422432134 >>>> ... >>>> >>>> This is causing a serious production issue. Please help if you can. >>>> >>>> Thanks >>>> David >>>> >>>> >>>> >>>>