That's correct. nodetool removenode is strongly preferred when your node is already down. If the node is still functional, use nodetool decommission on the node instead.

On 03/04/2023 16:32, Jeff Jirsa wrote:
FWIW, `nodetool decommission` is strongly preferred. `nodetool removenode` is designed to be run when a host is offline. Only decommission is guaranteed to maintain consistency / correctness, and removemode probably streams a lot more data around than decommission.


On Mon, Apr 3, 2023 at 6:47 AM Bowen Song via user <user@cassandra.apache.org> wrote:

    Use nodetool removenode is strongly preferred in most
    circumstances, and only resort to assassinate if you do not care
    about data consistency or you know there won't be any consistency
    issue (e.g. no new writes and did not run nodetool cleanup).

    Since the size of data on the new node is small, nodetool
    removenode should finish fairly quickly and bring your cluster back.

    Next time when you are doing something like this again, please
    test it out on a non-production environment, make sure everything
    works as expected before moving onto the production.


    On 03/04/2023 06:28, David Tinker wrote:
    Should I use assassinate or removenode? Given that there is some
    data on the node. Or will that be found on the other nodes? Sorry
    for all the questions but I really don't want to mess up.

    On Mon, Apr 3, 2023 at 7:21 AM Carlos Diaz <crdiaz...@gmail.com>
    wrote:

        That's what nodetool assassinte will do.

        On Sun, Apr 2, 2023 at 10:19 PM David Tinker
        <david.tin...@gmail.com> wrote:

            Is it possible for me to remove the node from the cluster
            i.e. to undo this mess and get the cluster operating again?

            On Mon, Apr 3, 2023 at 7:13 AM Carlos Diaz
            <crdiaz...@gmail.com> wrote:

                You can leave it in the seed list of the other nodes,
                just make sure it's not included in this node's seed
                list.  However, if you do decide to fix the issue
                with the racks first assassinate this node (nodetool
                assassinate <ip>), and update the rack name before
                you restart.

                On Sun, Apr 2, 2023 at 10:06 PM David Tinker
                <david.tin...@gmail.com> wrote:

                    It is also in the seeds list for the other nodes.
                    Should I remove it from those, restart them one
                    at a time, then restart it?

                    /etc/cassandra # grep -i bootstrap *
                    doesn't show anything so I don't think I have
                    auto_bootstrap false.

                    Thanks very much for the help.


                    On Mon, Apr 3, 2023 at 7:01 AM Carlos Diaz
                    <crdiaz...@gmail.com> wrote:

                        Just remove it from the seed list in the
                        cassandra.yaml file and restart the node. 
                        Make sure that auto_bootstrap is set to true
                        first though.

                        On Sun, Apr 2, 2023 at 9:59 PM David Tinker
                        <david.tin...@gmail.com> wrote:

                            So likely because I made it a seed node
                            when I added it to the cluster it didn't
                            do the bootstrap process. How can I
                            recover this?

                            On Mon, Apr 3, 2023 at 6:41 AM David
                            Tinker <david.tin...@gmail.com> wrote:

                                Yes replication factor is 3.

                                I ran nodetool repair -pr on all the
                                nodes (one at a time) and am still
                                having issues getting data back from
                                queries.

                                I did make the new node a seed node.

                                Re "rack4": I assumed that was just
                                an indication as to the physical
                                location of the server for
                                redundancy. This one is separate from
                                the others so I used rack4.

                                On Mon, Apr 3, 2023 at 6:30 AM Carlos
                                Diaz <crdiaz...@gmail.com> wrote:

                                    I'm assuming that your
                                    replication factor is 3.  If
                                    that's the case, did you
                                    intentionally put this node in
                                    rack 4?  Typically, you want to
                                    add nodes in multiples of your
                                    replication factor in order to
                                    keep the "racks" balanced.  In
                                    other words, this node should
                                    have been added to rack 1, 2 or 3.

                                    Having said that, you should be
                                    able to easily fix your problem
                                    by running a nodetool repair -pr
                                    on the new node.

                                    On Sun, Apr 2, 2023 at 8:16 PM
                                    David Tinker
                                    <david.tin...@gmail.com> wrote:

                                        Hi All

                                        I recently added a node to my
                                        3 node Cassandra 4.0.5
                                        cluster and now many reads
                                        are not returning rows! What
                                        do I need to do to fix this?
                                        There weren't any errors in
                                        the logs or other problems
                                        that I could see. I expected
                                        the cluster to balance itself
                                        but this hasn't happened
                                        (yet?). The nodes are similar
                                        so I have num_tokens=256 for
                                        each. I am using the
                                        Murmur3Partitioner.

                                        # nodetool status
                                        Datacenter: dc1
                                        ===============
                                        Status=Up/Down
                                        |/
                                        State=Normal/Leaving/Joining/Moving
                                        --  Address        Load    
                                        Tokens  Owns (effective)
                                         Host ID           Rack
                                        UN  xxx.xxx.xxx.105  2.65 TiB
                                        256     72.9%
                                        afd02287-3f88-4c6f-8b27-06f7a8192402
                                         rack3
                                        UN  xxx.xxx.xxx.253  2.6 TiB
                                         256     73.9%
                                        e1af72be-e5df-4c6b-a124-c7bc48c6602a
                                         rack2
                                        UN  xxx.xxx.xxx.24   93.82
                                        KiB  256     80.0%
                                        c4e8b4a0-f014-45e6-afb4-648aad4f8500
                                         rack4
                                        UN  xxx.xxx.xxx.107  2.65 TiB
                                        256     73.2%
                                        ab72f017-be96-41d2-9bef-a551dec2c7b5
                                         rack1

                                        # nodetool netstats
                                        Mode: NORMAL
                                        Not sending any streams.
                                        Read Repair Statistics:
                                        Attempted: 0
                                        Mismatch (Blocking): 0
                                        Mismatch (Background): 0
                                        Pool Name  Active Pending
                                         Completed Dropped
                                        Large messages    n/a 0  71754 0
                                        Small messages    n/a 0
                                         8398184  14
                                        Gossip messages           n/a
                                                0    1303634     0

                                        # nodetool ring
                                        Datacenter: dc1
                                        ==========
                                        Address         Rack    
                                         Status State   Load        
                                         Owns  Token
                                         9189523899826545641
                                        xxx.xxx.xxx.24        rack4  
                                          Up Normal  93.82 KiB 79.95%
                                         -9194674091837769168
                                        xxx.xxx.xxx.107       rack1  
                                            Up     Normal  2.65 TiB  
                                             73.25%  -9168781258594813088
                                        xxx.xxx.xxx.253       rack2  
                                            Up     Normal  2.6 TiB  
                                              73.92%
                                         -9163037340977721917
                                        xxx.xxx.xxx.105       rack3  
                                            Up     Normal  2.65 TiB  
                                             72.88%  -9148860739730046229
                                        xxx.xxx.xxx.107       rack1  
                                            Up     Normal  2.65 TiB  
                                             73.25%  -9125240034139323535
                                        xxx.xxx.xxx.253       rack2  
                                            Up     Normal  2.6 TiB  
                                              73.92%
                                         -9112518853051755414
                                        xxx.xxx.xxx.105       rack3  
                                            Up     Normal  2.65 TiB  
                                             72.88%  -9100516173422432134
                                        ...

                                        This is causing a serious
                                        production issue. Please help
                                        if you can.

                                        Thanks
                                        David


Reply via email to