Should I use assassinate or removenode? Given that
there is some data on the node. Or will that be
found on the other nodes? Sorry for all the
questions but I really don't want to mess up.
On Mon, Apr 3, 2023 at 7:21 AM Carlos Diaz
<crdiaz...@gmail.com> wrote:
That's what nodetool assassinte will do.
On Sun, Apr 2, 2023 at 10:19 PM David Tinker
<david.tin...@gmail.com> wrote:
Is it possible for me to remove the node
from the cluster i.e. to undo this mess
and get the cluster operating again?
On Mon, Apr 3, 2023 at 7:13 AM Carlos Diaz
<crdiaz...@gmail.com> wrote:
You can leave it in the seed list of
the other nodes, just make sure it's
not included in this node's seed list.
However, if you do decide to fix the
issue with the racks first assassinate
this node (nodetool assassinate <ip>),
and update the rack name before you
restart.
On Sun, Apr 2, 2023 at 10:06 PM David
Tinker <david.tin...@gmail.com> wrote:
It is also in the seeds list for
the other nodes. Should I remove
it from those, restart them one at
a time, then restart it?
/etc/cassandra # grep -i bootstrap *
doesn't show anything so I don't
think I have auto_bootstrap false.
Thanks very much for the help.
On Mon, Apr 3, 2023 at 7:01 AM
Carlos Diaz <crdiaz...@gmail.com>
wrote:
Just remove it from the seed
list in the cassandra.yaml
file and restart the node.
Make sure that auto_bootstrap
is set to true first though.
On Sun, Apr 2, 2023 at 9:59 PM
David Tinker
<david.tin...@gmail.com> wrote:
So likely because I made
it a seed node when I
added it to the cluster it
didn't do the bootstrap
process. How can I recover
this?
On Mon, Apr 3, 2023 at
6:41 AM David Tinker
<david.tin...@gmail.com>
wrote:
Yes replication factor
is 3.
I ran nodetool repair
-pr on all the nodes
(one at a time) and am
still having issues
getting data back from
queries.
I did make the new
node a seed node.
Re "rack4": I assumed
that was just an
indication as to the
physical location of
the server for
redundancy. This one
is separate from the
others so I used rack4.
On Mon, Apr 3, 2023 at
6:30 AM Carlos Diaz
<crdiaz...@gmail.com>
wrote:
I'm assuming that
your replication
factor is 3. If
that's the case,
did you
intentionally put
this node in rack
4? Typically, you
want to add nodes
in multiples of
your replication
factor in order to
keep the "racks"
balanced. In
other words, this
node should have
been added to rack
1, 2 or 3.
Having said that,
you should be able
to easily fix your
problem by running
a nodetool repair
-pr on the new node.
On Sun, Apr 2,
2023 at 8:16 PM
David Tinker
<david.tin...@gmail.com>
wrote:
Hi All
I recently
added a node
to my 3 node
Cassandra
4.0.5 cluster
and now many
reads are not
returning
rows! What do
I need to do
to fix this?
There weren't
any errors in
the logs or
other problems
that I could
see. I
expected the
cluster to
balance itself
but this
hasn't
happened
(yet?). The
nodes are
similar so I
have
num_tokens=256
for each. I am
using the
Murmur3Partitioner.
# nodetool status
Datacenter: dc1
===============
Status=Up/Down
|/
State=Normal/Leaving/Joining/Moving
-- Address
Load
Tokens Owns
(effective)
Host ID
Rack
UN
xxx.xxx.xxx.105
2.65 TiB 256
72.9%
afd02287-3f88-4c6f-8b27-06f7a8192402
rack3
UN
xxx.xxx.xxx.253
2.6 TiB 256
73.9%
e1af72be-e5df-4c6b-a124-c7bc48c6602a
rack2
UN
xxx.xxx.xxx.24
93.82 KiB
256 80.0%
c4e8b4a0-f014-45e6-afb4-648aad4f8500
rack4
UN
xxx.xxx.xxx.107
2.65 TiB 256
73.2%
ab72f017-be96-41d2-9bef-a551dec2c7b5
rack1
# nodetool
netstats
Mode: NORMAL
Not sending
any streams.
Read Repair
Statistics:
Attempted: 0
Mismatch
(Blocking): 0
Mismatch
(Background): 0
Pool Name
Active
Pending
Completed Dropped
Large messages
n/a 0 71754 0
Small messages
n/a 0
8398184 14
Gossip
messages
n/a
0 1303634
0
# nodetool ring
Datacenter: dc1
==========
Address
Rack
Status State
Load
Owns Token
9189523899826545641
xxx.xxx.xxx.24
rack4
Up Normal
93.82 KiB
79.95%
-9194674091837769168
xxx.xxx.xxx.107
rack1
Up
Normal 2.65
TiB
73.25%
-9168781258594813088
xxx.xxx.xxx.253
rack2
Up
Normal 2.6
TiB
73.92%
-9163037340977721917
xxx.xxx.xxx.105
rack3
Up
Normal 2.65
TiB
72.88%
-9148860739730046229
xxx.xxx.xxx.107
rack1
Up
Normal 2.65
TiB
73.25%
-9125240034139323535
xxx.xxx.xxx.253
rack2
Up
Normal 2.6
TiB
73.92%
-9112518853051755414
xxx.xxx.xxx.105
rack3
Up
Normal 2.65
TiB
72.88%
-9100516173422432134
...
This is
causing a
serious
production
issue. Please
help if you can.
Thanks
David