Re: initial token crashes cassandra

2014-05-17 Thread Colin
You may have used the old random partitioner token generator.  Use the murmur 
partitioner token generator instead.

--
Colin
320-221-9531


 On May 17, 2014, at 1:15 PM, Tim Dunphy bluethu...@gmail.com wrote:
 
 Hey all,
 
  I've set my initial_token in cassandra 2.0.7 using a python script I found 
 at the datastax wiki. 
 
 I've set the value like this:
 
 initial_token: 85070591730234615865843651857942052864
 
 And cassandra crashes when I try to start it:
 
 [root@beta:/etc/alternatives/cassandrahome] #./bin/cassandra -f
  INFO 18:14:38,511 Logging initialized
  INFO 18:14:38,560 Loading settings from 
 file:/usr/local/apache-cassandra-2.0.7/conf/cassandra.yaml
  INFO 18:14:39,151 Data files directories: [/var/lib/cassandra/data]
  INFO 18:14:39,152 Commit log directory: /var/lib/cassandra/commitlog
  INFO 18:14:39,153 DiskAccessMode 'auto' determined to be mmap, 
 indexAccessMode is mmap
  INFO 18:14:39,153 disk_failure_policy is stop
  INFO 18:14:39,153 commit_failure_policy is stop
  INFO 18:14:39,161 Global memtable threshold is enabled at 251MB
  INFO 18:14:39,362 Not using multi-threaded compaction
 ERROR 18:14:39,365 Fatal configuration error
 org.apache.cassandra.exceptions.ConfigurationException: For input string: 
 85070591730234615865843651857942052864
 at 
 org.apache.cassandra.dht.Murmur3Partitioner$1.validate(Murmur3Partitioner.java:178)
 at 
 org.apache.cassandra.config.DatabaseDescriptor.applyConfig(DatabaseDescriptor.java:440)
 at 
 org.apache.cassandra.config.DatabaseDescriptor.clinit(DatabaseDescriptor.java:111)
 at 
 org.apache.cassandra.service.CassandraDaemon.setup(CassandraDaemon.java:153)
 at 
 org.apache.cassandra.service.CassandraDaemon.activate(CassandraDaemon.java:471)
 at 
 org.apache.cassandra.service.CassandraDaemon.main(CassandraDaemon.java:560)
 For input string: 85070591730234615865843651857942052864
 Fatal configuration error; unable to start. See log for stacktrace.
 
 I really need to get replication going between 2 nodes. Can someone clue me 
 into why this may be crashing?
 
 Thanks!
 Tim
 
 -- 
 GPG me!!
 
 gpg --keyserver pool.sks-keyservers.net --recv-keys F186197B
 


Re: initial token crashes cassandra

2014-05-17 Thread Tim Dunphy
Hi and thanks for your response.

The puzzling thing is that yes I am using the murmur partition, yet I am
still getting the error I just told you guys about:

[root@beta:/etc/alternatives/cassandrahome] #grep -i partition
conf/cassandra.yaml | grep -v '#'
partitioner: org.apache.cassandra.dht.Murmur3Partitioner

Thanks
Tim


On Sat, May 17, 2014 at 3:23 PM, Colin colpcl...@gmail.com wrote:

 You may have used the old random partitioner token generator.  Use the
 murmur partitioner token generator instead.

 --
 Colin
 320-221-9531


 On May 17, 2014, at 1:15 PM, Tim Dunphy bluethu...@gmail.com wrote:

 Hey all,

  I've set my initial_token in cassandra 2.0.7 using a python script I
 found at the datastax wiki.

 I've set the value like this:

 initial_token: 85070591730234615865843651857942052864

 And cassandra crashes when I try to start it:

 [root@beta:/etc/alternatives/cassandrahome] #./bin/cassandra -f
  INFO 18:14:38,511 Logging initialized
  INFO 18:14:38,560 Loading settings from
 file:/usr/local/apache-cassandra-2.0.7/conf/cassandra.yaml
  INFO 18:14:39,151 Data files directories: [/var/lib/cassandra/data]
  INFO 18:14:39,152 Commit log directory: /var/lib/cassandra/commitlog
  INFO 18:14:39,153 DiskAccessMode 'auto' determined to be mmap,
 indexAccessMode is mmap
  INFO 18:14:39,153 disk_failure_policy is stop
  INFO 18:14:39,153 commit_failure_policy is stop
  INFO 18:14:39,161 Global memtable threshold is enabled at 251MB
  INFO 18:14:39,362 Not using multi-threaded compaction
 ERROR 18:14:39,365 Fatal configuration error
 org.apache.cassandra.exceptions.ConfigurationException: For input string:
 85070591730234615865843651857942052864
 at
 org.apache.cassandra.dht.Murmur3Partitioner$1.validate(Murmur3Partitioner.java:178)
 at
 org.apache.cassandra.config.DatabaseDescriptor.applyConfig(DatabaseDescriptor.java:440)
 at
 org.apache.cassandra.config.DatabaseDescriptor.clinit(DatabaseDescriptor.java:111)
 at
 org.apache.cassandra.service.CassandraDaemon.setup(CassandraDaemon.java:153)
 at
 org.apache.cassandra.service.CassandraDaemon.activate(CassandraDaemon.java:471)
 at
 org.apache.cassandra.service.CassandraDaemon.main(CassandraDaemon.java:560)
 For input string: 85070591730234615865843651857942052864
 Fatal configuration error; unable to start. See log for stacktrace.

 I really need to get replication going between 2 nodes. Can someone clue
 me into why this may be crashing?

 Thanks!
 Tim

 --
 GPG me!!

 gpg --keyserver pool.sks-keyservers.net --recv-keys F186197B




-- 
GPG me!!

gpg --keyserver pool.sks-keyservers.net --recv-keys F186197B


Re: initial token crashes cassandra

2014-05-17 Thread Colin Clark
You probably generated the wrong token type.  Look for a murmur token
generator on the Datastax site.

--
Colin
320-221-9531


On May 17, 2014, at 7:00 PM, Tim Dunphy bluethu...@gmail.com wrote:

Hi and thanks for your response.

The puzzling thing is that yes I am using the murmur partition, yet I am
still getting the error I just told you guys about:

[root@beta:/etc/alternatives/cassandrahome] #grep -i partition
conf/cassandra.yaml | grep -v '#'
partitioner: org.apache.cassandra.dht.Murmur3Partitioner

Thanks
Tim


On Sat, May 17, 2014 at 3:23 PM, Colin colpcl...@gmail.com wrote:

 You may have used the old random partitioner token generator.  Use the
 murmur partitioner token generator instead.

 --
 Colin
 320-221-9531


 On May 17, 2014, at 1:15 PM, Tim Dunphy bluethu...@gmail.com wrote:

 Hey all,

  I've set my initial_token in cassandra 2.0.7 using a python script I
 found at the datastax wiki.

 I've set the value like this:

 initial_token: 85070591730234615865843651857942052864

 And cassandra crashes when I try to start it:

 [root@beta:/etc/alternatives/cassandrahome] #./bin/cassandra -f
  INFO 18:14:38,511 Logging initialized
  INFO 18:14:38,560 Loading settings from
 file:/usr/local/apache-cassandra-2.0.7/conf/cassandra.yaml
  INFO 18:14:39,151 Data files directories: [/var/lib/cassandra/data]
  INFO 18:14:39,152 Commit log directory: /var/lib/cassandra/commitlog
  INFO 18:14:39,153 DiskAccessMode 'auto' determined to be mmap,
 indexAccessMode is mmap
  INFO 18:14:39,153 disk_failure_policy is stop
  INFO 18:14:39,153 commit_failure_policy is stop
  INFO 18:14:39,161 Global memtable threshold is enabled at 251MB
  INFO 18:14:39,362 Not using multi-threaded compaction
 ERROR 18:14:39,365 Fatal configuration error
 org.apache.cassandra.exceptions.ConfigurationException: For input string:
 85070591730234615865843651857942052864
 at
 org.apache.cassandra.dht.Murmur3Partitioner$1.validate(Murmur3Partitioner.java:178)
 at
 org.apache.cassandra.config.DatabaseDescriptor.applyConfig(DatabaseDescriptor.java:440)
 at
 org.apache.cassandra.config.DatabaseDescriptor.clinit(DatabaseDescriptor.java:111)
 at
 org.apache.cassandra.service.CassandraDaemon.setup(CassandraDaemon.java:153)
 at
 org.apache.cassandra.service.CassandraDaemon.activate(CassandraDaemon.java:471)
 at
 org.apache.cassandra.service.CassandraDaemon.main(CassandraDaemon.java:560)
 For input string: 85070591730234615865843651857942052864
 Fatal configuration error; unable to start. See log for stacktrace.

 I really need to get replication going between 2 nodes. Can someone clue
 me into why this may be crashing?

 Thanks!
 Tim

 --
 GPG me!!

 gpg --keyserver pool.sks-keyservers.net --recv-keys F186197B




-- 
GPG me!!

gpg --keyserver pool.sks-keyservers.net --recv-keys F186197B


Re: initial token crashes cassandra

2014-05-17 Thread Dave Brosius
What Colin is saying is that the tool you used to create the token, is 
not creating tokens usable for the Murmur3Partitioner. That tool is 
probably generating tokens for the (original) RandomPartitioner, which 
has a different range.



On 05/17/2014 07:20 PM, Tim Dunphy wrote:

Hi and thanks for your response.

The puzzling thing is that yes I am using the murmur partition, yet I 
am still getting the error I just told you guys about:


[root@beta:/etc/alternatives/cassandrahome] #grep -i partition 
conf/cassandra.yaml | grep -v '#'

partitioner: org.apache.cassandra.dht.Murmur3Partitioner

Thanks
Tim


On Sat, May 17, 2014 at 3:23 PM, Colin colpcl...@gmail.com 
mailto:colpcl...@gmail.com wrote:


You may have used the old random partitioner token generator.  Use
the murmur partitioner token generator instead.

-- 
Colin

320-221-9531 tel:320-221-9531


On May 17, 2014, at 1:15 PM, Tim Dunphy bluethu...@gmail.com
mailto:bluethu...@gmail.com wrote:


Hey all,

 I've set my initial_token in cassandra 2.0.7 using a python
script I found at the datastax wiki.

I've set the value like this:

initial_token: 85070591730234615865843651857942052864

And cassandra crashes when I try to start it:

[root@beta:/etc/alternatives/cassandrahome] #./bin/cassandra -f
 INFO 18:14:38,511 Logging initialized
 INFO 18:14:38,560 Loading settings from
file:/usr/local/apache-cassandra-2.0.7/conf/cassandra.yaml
 INFO 18:14:39,151 Data files directories: [/var/lib/cassandra/data]
 INFO 18:14:39,152 Commit log directory: /var/lib/cassandra/commitlog
 INFO 18:14:39,153 DiskAccessMode 'auto' determined to be mmap,
indexAccessMode is mmap
 INFO 18:14:39,153 disk_failure_policy is stop
 INFO 18:14:39,153 commit_failure_policy is stop
 INFO 18:14:39,161 Global memtable threshold is enabled at 251MB
 INFO 18:14:39,362 Not using multi-threaded compaction
ERROR 18:14:39,365 Fatal configuration error
org.apache.cassandra.exceptions.ConfigurationException: For input
string: 85070591730234615865843651857942052864
at

org.apache.cassandra.dht.Murmur3Partitioner$1.validate(Murmur3Partitioner.java:178)
at

org.apache.cassandra.config.DatabaseDescriptor.applyConfig(DatabaseDescriptor.java:440)
at

org.apache.cassandra.config.DatabaseDescriptor.clinit(DatabaseDescriptor.java:111)
at
org.apache.cassandra.service.CassandraDaemon.setup(CassandraDaemon.java:153)
at

org.apache.cassandra.service.CassandraDaemon.activate(CassandraDaemon.java:471)
at
org.apache.cassandra.service.CassandraDaemon.main(CassandraDaemon.java:560)
For input string: 85070591730234615865843651857942052864
Fatal configuration error; unable to start. See log for stacktrace.

I really need to get replication going between 2 nodes. Can
someone clue me into why this may be crashing?

Thanks!
Tim

-- 
GPG me!!


gpg --keyserver pool.sks-keyservers.net
http://pool.sks-keyservers.net --recv-keys F186197B





--
GPG me!!

gpg --keyserver pool.sks-keyservers.net 
http://pool.sks-keyservers.net --recv-keys F186197B






Re: initial token crashes cassandra

2014-05-17 Thread Tim Dunphy

 You probably generated the wrong token type.  Look for a murmur token
 generator on the Datastax site.

What Colin is saying is that the tool you used to create the token, is not
 creating tokens usable for the Murmur3Partitioner. That tool is probably
 generating tokens for the (original) RandomPartitioner, which has a
 different range.


Thanks guys for your input. And I apologize for reading  Colin's initial
response too quickly which lets me know that I was probably using the wrong
token generator for the wrong partition type. That of course was the case.
So what I've done is use this token generator form the datastax website:

python -c 'print [str(((2**64 / number_of_tokens) * i) - 2**63) for i
in range(number_of_tokens)]


That algorithm generated a token I could use to start Cassandra on my
second node.


However at this stage I have both nodes running and I believe their
gossiping if I understand what I see here correctly:


 INFO 02:44:13,823 No gossip backlog; proceeding


However I've setup web pages for each of the two web servers that are
running Cassandra. And it looks like the seed node with all the data
is rendering correctly. But the node that's downstream from the seed
node is not receiving any of its data despite the message that I've
just shown you.


And if I go to the seed node and do a describe keyspaces I see the
keyspace that drives the website listed. It's called 'joke_fire1'


cqlsh describe keyspaces;

system  joke_fire1  system_traces

And if I go to the node that's downstream from the seed node and run
the same command:


cqlsh describe keyspaces;

system  system_traces


I don't see the important keyspace that runs the site.


I have the seed node's IP listed in 'seeds' in the cassandra.yaml on
the downstream node. So I'm not really sure why its' not receiving the
seed's data. If there's some command I need to run to flush the system
or something like that.


And if I do a nodetool ring command on the first (seed) host I don't
see the IP of the downstream node listed:








[root@beta-new:~] #nodetool ring | head -10

Note: Ownership information does not include topology; for complete
information, specify a keyspace


Datacenter: datacenter1

==

Address RackStatus State   LoadOwns
Token


10.10.1.94  rack1   Up Normal  150.64 KB   100.00%
-9173731940639284976

10.10.1.94  rack1   Up Normal  150.64 KB   100.00%
-9070607847117718988

10.10.1.94  rack1   kUp Normal  150.64 KB   100.00%
 -9060190512633067546

10.10.1.94  rack1   Up Normal  150.64 KB   100.00%
-8935690644016753923


And if I look on the downstream node and run nodetool ring I see only
the IP of the downstream node and not the seed listed:









[root@beta:/var/lib/cassandra] #nodetool ring | head -15


Datacenter: datacenter1

==

Address  RackStatus State   LoadOwns
 Token


10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-9223372036854775808

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-9151314442816847873

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-9079256848778919937

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-9007199254740992001

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-8935141660703064065

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-886308405136129

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-8791026472627208193

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-8718968878589280257

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-8646911284551352321

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-8574853690513424385


Yet in my seeds entry in cassandra.yaml I have the correct IP of my
seed node listed:


seed_provider:

- class_name: org.apache.cassandra.locator.SimpleSeedProvider

  # seeds is actually a comma-delimited list of addresses.

  - seeds: 10.10.1.94


So I'm just wondering what I'm missing in trying to get these two
nodes to communicate via gossip at this point.


Thanks!

Tim








On Sat, May 17, 2014 at 8:54 PM, Dave Brosius dbros...@mebigfatguy.comwrote:

  What Colin is saying is that the tool you used to create the token, is
 not creating tokens usable for the Murmur3Partitioner. That tool is
 probably generating tokens for the (original) RandomPartitioner, which has
 a different range.



 On 05/17/2014 07:20 PM, Tim Dunphy wrote:

 Hi and thanks for your response.

  The puzzling thing is that yes I am using the murmur partition, yet I am
 still getting the error I just told you guys about:

   [root@beta:/etc/alternatives/cassandrahome] #grep -i partition
 conf/cassandra.yaml | grep -v '#'
 partitioner: org.apache.cassandra.dht.Murmur3Partitioner

  Thanks
 Tim


 On Sat, May 17, 2014 at 3:23 PM, 

Re: initial token crashes cassandra

2014-05-17 Thread Colin Clark
Looks like you may have put the token next to num-tokens property in the
yaml file for one node.  I would double check the yaml's to make sure the
tokens are setup correctly and that the ip addresses are associated with
the right entries as well.

Compare them to a fresh download if possible to see what you've changed.

--
Colin
320-221-9531


On May 17, 2014, at 10:29 PM, Tim Dunphy bluethu...@gmail.com wrote:

You probably generated the wrong token type.  Look for a murmur token
 generator on the Datastax site.

What Colin is saying is that the tool you used to create the token, is not
 creating tokens usable for the Murmur3Partitioner. That tool is probably
 generating tokens for the (original) RandomPartitioner, which has a
 different range.


Thanks guys for your input. And I apologize for reading  Colin's initial
response too quickly which lets me know that I was probably using the wrong
token generator for the wrong partition type. That of course was the case.
So what I've done is use this token generator form the datastax website:

python -c 'print [str(((2**64 / number_of_tokens) * i) - 2**63) for i
in range(number_of_tokens)]


That algorithm generated a token I could use to start Cassandra on my
second node.


However at this stage I have both nodes running and I believe their
gossiping if I understand what I see here correctly:


 INFO 02:44:13,823 No gossip backlog; proceeding


However I've setup web pages for each of the two web servers that are
running Cassandra. And it looks like the seed node with all the data
is rendering correctly. But the node that's downstream from the seed
node is not receiving any of its data despite the message that I've
just shown you.


And if I go to the seed node and do a describe keyspaces I see the
keyspace that drives the website listed. It's called 'joke_fire1'


cqlsh describe keyspaces;

system  joke_fire1  system_traces

And if I go to the node that's downstream from the seed node and run
the same command:


cqlsh describe keyspaces;

system  system_traces


I don't see the important keyspace that runs the site.


I have the seed node's IP listed in 'seeds' in the cassandra.yaml on
the downstream node. So I'm not really sure why its' not receiving the
seed's data. If there's some command I need to run to flush the system
or something like that.


And if I do a nodetool ring command on the first (seed) host I don't
see the IP of the downstream node listed:







[root@beta-new:~] #nodetool ring | head -10

Note: Ownership information does not include topology; for complete
information, specify a keyspace


Datacenter: datacenter1

==

Address RackStatus State   LoadOwns
Token


10.10.1.94  rack1   Up Normal  150.64 KB   100.00%
-9173731940639284976

10.10.1.94  rack1   Up Normal  150.64 KB   100.00%
-9070607847117718988

10.10.1.94  rack1   kUp Normal  150.64 KB   100.00%
 -9060190512633067546

10.10.1.94  rack1   Up Normal  150.64 KB   100.00%
-8935690644016753923


And if I look on the downstream node and run nodetool ring I see only
the IP of the downstream node and not the seed listed:









[root@beta:/var/lib/cassandra] #nodetool ring | head -15


Datacenter: datacenter1

==

Address  RackStatus State   LoadOwns
 Token


10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-9223372036854775808

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-9151314442816847873

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-9079256848778919937

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-9007199254740992001

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-8935141660703064065

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-886308405136129

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-8791026472627208193

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-8718968878589280257

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-8646911284551352321

10.10.1.98  rack1   Up Normal  91.06 KB99.99%
-8574853690513424385


Yet in my seeds entry in cassandra.yaml I have the correct IP of my
seed node listed:


seed_provider:

- class_name: org.apache.cassandra.locator.SimpleSeedProvider

  # seeds is actually a comma-delimited list of addresses.

  - seeds: 10.10.1.94


So I'm just wondering what I'm missing in trying to get these two
nodes to communicate via gossip at this point.


Thanks!

Tim








On Sat, May 17, 2014 at 8:54 PM, Dave Brosius dbros...@mebigfatguy.comwrote:

  What Colin is saying is that the tool you used to create the token, is
 not creating tokens usable for the Murmur3Partitioner. That tool is
 probably generating tokens for the (original) RandomPartitioner, which has
 a different range.



 On 

Re: initial token crashes cassandra

2014-05-17 Thread Tim Dunphy
Hey Colin,

Looks like you may have put the token next to num-tokens property in the
 yaml file for one node.  I would double check the yaml's to make sure the
 tokens are setup correctly and that the ip addresses are associated with
 the right entries as well.
 Compare them to a fresh download if possible to see what you've changed.


 Thanks! I did that and now things are working perfectly:

[root@beta-new:~] #nodetool status

Datacenter: datacenter1

===

Status=Up/Down

|/ State=Normal/Leaving/Joining/Moving

--  Address Load   Tokens  Owns   Host ID
Rack

UN  10.10.1.94  164.39 KB  256 49.4%
fd2f76ae-8dcf-4e93-a37f-bf1e9088696e  rack1

UN  10.10.1.98 99.08 KB   256 50.6%
f2a48fc7-a362-43f5-9061-4bb3739fdeaf  rack1


Thanks again for your help!


Tim


On Sun, May 18, 2014 at 12:35 AM, Colin Clark co...@clark.ws wrote:

 Looks like you may have put the token next to num-tokens property in the
 yaml file for one node.  I would double check the yaml's to make sure the
 tokens are setup correctly and that the ip addresses are associated with
 the right entries as well.

 Compare them to a fresh download if possible to see what you've changed.

 --
 Colin
 320-221-9531


 On May 17, 2014, at 10:29 PM, Tim Dunphy bluethu...@gmail.com wrote:

 You probably generated the wrong token type.  Look for a murmur token
 generator on the Datastax site.

 What Colin is saying is that the tool you used to create the token, is not
 creating tokens usable for the Murmur3Partitioner. That tool is probably
 generating tokens for the (original) RandomPartitioner, which has a
 different range.


 Thanks guys for your input. And I apologize for reading  Colin's initial
 response too quickly which lets me know that I was probably using the wrong
 token generator for the wrong partition type. That of course was the case.
 So what I've done is use this token generator form the datastax website:

 python -c 'print [str(((2**64 / number_of_tokens) * i) - 2**63) for i in 
 range(number_of_tokens)]


 That algorithm generated a token I could use to start Cassandra on my second 
 node.


 However at this stage I have both nodes running and I believe their gossiping 
 if I understand what I see here correctly:


  INFO 02:44:13,823 No gossip backlog; proceeding


 However I've setup web pages for each of the two web servers that are running 
 Cassandra. And it looks like the seed node with all the data is rendering 
 correctly. But the node that's downstream from the seed node is not receiving 
 any of its data despite the message that I've just shown you.


 And if I go to the seed node and do a describe keyspaces I see the keyspace 
 that drives the website listed. It's called 'joke_fire1'


 cqlsh describe keyspaces;

 system  joke_fire1  system_traces

 And if I go to the node that's downstream from the seed node and run the same 
 command:


 cqlsh describe keyspaces;

 system  system_traces


 I don't see the important keyspace that runs the site.


 I have the seed node's IP listed in 'seeds' in the cassandra.yaml on the 
 downstream node. So I'm not really sure why its' not receiving the seed's 
 data. If there's some command I need to run to flush the system or something 
 like that.


 And if I do a nodetool ring command on the first (seed) host I don't see the 
 IP of the downstream node listed:






 [root@beta-new:~] #nodetool ring | head -10

 Note: Ownership information does not include topology; for complete 
 information, specify a keyspace


 Datacenter: datacenter1

 ==

 Address RackStatus State   LoadOwns   
  Token


 10.10.1.94  rack1   Up Normal  150.64 KB   100.00% 
 -9173731940639284976

 10.10.1.94  rack1   Up Normal  150.64 KB   100.00% 
 -9070607847117718988

 10.10.1.94  rack1   kUp Normal  150.64 KB   100.00% 
 -9060190512633067546

 10.10.1.94  rack1   Up Normal  150.64 KB   100.00% 
 -8935690644016753923


 And if I look on the downstream node and run nodetool ring I see only the IP 
 of the downstream node and not the seed listed:









 [root@beta:/var/lib/cassandra] #nodetool ring | head -15


 Datacenter: datacenter1

 ==

 Address  RackStatus State   LoadOwns
 Token


 10.10.1.98  rack1   Up Normal  91.06 KB99.99%  
 -9223372036854775808

 10.10.1.98  rack1   Up Normal  91.06 KB99.99%  
 -9151314442816847873

 10.10.1.98  rack1   Up Normal  91.06 KB99.99%  
 -9079256848778919937

 10.10.1.98  rack1   Up Normal  91.06 KB99.99%  
 -9007199254740992001

 10.10.1.98  rack1   Up Normal  91.06 KB99.99%  
 -8935141660703064065

 10.10.1.98  rack1   Up Normal  91.06 KB99.99%