[jira] [Commented] (CASSANDRA-2356) make the debian package never start by default

2012-11-20 Thread Tamar Fraenkel (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-2356?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13501723#comment-13501723
 ] 

Tamar Fraenkel commented on CASSANDRA-2356:
---

I would also like that fix. The auto restart is a pain after upgrading, since I 
need to merge my cassandra.yaml and cassander-env.sh, and don't want the node 
to restart on its own.

> make the debian package never start by default
> --
>
> Key: CASSANDRA-2356
> URL: https://issues.apache.org/jira/browse/CASSANDRA-2356
> Project: Cassandra
>  Issue Type: Improvement
>  Components: Packaging
>Reporter: Jeremy Hanna
>Priority: Minor
>  Labels: debian, packaging
> Attachments: 2356.txt
>
>
> Currently the debian package that installs cassandra starts cassandra by 
> default.  It sounds like that is a standard debian packaging convention.  
> However, if you want to bootstrap a new node and want to configure it before 
> it creates any sort of state information, it's a pain.  I would think that 
> the common use case would be to have it install all of the init scripts and 
> such but *not* have it start up by default.  That way an admin can configure 
> cassandra with seed, token, host, etc. information and then start it.  That 
> makes it easier to programmatically do this as well - have chef/puppet 
> install cassandra, do some configuration, then do the service start.
> With the current setup, it sounds like cassandra creates state on startup 
> that has to be cleaned before a new configuration can take effect.  So the 
> process of installing turns into:
> * install debian package
> * shutdown cassandra
> * clean out state (data/log dirs)
> * configure cassandra
> * start cassandra
> That seems suboptimal for the default case, especially when trying to 
> automate new nodes being bootstrapped.
> Another case might be when a downed node comes back up and starts by default 
> and tries to claim a token that has already been claimed by another newly 
> bootstrapped node.  Rob is more familiar with that case so I'll let him 
> explain it in the comments.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (CASSANDRA-4446) nodetool drain sometimes doesn't mark commitlog fully flushed

2012-11-20 Thread Tamar Fraenkel (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-4446?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13501722#comment-13501722
 ] 

Tamar Fraenkel commented on CASSANDRA-4446:
---

I had the same experience, when I upgraded my cluster from 1.0.9 to 1.0.11. I 
ran drain before the upgrade, upgrade on the node finished and node restarted 
at 2012-11-20 10:20:58, but then I see in the logs reply of commit log:
{quote} 
 INFO [main] 2012-11-20 09:41:13,918 CommitLog.java (line 172) Replaying 
/raid0/cassandra/commitlog/CommitLog-1353402218337.log
 INFO [main] 2012-11-20 09:41:20,360 CommitLog.java (line 179) Log replay 
complete, 0 replayed mutations
 INFO [main] 2012-11-20 10:11:35,635 CommitLog.java (line 167) No commitlog 
files found; skipping replay
 INFO [main] 2012-11-20 10:21:11,631 CommitLog.java (line 172) Replaying 
/raid0/cassandra/commitlog/CommitLog-1353404473899.log
 INFO [main] 2012-11-20 10:21:18,119 CommitLog.java (line 179) Log replay 
complete, 6413 replayed mutations
 INFO [main] 2012-11-20 10:55:46,435 CommitLog.java (line 172) Replaying 
/raid0/cassandra/commitlog/CommitLog-1353406871619.log
 INFO [main] 2012-11-20 10:55:54,139 CommitLog.java (line 179) Log replay 
complete, 3 replayed mutations
{quote} 
This caused over increment of counters


> nodetool drain sometimes doesn't mark commitlog fully flushed
> -
>
> Key: CASSANDRA-4446
> URL: https://issues.apache.org/jira/browse/CASSANDRA-4446
> Project: Cassandra
>  Issue Type: Bug
>Affects Versions: 1.0.10
> Environment: ubuntu 10.04 64bit
> Linux HOSTNAME 2.6.32-345-ec2 #48-Ubuntu SMP Wed May 2 19:29:55 UTC 2012 
> x86_64 GNU/Linux
> sun JVM
> cassandra 1.0.10 installed from apache deb
>Reporter: Robert Coli
> Attachments: 
> cassandra.1.0.10.replaying.log.after.exception.during.drain.txt
>
>
> I recently wiped a customer's QA cluster. I drained each node and verified 
> that they were drained. When I restarted the nodes, I saw the commitlog 
> replay create a memtable and then flush it. I have attached a sanitized log 
> snippet from a representative node at the time. 
> It appears to show the following :
> 1) Drain begins
> 2) Drain triggers flush
> 3) Flush triggers compaction
> 4) StorageService logs DRAINED message
> 5) compaction thread excepts
> 6) on restart, same CF creates a memtable
> 7) and then flushes it [1]
> The columnfamily involved in the replay in 7) is the CF for which the 
> compaction thread excepted in 5). This seems to suggest a timing issue 
> whereby the exception in 5) prevents the flush in 3) from marking all the 
> segments flushed, causing them to replay after restart.
> In case it might be relevant, I did an online change of compaction strategy 
> from Leveled to SizeTiered during the uptime period preceding this drain.
> [1] Isn't commitlog replay not supposed to automatically trigger a flush in 
> modern cassandra?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Comment Edited] (CASSANDRA-3736) -Dreplace_token leaves old node (IP) in the gossip with the token.

2012-11-13 Thread Tamar Fraenkel (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-3736?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13496136#comment-13496136
 ] 

Tamar Fraenkel edited comment on CASSANDRA-3736 at 11/13/12 12:32 PM:
--

I have the same issue with Cassandra 1.0.11 (used DataStax AMI).
I thought it was supposed to be solved already.
I see those messages on the node that was started using 
-Dcassandra.replace_token=.

>From time to time I also see 
{color:blue} 
 INFO [GossipTasks:1] 2012-11-13 12:26:38,195 Gossiper.java (line 818) 
InetAddress / is now dead.
 INFO [GossipTasks:1] 2012-11-13 12:26:58,203 Gossiper.java (line 632) 
FatClient / has been silent for 3ms, removing from gossip
 INFO [GossipStage:1] 2012-11-13 12:27:59,210 Gossiper.java (line 838) Node 
/ is now part of the cluster
 INFO [GossipStage:1] 2012-11-13 12:27:59,210 Gossiper.java (line 804) 
InetAddress / is now UP
 INFO [GossipStage:1] 2012-11-13 12:27:59,210 StorageService.java (line 1017) 
Nodes / and / have the same token 
113427455640312821154458202477256070484.  Ignoring /

{color}



  was (Author: tamarfraenkel):
I have the same issue with Cassandra 1.0.11 (used DataStax AMI).
I thought it was supposed to be solved already.
I see those messages on the node that was started using 
-Dcassandra.replace_token=.


  
> -Dreplace_token leaves old node (IP) in the gossip with the token.
> --
>
> Key: CASSANDRA-3736
> URL: https://issues.apache.org/jira/browse/CASSANDRA-3736
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
>Affects Versions: 1.0.0
>Reporter: Jackson Chung
>Assignee: Vijay
> Fix For: 1.0.8, 1.1.0
>
> Attachments: 0001-CASSANDRA-3736.patch
>
>
> https://issues.apache.org/jira/browse/CASSANDRA-957 introduce a 
> -Dreplace_token,
> however, the replaced IP keeps on showing up in the Gossiper when starting 
> the replacement node:
> {noformat}
>  INFO [Thread-2] 2012-01-12 23:59:35,162 CassandraDaemon.java (line 213) 
> Listening for thrift clients...
>  INFO [GossipStage:1] 2012-01-12 23:59:35,173 Gossiper.java (line 836) Node 
> /50.56.59.68 has restarted, now UP
>  INFO [GossipStage:1] 2012-01-12 23:59:35,174 Gossiper.java (line 804) 
> InetAddress /50.56.59.68 is now UP
>  INFO [GossipStage:1] 2012-01-12 23:59:35,175 StorageService.java (line 988) 
> Node /50.56.59.68 state jump to normal
>  INFO [GossipStage:1] 2012-01-12 23:59:35,176 Gossiper.java (line 836) Node 
> /50.56.58.55 has restarted, now UP
>  INFO [GossipStage:1] 2012-01-12 23:59:35,176 Gossiper.java (line 804) 
> InetAddress /50.56.58.55 is now UP
>  INFO [GossipStage:1] 2012-01-12 23:59:35,177 StorageService.java (line 1016) 
> Nodes /50.56.58.55 and action-quick2/50.56.31.186 have the same token 
> 85070591730234615865843651857942052864.  Ignoring /50.56.58.55
>  INFO [GossipTasks:1] 2012-01-12 23:59:45,048 Gossiper.java (line 818) 
> InetAddress /50.56.58.55 is now dead.
>  INFO [GossipTasks:1] 2012-01-13 00:00:06,062 Gossiper.java (line 632) 
> FatClient /50.56.58.55 has been silent for 3ms, removing from gossip
>  INFO [GossipStage:1] 2012-01-13 00:01:06,320 Gossiper.java (line 838) Node 
> /50.56.58.55 is now part of the cluster
>  INFO [GossipStage:1] 2012-01-13 00:01:06,320 Gossiper.java (line 804) 
> InetAddress /50.56.58.55 is now UP
>  INFO [GossipStage:1] 2012-01-13 00:01:06,321 StorageService.java (line 1016) 
> Nodes /50.56.58.55 and action-quick2/50.56.31.186 have the same token 
> 85070591730234615865843651857942052864.  Ignoring /50.56.58.55
>  INFO [GossipTasks:1] 2012-01-13 00:01:16,106 Gossiper.java (line 818) 
> InetAddress /50.56.58.55 is now dead.
>  INFO [GossipTasks:1] 2012-01-13 00:01:37,121 Gossiper.java (line 632) 
> FatClient /50.56.58.55 has been silent for 3ms, removing from gossip
>  INFO [GossipStage:1] 2012-01-13 00:02:37,352 Gossiper.java (line 838) Node 
> /50.56.58.55 is now part of the cluster
>  INFO [GossipStage:1] 2012-01-13 00:02:37,353 Gossiper.java (line 804) 
> InetAddress /50.56.58.55 is now UP
>  INFO [GossipStage:1] 2012-01-13 00:02:37,353 StorageService.java (line 1016) 
> Nodes /50.56.58.55 and action-quick2/50.56.31.186 have the same token 
> 85070591730234615865843651857942052864.  Ignoring /50.56.58.55
>  INFO [GossipTasks:1] 2012-01-13 00:02:47,158 Gossiper.java (line 818) 
> InetAddress /50.56.58.55 is now dead.
>  INFO [GossipStage:1] 2012-01-13 00:02:50,162 Gossiper.java (line 818) 
> InetAddress /50.56.58.55 is now dead.
>  INFO [GossipStage:1] 2012-01-13 00:02:50,163 StorageService.java (line 1156) 
> Removing token 122029383590318827259508597176866581733 for /50.56.58.55
> {noformat}
> in the above, /50.56.58.55 was the replaced IP.
> tried adding the "Gossiper.instance.r

[jira] [Commented] (CASSANDRA-3736) -Dreplace_token leaves old node (IP) in the gossip with the token.

2012-11-13 Thread Tamar Fraenkel (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-3736?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13496136#comment-13496136
 ] 

Tamar Fraenkel commented on CASSANDRA-3736:
---

I have the same issue with Cassandra 1.0.11 (used DataStax AMI).
I thought it was supposed to be solved already.
I see those messages on the node that was started using 
-Dcassandra.replace_token=.



> -Dreplace_token leaves old node (IP) in the gossip with the token.
> --
>
> Key: CASSANDRA-3736
> URL: https://issues.apache.org/jira/browse/CASSANDRA-3736
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
>Affects Versions: 1.0.0
>Reporter: Jackson Chung
>Assignee: Vijay
> Fix For: 1.0.8, 1.1.0
>
> Attachments: 0001-CASSANDRA-3736.patch
>
>
> https://issues.apache.org/jira/browse/CASSANDRA-957 introduce a 
> -Dreplace_token,
> however, the replaced IP keeps on showing up in the Gossiper when starting 
> the replacement node:
> {noformat}
>  INFO [Thread-2] 2012-01-12 23:59:35,162 CassandraDaemon.java (line 213) 
> Listening for thrift clients...
>  INFO [GossipStage:1] 2012-01-12 23:59:35,173 Gossiper.java (line 836) Node 
> /50.56.59.68 has restarted, now UP
>  INFO [GossipStage:1] 2012-01-12 23:59:35,174 Gossiper.java (line 804) 
> InetAddress /50.56.59.68 is now UP
>  INFO [GossipStage:1] 2012-01-12 23:59:35,175 StorageService.java (line 988) 
> Node /50.56.59.68 state jump to normal
>  INFO [GossipStage:1] 2012-01-12 23:59:35,176 Gossiper.java (line 836) Node 
> /50.56.58.55 has restarted, now UP
>  INFO [GossipStage:1] 2012-01-12 23:59:35,176 Gossiper.java (line 804) 
> InetAddress /50.56.58.55 is now UP
>  INFO [GossipStage:1] 2012-01-12 23:59:35,177 StorageService.java (line 1016) 
> Nodes /50.56.58.55 and action-quick2/50.56.31.186 have the same token 
> 85070591730234615865843651857942052864.  Ignoring /50.56.58.55
>  INFO [GossipTasks:1] 2012-01-12 23:59:45,048 Gossiper.java (line 818) 
> InetAddress /50.56.58.55 is now dead.
>  INFO [GossipTasks:1] 2012-01-13 00:00:06,062 Gossiper.java (line 632) 
> FatClient /50.56.58.55 has been silent for 3ms, removing from gossip
>  INFO [GossipStage:1] 2012-01-13 00:01:06,320 Gossiper.java (line 838) Node 
> /50.56.58.55 is now part of the cluster
>  INFO [GossipStage:1] 2012-01-13 00:01:06,320 Gossiper.java (line 804) 
> InetAddress /50.56.58.55 is now UP
>  INFO [GossipStage:1] 2012-01-13 00:01:06,321 StorageService.java (line 1016) 
> Nodes /50.56.58.55 and action-quick2/50.56.31.186 have the same token 
> 85070591730234615865843651857942052864.  Ignoring /50.56.58.55
>  INFO [GossipTasks:1] 2012-01-13 00:01:16,106 Gossiper.java (line 818) 
> InetAddress /50.56.58.55 is now dead.
>  INFO [GossipTasks:1] 2012-01-13 00:01:37,121 Gossiper.java (line 632) 
> FatClient /50.56.58.55 has been silent for 3ms, removing from gossip
>  INFO [GossipStage:1] 2012-01-13 00:02:37,352 Gossiper.java (line 838) Node 
> /50.56.58.55 is now part of the cluster
>  INFO [GossipStage:1] 2012-01-13 00:02:37,353 Gossiper.java (line 804) 
> InetAddress /50.56.58.55 is now UP
>  INFO [GossipStage:1] 2012-01-13 00:02:37,353 StorageService.java (line 1016) 
> Nodes /50.56.58.55 and action-quick2/50.56.31.186 have the same token 
> 85070591730234615865843651857942052864.  Ignoring /50.56.58.55
>  INFO [GossipTasks:1] 2012-01-13 00:02:47,158 Gossiper.java (line 818) 
> InetAddress /50.56.58.55 is now dead.
>  INFO [GossipStage:1] 2012-01-13 00:02:50,162 Gossiper.java (line 818) 
> InetAddress /50.56.58.55 is now dead.
>  INFO [GossipStage:1] 2012-01-13 00:02:50,163 StorageService.java (line 1156) 
> Removing token 122029383590318827259508597176866581733 for /50.56.58.55
> {noformat}
> in the above, /50.56.58.55 was the replaced IP.
> tried adding the "Gossiper.instance.removeEndpoint(endpoint);" in the 
> StorageService.java where the message 'Nodes %s and %s have the same token 
> %s.  Ignoring %s",' seems only have fixed this temporary. Here is a ring 
> output:
> {noformat}
> riptano@action-quick:~/work/cassandra$ ./bin/nodetool -h localhost ring
> Address DC  RackStatus State   LoadOwns   
>  Token   
>   
>  85070591730234615865843651857942052864  
> 50.56.59.68 datacenter1 rack1   Up Normal  6.67 KB 85.56% 
>  60502102442797279294142560823234402248  
> 50.56.31.186datacenter1 rack1   Up Normal  11.12 KB14.44% 
>  85070591730234615865843651857942052864 
> {noformat}
> gossipinfo:
> {noformat}
> $ ./bin/nodetool -h localhost gossipinfo
> /50.56.58.55
>   LOAD:6835.0
>   SCHEMA:--1000--
>   RPC_AD

[jira] [Created] (CASSANDRA-4742) key cache exist for dropped column families

2012-10-01 Thread Tamar Fraenkel (JIRA)
Tamar Fraenkel created CASSANDRA-4742:
-

 Summary: key cache exist for dropped column families
 Key: CASSANDRA-4742
 URL: https://issues.apache.org/jira/browse/CASSANDRA-4742
 Project: Cassandra
  Issue Type: Bug
  Components: Core
Affects Versions: 1.0.8
 Environment: DataStax AMI on Amazon
Reporter: Tamar Fraenkel


I have couple of dropped column families, and in the JMX console I don't see 
them under org.apache.cassandra.db.ColumnFamilies, BUT I do see them under 
org.apache.cassandra.db.Caches, and the cache is not empty!

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira