[jira] [Created] (CASSANDRA-10464) "nodetool compactionhistory" output should be sorted on compacted_at column and the timestamp shown in human readable format

2015-10-06 Thread Wei Deng (JIRA)
Wei Deng created CASSANDRA-10464:


 Summary: "nodetool compactionhistory" output should be sorted on 
compacted_at column and the timestamp shown in human readable format
 Key: CASSANDRA-10464
 URL: https://issues.apache.org/jira/browse/CASSANDRA-10464
 Project: Cassandra
  Issue Type: Improvement
Reporter: Wei Deng
Priority: Minor


"nodetool compactionhistory" (introduced in CASSANDRA-5078) is a useful tool 
for Cassandra DBAs. However, the current output limits its usefulness without 
some additional parsing.

We should improve it in the following two areas:
1. The output should be sorted on the compacted_at column, so that the most 
recently finished compaction will show up last (which is what the DBAs would 
expect);
2. The compacted_at column should be printed in human-readable timestamp.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10445) Cassandra-stress throws max frame size error when SSL certification is enabled

2015-10-06 Thread Philip Thompson (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10445?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945157#comment-14945157
 ] 

Philip Thompson commented on CASSANDRA-10445:
-

If you compile and use the cassandra stress from trunk against your 2.1 
cluster, does this still reproduce?

> Cassandra-stress throws max frame size error when SSL certification is enabled
> --
>
> Key: CASSANDRA-10445
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10445
> Project: Cassandra
>  Issue Type: Bug
>Reporter: Sam Goldberg
> Fix For: 2.1.x
>
>
> Running cassandra-stress when SSL is enabled gives the following error and 
> does not finish executing:
> {quote}
> cassandra-stress write n=100
> Exception in thread "main" java.lang.RuntimeException: 
> org.apache.thrift.transport.TTransportException: Frame size (352518912) 
> larger than max length (15728640)!
> at 
> org.apache.cassandra.stress.settings.StressSettings.getRawThriftClient(StressSettings.java:144)
> at 
> org.apache.cassandra.stress.settings.StressSettings.getRawThriftClient(StressSettings.java:110)
> at 
> org.apache.cassandra.stress.settings.SettingsSchema.createKeySpacesThrift(SettingsSchema.java:111)
> at 
> org.apache.cassandra.stress.settings.SettingsSchema.createKeySpaces(SettingsSchema.java:59)
> at 
> org.apache.cassandra.stress.settings.StressSettings.maybeCreateKeyspaces(StressSettings.java:205)
> at org.apache.cassandra.stress.StressAction.run(StressAction.java:55)
> at org.apache.cassandra.stress.Stress.main(Stress.java:109)
> {quote}
> I was able to reproduce this issue consistently via the following steps:
> 1) Spin up 3 node cassandra cluster running 2.1.8
> 2) Perform cassandra-stress write n=100
> 3) Everything works!
> 4) Generate keystore and truststore for each node in the cluster and 
> distribute appropriately 
> 5) Modify cassandra.yaml on each node to enable SSL:
> client_encryption_options:
> enabled: true
> keystore: /
> # require_client_auth: false
> # Set trustore and truststore_password if require_client_auth is true
> truststore:  /
> truststore_password: 
> # More advanced defaults below:
> protocol: ssl
> 6) Restart each node.
> 7) Perform cassandra-stress write n=100
> 8) Get Frame Size error, cassandra-stress fails
> This may be related to CASSANDRA-9325.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10445) Cassandra-stress throws max frame size error when SSL certification is enabled

2015-10-06 Thread Philip Thompson (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Philip Thompson updated CASSANDRA-10445:

Fix Version/s: 2.1.x

> Cassandra-stress throws max frame size error when SSL certification is enabled
> --
>
> Key: CASSANDRA-10445
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10445
> Project: Cassandra
>  Issue Type: Bug
>Reporter: Sam Goldberg
> Fix For: 2.1.x
>
>
> Running cassandra-stress when SSL is enabled gives the following error and 
> does not finish executing:
> {quote}
> cassandra-stress write n=100
> Exception in thread "main" java.lang.RuntimeException: 
> org.apache.thrift.transport.TTransportException: Frame size (352518912) 
> larger than max length (15728640)!
> at 
> org.apache.cassandra.stress.settings.StressSettings.getRawThriftClient(StressSettings.java:144)
> at 
> org.apache.cassandra.stress.settings.StressSettings.getRawThriftClient(StressSettings.java:110)
> at 
> org.apache.cassandra.stress.settings.SettingsSchema.createKeySpacesThrift(SettingsSchema.java:111)
> at 
> org.apache.cassandra.stress.settings.SettingsSchema.createKeySpaces(SettingsSchema.java:59)
> at 
> org.apache.cassandra.stress.settings.StressSettings.maybeCreateKeyspaces(StressSettings.java:205)
> at org.apache.cassandra.stress.StressAction.run(StressAction.java:55)
> at org.apache.cassandra.stress.Stress.main(Stress.java:109)
> {quote}
> I was able to reproduce this issue consistently via the following steps:
> 1) Spin up 3 node cassandra cluster running 2.1.8
> 2) Perform cassandra-stress write n=100
> 3) Everything works!
> 4) Generate keystore and truststore for each node in the cluster and 
> distribute appropriately 
> 5) Modify cassandra.yaml on each node to enable SSL:
> client_encryption_options:
> enabled: true
> keystore: /
> # require_client_auth: false
> # Set trustore and truststore_password if require_client_auth is true
> truststore:  /
> truststore_password: 
> # More advanced defaults below:
> protocol: ssl
> 6) Restart each node.
> 7) Perform cassandra-stress write n=100
> 8) Get Frame Size error, cassandra-stress fails
> This may be related to CASSANDRA-9325.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10412) Could not initialize class org.apache.cassandra.config.DatabaseDescriptor

2015-10-06 Thread Paulo Motta (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10412?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paulo Motta updated CASSANDRA-10412:

Reviewer: Paulo Motta

> Could not initialize class org.apache.cassandra.config.DatabaseDescriptor
> -
>
> Key: CASSANDRA-10412
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10412
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
> Environment: OS: Windows 2012 R2
> Java JRE: 1.8.0_51
>Reporter: Eric Simmons
>Assignee: Carl Yeksigian
>Priority: Minor
> Attachments: cassandra-env.ps1, cassandra.yaml
>
>
> We are unable to start version 2.2.1 on our Windows 2012 R2 systems, however 
> we can use the same environment to start version 2.1.2
> I have attached our Cassandra.yaml and cassandra-env.ps1 file for reference.  
> I have also attached the system.log file displaying the error.
> I have also included an excerpt of the log file showing the stack trace of 
> the error.
> ERROR [ScheduledTasks:1] 2015-09-29 07:03:47,482 
> DebuggableThreadPoolExecutor.java:242 - Error in ThreadPoolExecutor
> java.lang.NoClassDefFoundError: Could not initialize class 
> org.apache.cassandra.config.DatabaseDescriptor
>   at 
> org.apache.cassandra.utils.JVMStabilityInspector.inspectThrowable(JVMStabilityInspector.java:57)
>  ~[apache-cassandra-2.2.1.jar:2.2.1]
>   at 
> org.apache.cassandra.concurrent.DebuggableScheduledThreadPoolExecutor$UncomplainingRunnable.run(DebuggableScheduledThreadPoolExecutor.java:122)
>  ~[apache-cassandra-2.2.1.jar:2.2.1]
>   at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) 
> ~[na:1.8.0_51]
>   at java.util.concurrent.FutureTask.runAndReset(Unknown Source) 
> ~[na:1.8.0_51]
>   at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(Unknown
>  Source) ~[na:1.8.0_51]
>   at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(Unknown
>  Source) ~[na:1.8.0_51]
>   at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) 
> [na:1.8.0_51]
>   at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) 
> [na:1.8.0_51]
>   at java.lang.Thread.run(Unknown Source) [na:1.8.0_51]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10449) OOM on bootstrap due to long GC pause

2015-10-06 Thread Philip Thompson (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10449?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945149#comment-14945149
 ] 

Philip Thompson commented on CASSANDRA-10449:
-

[~yukim], do you want to look at this as well?

> OOM on bootstrap due to long GC pause
> -
>
> Key: CASSANDRA-10449
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10449
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
> Environment: Ubuntu 14.04, AWS
>Reporter: Robbie Strickland
>  Labels: gc
> Fix For: 2.1.x
>
> Attachments: system.log.10-05
>
>
> I have a 20-node cluster (i2.4xlarge) with vnodes (default of 256) and 
> 500-700GB per node.  SSTable counts are <10 per table.  I am attempting to 
> provision additional nodes, but bootstrapping OOMs every time after about 10 
> hours with a sudden long GC pause:
> {noformat}
> INFO  [Service Thread] 2015-10-05 23:33:33,373 GCInspector.java:252 - G1 Old 
> Generation GC in 1586126ms.  G1 Old Gen: 49213756976 -> 49072277176;
> ...
> ERROR [MemtableFlushWriter:454] 2015-10-05 23:33:33,380 
> CassandraDaemon.java:223 - Exception in thread 
> Thread[MemtableFlushWriter:454,5,main]
> java.lang.OutOfMemoryError: Java heap space
> {noformat}
> I have tried increasing max heap to 48G just to get through the bootstrap, to 
> no avail.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10242) Validate rack information on startup

2015-10-06 Thread Paulo Motta (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10242?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paulo Motta updated CASSANDRA-10242:

Reviewer: Paulo Motta

> Validate rack information on startup
> 
>
> Key: CASSANDRA-10242
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10242
> Project: Cassandra
>  Issue Type: Improvement
>Reporter: Jonathan Ellis
>Assignee: Carl Yeksigian
> Fix For: 2.1.x
>
>
> Moving to a new rack means that different data should be stored on a node.  
> We already persist rack information in a system table; we should fail startup 
> if this doesn't match what the snitch thinks it should be.  (Either the 
> snitch is wrong, and needs to be fixed, or the machine has been moved and 
> needs to be rebootstrapped.)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10389) Repair session exception Validation failed

2015-10-06 Thread Philip Thompson (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10389?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Philip Thompson updated CASSANDRA-10389:

Fix Version/s: 2.2.x

> Repair session exception Validation failed
> --
>
> Key: CASSANDRA-10389
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10389
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
> Environment: Debian 8, Java 1.8.0_60, Cassandra 2.2.1 (datastax 
> compilation)
>Reporter: Jędrzej Sieracki
> Fix For: 2.2.x
>
>
> I'm running a repair on a ring of nodes, that was recently extented from 3 to 
> 13 nodes. The extension was done two days ago, the repair was attempted 
> yesterday.
> {quote}
> [2015-09-22 11:55:55,266] Starting repair command #9, repairing keyspace 
> perspectiv with repair options (parallelism: parallel, primary range: false, 
> incremental: true, job threads: 1, ColumnFamilies: [], dataCenters: [], 
> hosts: [], # of ranges: 517)
> [2015-09-22 11:55:58,043] Repair session 1f7c50c0-6110-11e5-b992-9f13fa8664c8 
> for range (-5927186132136652665,-5917344746039874798] failed with error 
> [repair #1f7c50c0-6110-11e5-b992-9f13fa8664c8 on 
> perspectiv/stock_increment_agg, (-5927186132136652665,-5917344746039874798]] 
> Validation failed in cblade1.XXX/XXX (progress: 0%)
> {quote}
> BTW, I am ignoring the LEAK errors for now, that's outside of the scope of 
> the main issue:
> {quote}
> ERROR [Reference-Reaper:1] 2015-09-22 11:58:27,843 Ref.java:187 - LEAK 
> DETECTED: a reference 
> (org.apache.cassandra.utils.concurrent.Ref$State@4d25ad8f) to class 
> org.apache.cassandra.io.sstable.format.SSTableReader$InstanceTidier@896826067:/var/lib/cassandra/data/perspectiv/stock_increment_agg-840cad405de711e5b9929f13fa8664c8/la-73-big
>  was not released before the reference was garbage collected
> {quote}
> I scrubbed the sstable with failed validation on cblade1 with nodetool scrub 
> perspectiv stock_increment_agg:
> {quote}
> INFO  [CompactionExecutor:1704] 2015-09-22 12:05:31,615 OutputHandler.java:42 
> - Scrubbing 
> BigTableReader(path='/var/lib/cassandra/data/perspectiv/stock_increment_agg-840cad405de711e5b9929f13fa8664c8/la-83-big-Data.db')
>  (345466609 bytes)
> INFO  [CompactionExecutor:1703] 2015-09-22 12:05:31,615 OutputHandler.java:42 
> - Scrubbing 
> BigTableReader(path='/var/lib/cassandra/data/perspectiv/stock_increment_agg-840cad405de711e5b9929f13fa8664c8/la-82-big-Data.db')
>  (60496378 bytes)
> ERROR [Reference-Reaper:1] 2015-09-22 12:05:31,676 Ref.java:187 - LEAK 
> DETECTED: a reference 
> (org.apache.cassandra.utils.concurrent.Ref$State@4ca8951e) to class 
> org.apache.cassandra.io.sstable.format.SSTableReader$InstanceTidier@114161559:/var/lib/cassandra/data/perspectiv/receipt_agg_total-76abb0625de711e59f6e0b7d98a25b6e/la-48-big
>  was not released before the reference was garbage collected
> ERROR [Reference-Reaper:1] 2015-09-22 12:05:31,676 Ref.java:187 - LEAK 
> DETECTED: a reference 
> (org.apache.cassandra.utils.concurrent.Ref$State@eeb6383) to class 
> org.apache.cassandra.io.sstable.format.SSTableReader$InstanceTidier@1612685364:/var/lib/cassandra/data/perspectiv/stock_increment_agg-840cad405de711e5b9929f13fa8664c8/la-83-big
>  was not released before the reference was garbage collected
> ERROR [Reference-Reaper:1] 2015-09-22 12:05:31,676 Ref.java:187 - LEAK 
> DETECTED: a reference 
> (org.apache.cassandra.utils.concurrent.Ref$State@1de90543) to class 
> org.apache.cassandra.io.sstable.format.SSTableReader$InstanceTidier@2058626950:/var/lib/cassandra/data/perspectiv/receipt_agg_total-76abb0625de711e59f6e0b7d98a25b6e/la-49-big
>  was not released before the reference was garbage collected
> ERROR [Reference-Reaper:1] 2015-09-22 12:05:31,676 Ref.java:187 - LEAK 
> DETECTED: a reference 
> (org.apache.cassandra.utils.concurrent.Ref$State@15616385) to class 
> org.apache.cassandra.io.sstable.format.SSTableReader$InstanceTidier@1386628428:/var/lib/cassandra/data/perspectiv/receipt_agg_total-76abb0625de711e59f6e0b7d98a25b6e/la-47-big
>  was not released before the reference was garbage collected
> INFO  [CompactionExecutor:1703] 2015-09-22 12:05:35,098 OutputHandler.java:42 
> - Scrub of 
> BigTableReader(path='/var/lib/cassandra/data/perspectiv/stock_increment_agg-840cad405de711e5b9929f13fa8664c8/la-82-big-Data.db')
>  complete: 51397 rows in new sstable and 0 empty (tombstoned) rows dropped
> INFO  [CompactionExecutor:1704] 2015-09-22 12:05:47,605 OutputHandler.java:42 
> - Scrub of 
> BigTableReader(path='/var/lib/cassandra/data/perspectiv/stock_increment_agg-840cad405de711e5b9929f13fa8664c8/la-83-big-Data.db')
>  complete: 292600 rows in new sstable and 0 empty (tombstoned) rows dropped
> {quote}
> Now, after scrubbing, another repair was attempted, it did finish, but with 
> lots of errors from 

[jira] [Commented] (CASSANDRA-10393) LEAK DETECTED: a reference (org.apache.cassandra.utils.concurrent.Ref)

2015-10-06 Thread Philip Thompson (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10393?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945162#comment-14945162
 ] 

Philip Thompson commented on CASSANDRA-10393:
-

[~yukim], can you look at this repair issue?

> LEAK DETECTED: a reference (org.apache.cassandra.utils.concurrent.Ref)
> --
>
> Key: CASSANDRA-10393
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10393
> Project: Cassandra
>  Issue Type: Bug
> Environment: v 2.2.1 (from apt)
> -> lsb_release -a
> No LSB modules are available.
> Distributor ID:   Debian
> Description:  Debian GNU/Linux 7.8 (wheezy)
> Release:  7.8
> Codename: wheezy
> -> java -version
> java version "1.8.0_60"
> Java(TM) SE Runtime Environment (build 1.8.0_60-b27)
> Java HotSpot(TM) 64-Bit Server VM (build 25.60-b23, mixed mode)
>Reporter: Christian Winther
>  Labels: memory-leak, sstable
> Fix For: 2.2.x
>
>
> When trying to repair full on a table with the following schema my nodes 
> stall 
> and end up with spamming this 
> I've recently changed the table from SizeTieredCompactionStrategy to 
> LeveledCompactionStrategy.
> Coming from 2.1.9 -> 2.2.0 -> 2.2.1 i ran upgradesstable without issue as well
> When trying to full repair post compaction change, I got "out of order" 
> errors. A few google searches later, I was told to "scrub" the keyspace - did 
> that during the night (no problems logged, and no data lost)
> Now a repair just stalls and output memory leaks all over the place 
> {code}
> CREATE KEYSPACE sessions WITH replication = {'class': 'SimpleStrategy', 
> 'replication_factor': '3'}  AND durable_writes = true;
> CREATE TABLE sessions.sessions (
> id text PRIMARY KEY,
> client_ip text,
> controller text,
> controller_action text,
> created timestamp,
> data text,
> expires timestamp,
> http_host text,
> modified timestamp,
> request_agent text,
> request_agent_bot boolean,
> request_path text,
> site_id int,
> user_id int
> ) WITH bloom_filter_fp_chance = 0.01
> AND caching = '{"keys":"NONE", "rows_per_partition":"NONE"}'
> AND comment = ''
> AND compaction = {'class': 
> 'org.apache.cassandra.db.compaction.LeveledCompactionStrategy'}
> AND compression = {'sstable_compression': 
> 'org.apache.cassandra.io.compress.LZ4Compressor'}
> AND dclocal_read_repair_chance = 0.1
> AND default_time_to_live = 0
> AND gc_grace_seconds = 864000
> AND max_index_interval = 2048
> AND memtable_flush_period_in_ms = 0
> AND min_index_interval = 128
> AND read_repair_chance = 0.0
> AND speculative_retry = '99.0PERCENTILE';
> {code}
> {code}
> ERROR [Reference-Reaper:1] 2015-09-24 10:25:28,475 Ref.java:187 - LEAK 
> DETECTED: a reference 
> (org.apache.cassandra.utils.concurrent.Ref$State@4428a373) to class 
> org.apache.cassandra.io.sstable.format.SSTableReader$InstanceTidier@184765:/data/1/cassandra/sessions/sessions-77dd22f0ab9711e49cbc410c6b6f53a6/la-104037-big
>  was not released before the reference was garbage collected
> ERROR [Reference-Reaper:1] 2015-09-24 10:25:28,475 Ref.java:187 - LEAK 
> DETECTED: a reference 
> (org.apache.cassandra.utils.concurrent.Ref$State@368dd97) to class 
> org.apache.cassandra.io.sstable.format.SSTableReader$InstanceTidier@184765:/data/1/cassandra/sessions/sessions-77dd22f0ab9711e49cbc410c6b6f53a6/la-104037-big
>  was not released before the reference was garbage collected
> ERROR [Reference-Reaper:1] 2015-09-24 10:25:28,475 Ref.java:187 - LEAK 
> DETECTED: a reference 
> (org.apache.cassandra.utils.concurrent.Ref$State@66fb78be) to class 
> org.apache.cassandra.io.sstable.format.SSTableReader$InstanceTidier@184765:/data/1/cassandra/sessions/sessions-77dd22f0ab9711e49cbc410c6b6f53a6/la-104037-big
>  was not released before the reference was garbage collected
> ERROR [Reference-Reaper:1] 2015-09-24 10:25:28,475 Ref.java:187 - LEAK 
> DETECTED: a reference 
> (org.apache.cassandra.utils.concurrent.Ref$State@9fdd2e6) to class 
> org.apache.cassandra.io.sstable.format.SSTableReader$InstanceTidier@1460906269:/data/1/cassandra/sessions/sessions-77dd22f0ab9711e49cbc410c6b6f53a6/la-104788-big
>  was not released before the reference was garbage collected
> ERROR [Reference-Reaper:1] 2015-09-24 10:25:28,475 Ref.java:187 - LEAK 
> DETECTED: a reference 
> (org.apache.cassandra.utils.concurrent.Ref$State@84fcb91) to class 
> org.apache.cassandra.io.sstable.format.SSTableReader$InstanceTidier@1460906269:/data/1/cassandra/sessions/sessions-77dd22f0ab9711e49cbc410c6b6f53a6/la-104788-big
>  was not released before the reference was garbage collected
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10448) "Unknown type 0" Stream failure on Repair

2015-10-06 Thread Omri Iluz (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10448?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945447#comment-14945447
 ] 

Omri Iluz commented on CASSANDRA-10448:
---

all nodes are within one availability zone in Google's data center. We never 
had any network issues and any network test I am running in parallel to the 
repair session seems to return fine.

I thought this might be due to corruption in the sstables somewhere so I ran 
scrub, but that also didn't help.

Any other ideas or things I can try ?

> "Unknown type 0" Stream failure on Repair
> -
>
> Key: CASSANDRA-10448
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10448
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
> Environment: Cassandra 2.2.2
> 5 Nodes in Google Compute Engine
> Java 1.8.0_60
>Reporter: Omri Iluz
>Assignee: Yuki Morishita
> Fix For: 2.2.x
>
>
> While running repair after upgrading to 2.2.2 I am getting many stream fail 
> errors:
> {noformat}
> [2015-10-05 23:52:30,353] Repair session 4c181051-6bbb-11e5-acdb-d9a8bbd39330 
> for range (59694553044959221,86389982480621619] failed with error [repair 
> #4c181051-6bbb-11e5-acdb-d9a8bbd39330 on px/acti
> vities, (59694553044959221,86389982480621619]] Sync failed between 
> /10.240.81.104 and /10.240.134.221 (progress: 4%)
> {noformat}
> Logs from both sides of the stream:
> Sides 1 -
> {noformat}
> INFO  [STREAM-INIT-/10.240.81.104:52722] 2015-10-05 23:52:30,063 
> StreamResultFuture.java:111 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550 
> ID#0] Creating new streaming plan for Repair
> INFO  [STREAM-INIT-/10.240.81.104:52722] 2015-10-05 23:52:30,063 
> StreamResultFuture.java:118 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550, 
> ID#0] Received streaming plan for Repair
> INFO  [STREAM-INIT-/10.240.81.104:52723] 2015-10-05 23:52:30,063 
> StreamResultFuture.java:118 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550, 
> ID#0] Received streaming plan for Repair
> INFO  [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,098 
> StreamResultFuture.java:168 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550 
> ID#0] Prepare completed. Receiving 13 files(517391317 bytes), sending 10 
> files(469491729 bytes)
> ERROR [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,234 
> StreamSession.java:524 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Streaming error occurred
> java.lang.IllegalArgumentException: Unknown type 0
>   at 
> org.apache.cassandra.streaming.messages.StreamMessage$Type.get(StreamMessage.java:96)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:57)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at java.lang.Thread.run(Thread.java:745) [na:1.8.0_60]
> INFO  [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,302 
> StreamResultFuture.java:182 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Session with /10.240.81.104 is complete
> WARN  [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,302 
> StreamResultFuture.java:209 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Stream failed
> {noformat}
> Side 2 -
> {noformat}
> INFO  [AntiEntropyStage:1] 2015-10-05 23:52:30,060 StreamResultFuture.java:86 
> - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] Executing streaming plan for 
> Repair
> INFO  [StreamConnectionEstablisher:6] 2015-10-05 23:52:30,061 
> StreamSession.java:232 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Starting streaming to /10.240.134.221
> INFO  [StreamConnectionEstablisher:6] 2015-10-05 23:52:30,063 
> StreamCoordinator.java:213 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550, 
> ID#0] Beginning stream session with /10.240.134.221
> INFO  [STREAM-IN-/10.240.134.221] 2015-10-05 23:52:30,098 
> StreamResultFuture.java:168 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550 
> ID#0] Prepare completed. Receiving 10 files(469491729 bytes), sending 13 
> files(517391317 bytes)
> INFO  [STREAM-IN-/10.240.134.221] 2015-10-05 23:52:30,349 
> StreamResultFuture.java:182 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Session with /10.240.134.221 is complete
> ERROR [STREAM-OUT-/10.240.134.221] 2015-10-05 23:52:30,349 
> StreamSession.java:524 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Streaming error occurred
> org.apache.cassandra.io.FSReadError: java.io.IOException: Broken pipe
>   at 
> org.apache.cassandra.io.util.ChannelProxy.transferTo(ChannelProxy.java:144) 
> ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> org.apache.cassandra.streaming.compress.CompressedStreamWriter$1.apply(CompressedStreamWriter.java:79)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 

[jira] [Commented] (CASSANDRA-10438) Overwrites of rows in memtable produce incorrect deltas for indexing

2015-10-06 Thread Ariel Weisberg (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10438?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945552#comment-14945552
 ] 

Ariel Weisberg commented on CASSANDRA-10438:


Unless I am mistaken I can revert the fix to onPrimaryKeyLivenessInfo and the 
test still passes?

The fix itself and the description of the bug make sense to me. The test is 
convincing in that it checks the contents of the update case. I'm just not sure 
about onPrimaryKeyLivenessInfo.

I looked at the cassci results. 
org.apache.cassandra.cql3.validation.operations.CreateTest.testCQL3PartitionKeyOnlyTable
 is unusual, but I didn't reproduce it.

> Overwrites of rows in memtable produce incorrect deltas for indexing
> 
>
> Key: CASSANDRA-10438
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10438
> Project: Cassandra
>  Issue Type: Improvement
>Reporter: Sam Tunnicliffe
>Assignee: Sam Tunnicliffe
> Fix For: 3.0.0 rc2
>
>
> When a row in the memtable is updated, the delta is supplied to any 
> registered indexer. This consists of two {{Row}} objects, representing the 
> old and new data in the memtable. As per its javadoc, the contract of 
> {{Index.Indexer::updateRow}} is that these old & new rows contain only the 
> changed columns, so any column which was not affected by the update will 
> appear in neither the new nor old row. The {{RowDiffListener::onCell}} method 
> in {{SecondaryIndexManager.WriteTimeTransaction::onUpdated}} which produces 
> these deltas uses a reference equality check, where it should be checking 
> object equality. This results in unchanged, prexisting cells appearing in the 
> {{toInsert}} row.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10437) Remove offheap_objects option until 9472 re-introduces them

2015-10-06 Thread Ariel Weisberg (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10437?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945466#comment-14945466
 ] 

Ariel Weisberg commented on CASSANDRA-10437:


Created a branch for cassci to run it. Will report back.

> Remove offheap_objects option until 9472 re-introduces them
> ---
>
> Key: CASSANDRA-10437
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10437
> Project: Cassandra
>  Issue Type: Bug
>Reporter: Sylvain Lebresne
>Assignee: Sylvain Lebresne
>Priority: Trivial
> Fix For: 3.0.0 rc2
>
> Attachments: 0001-Remove-offheap_objects-option.txt
>
>
> We need to send a meaningful error message if user try to use 
> {{offheap_objects}} in 3.0 since it's not supported currently (pending 
> CASSANDRA-9472). And document the current removal in the NEWS file.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10451) Fix batch_test.TestBatch.logged_batch_gcgs_below_threshold_{multi,single}_table_test on CassCI

2015-10-06 Thread Paulo Motta (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945579#comment-14945579
 ] 

Paulo Motta commented on CASSANDRA-10451:
-

Created [dtest PR|https://github.com/riptano/cassandra-dtest/pull/583] with 
fixes. Basically ccm/dtest merges system.log and debug.log, so info/warn/error 
statements are duplicated. Modified tests to only assert logs are present, and 
not count them.

> Fix 
> batch_test.TestBatch.logged_batch_gcgs_below_threshold_{multi,single}_table_test
>  on CassCI
> --
>
> Key: CASSANDRA-10451
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10451
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Paulo Motta
> Fix For: 3.0.0 rc2
>
>
> This is a recent regression:
> http://cassci.datastax.com/view/trunk/job/trunk_dtest/lastCompletedBuild/testReport/batch_test/TestBatch/logged_batch_gcgs_below_threshold_multi_table_test/history/
> Based on [when it started 
> failing|http://cassci.datastax.com/view/trunk/job/trunk_dtest/623/changes], 
> it looks like the regression was introduced by [the recent logging 
> changes|https://github.com/apache/cassandra/commit/4a849efeb7c7c1a54bc12094d2d6a9f3f008a2fa].
>  Pinging [~blerer] and [~pauloricardomg] to have a look.
> My initial guess was that the logs the test depends on were moved to the new 
> debug log. However, the tests passes when I run it manually on openstack, so 
> it could just be a configuration issue on CassCI.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10455) Fix pep8 compliance in cqlsh.py

2015-10-06 Thread Philip Thompson (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10455?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945596#comment-14945596
 ] 

Philip Thompson commented on CASSANDRA-10455:
-

I missed this being broken when i reviewed CASSANDRA-9855.

http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/210/changes

> Fix pep8 compliance in cqlsh.py
> ---
>
> Key: CASSANDRA-10455
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10455
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Philip Thompson
>  Labels: cqlsh
> Fix For: 3.0.0 rc2
>
>
> cqlsh has fallen out of pep8 compliance:
> http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/cqlsh_tests.cqlsh_tests/TestCqlsh/test_pep8_compliance/
> [~philipthompson] has said off-JIRA that he'll fix it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10453) Fix compression_cql_options_test dtest

2015-10-06 Thread Paulo Motta (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945583#comment-14945583
 ] 

Paulo Motta commented on CASSANDRA-10453:
-

Created [dtest PR|https://github.com/riptano/cassandra-dtest/pull/583] with 
fixes. Basically ccm/dtest merges system.log and debug.log, so info/warn/error 
statements are duplicated. Modified tests to only assert logs are present, and 
not count them.

> Fix compression_cql_options_test dtest
> --
>
> Key: CASSANDRA-10453
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10453
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Paulo Motta
> Fix For: 3.0.0 rc2
>
>
> This has been failing hard for a while:
> http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/206/testReport/junit/compression_test/TestCompression/compression_cql_options_test/history/
> This regression, like CASSANDRA-10451, started when the recent log changes 
> were introduced:
> http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/206/changes
> I also can't reproduce on OpenStack. Pinging [~pauloricardomg] and [~blerer] 
> again.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10455) Fix pep8 compliance in cqlsh.py

2015-10-06 Thread Philip Thompson (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10455?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Philip Thompson updated CASSANDRA-10455:

Labels: cqlsh  (was: )

> Fix pep8 compliance in cqlsh.py
> ---
>
> Key: CASSANDRA-10455
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10455
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Philip Thompson
>  Labels: cqlsh
> Fix For: 3.0.0 rc2
>
>
> cqlsh has fallen out of pep8 compliance:
> http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/cqlsh_tests.cqlsh_tests/TestCqlsh/test_pep8_compliance/
> [~philipthompson] has said off-JIRA that he'll fix it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-10455) Fix pep8 compliance in cqlsh.py

2015-10-06 Thread Jim Witschey (JIRA)
Jim Witschey created CASSANDRA-10455:


 Summary: Fix pep8 compliance in cqlsh.py
 Key: CASSANDRA-10455
 URL: https://issues.apache.org/jira/browse/CASSANDRA-10455
 Project: Cassandra
  Issue Type: Sub-task
Reporter: Jim Witschey
Assignee: Philip Thompson
 Fix For: 3.0.0 rc2


cqlsh has fallen out of pep8 compliance:

http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/cqlsh_tests.cqlsh_tests/TestCqlsh/test_pep8_compliance/

[~philipthompson] has said off-JIRA that he'll fix it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-10454) fix failing cqlsh COPY dtests

2015-10-06 Thread Jim Witschey (JIRA)
Jim Witschey created CASSANDRA-10454:


 Summary: fix failing cqlsh COPY dtests
 Key: CASSANDRA-10454
 URL: https://issues.apache.org/jira/browse/CASSANDRA-10454
 Project: Cassandra
  Issue Type: Sub-task
Reporter: Jim Witschey
Assignee: Jim Witschey


The proximate cause of these failures:

http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/cqlsh_tests.cqlsh_copy_tests/CqlshCopyTest/test_null_as_null_indicator/
http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/cqlsh_tests.cqlsh_copy_tests/CqlshCopyTest/test_undefined_as_null_indicator/

is an error in the dtest logic. I'll take this one on; I've filed a dtest PR 
here:

https://github.com/riptano/cassandra-dtest/pull/582





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10438) Overwrites of rows in memtable produce incorrect deltas for indexing

2015-10-06 Thread Ariel Weisberg (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10438?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ariel Weisberg updated CASSANDRA-10438:
---
Reviewer: Ariel Weisberg

> Overwrites of rows in memtable produce incorrect deltas for indexing
> 
>
> Key: CASSANDRA-10438
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10438
> Project: Cassandra
>  Issue Type: Improvement
>Reporter: Sam Tunnicliffe
>Assignee: Sam Tunnicliffe
> Fix For: 3.0.0 rc2
>
>
> When a row in the memtable is updated, the delta is supplied to any 
> registered indexer. This consists of two {{Row}} objects, representing the 
> old and new data in the memtable. As per its javadoc, the contract of 
> {{Index.Indexer::updateRow}} is that these old & new rows contain only the 
> changed columns, so any column which was not affected by the update will 
> appear in neither the new nor old row. The {{RowDiffListener::onCell}} method 
> in {{SecondaryIndexManager.WriteTimeTransaction::onUpdated}} which produces 
> these deltas uses a reference equality check, where it should be checking 
> object equality. This results in unchanged, prexisting cells appearing in the 
> {{toInsert}} row.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10242) Validate rack information on startup

2015-10-06 Thread Paulo Motta (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10242?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945762#comment-14945762
 ] 

Paulo Motta commented on CASSANDRA-10242:
-

Patch LGTM overall, could you just add a simple dtest changing the rack of a 
single node and check the node will fail to initialize? {{replication_test.py}} 
have some example of changing snitch/rack information.

Also, is there any reason you changed the default rack and dc assignments of 
{{conf/cassandra-rackdc.properties}} ?

> Validate rack information on startup
> 
>
> Key: CASSANDRA-10242
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10242
> Project: Cassandra
>  Issue Type: Improvement
>Reporter: Jonathan Ellis
>Assignee: Carl Yeksigian
> Fix For: 2.1.x
>
>
> Moving to a new rack means that different data should be stored on a node.  
> We already persist rack information in a system table; we should fail startup 
> if this doesn't match what the snitch thinks it should be.  (Either the 
> snitch is wrong, and needs to be fixed, or the machine has been moved and 
> needs to be rebootstrapped.)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-10456) fix deprecated_repair_test dtests

2015-10-06 Thread Jim Witschey (JIRA)
Jim Witschey created CASSANDRA-10456:


 Summary: fix deprecated_repair_test dtests
 Key: CASSANDRA-10456
 URL: https://issues.apache.org/jira/browse/CASSANDRA-10456
 Project: Cassandra
  Issue Type: Sub-task
Reporter: Jim Witschey
Assignee: Jim Witschey


All the deprecated_repair_test dtests are failing on CassCI. I believe the 
immediate cause is some recent changes to logs and how ccm handles them, like 
the failures in CASSANDRA-10451 and CASSANDRA-10453. Test failures on CassCI 
here:

http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/deprecated_repair_test/

I'm assigning myself to track this, in case there are other failures to deal 
with.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10456) fix deprecated_repair_test dtests

2015-10-06 Thread Yuki Morishita (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10456?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945720#comment-14945720
 ] 

Yuki Morishita commented on CASSANDRA-10456:


I think these are from log change.
Can you try with the latest ccm?
My PR to fix log was merged. (https://github.com/pcmanus/ccm/pull/380)

> fix deprecated_repair_test dtests
> -
>
> Key: CASSANDRA-10456
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10456
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Jim Witschey
> Fix For: 3.0.0 rc2
>
>
> All the deprecated_repair_test dtests are failing on CassCI. I believe the 
> immediate cause is some recent changes to logs and how ccm handles them, like 
> the failures in CASSANDRA-10451 and CASSANDRA-10453. Test failures on CassCI 
> here:
> http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/deprecated_repair_test/
> I'm assigning myself to track this, in case there are other failures to deal 
> with.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10412) Could not initialize class org.apache.cassandra.config.DatabaseDescriptor

2015-10-06 Thread Joshua McKenzie (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10412?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945744#comment-14945744
 ] 

Joshua McKenzie commented on CASSANDRA-10412:
-

Windows timer init's pretty flexible, so move it around in that method at-will.

> Could not initialize class org.apache.cassandra.config.DatabaseDescriptor
> -
>
> Key: CASSANDRA-10412
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10412
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
> Environment: OS: Windows 2012 R2
> Java JRE: 1.8.0_51
>Reporter: Eric Simmons
>Assignee: Carl Yeksigian
>Priority: Minor
> Fix For: 2.2.x
>
> Attachments: cassandra-env.ps1, cassandra.yaml
>
>
> We are unable to start version 2.2.1 on our Windows 2012 R2 systems, however 
> we can use the same environment to start version 2.1.2
> I have attached our Cassandra.yaml and cassandra-env.ps1 file for reference.  
> I have also attached the system.log file displaying the error.
> I have also included an excerpt of the log file showing the stack trace of 
> the error.
> ERROR [ScheduledTasks:1] 2015-09-29 07:03:47,482 
> DebuggableThreadPoolExecutor.java:242 - Error in ThreadPoolExecutor
> java.lang.NoClassDefFoundError: Could not initialize class 
> org.apache.cassandra.config.DatabaseDescriptor
>   at 
> org.apache.cassandra.utils.JVMStabilityInspector.inspectThrowable(JVMStabilityInspector.java:57)
>  ~[apache-cassandra-2.2.1.jar:2.2.1]
>   at 
> org.apache.cassandra.concurrent.DebuggableScheduledThreadPoolExecutor$UncomplainingRunnable.run(DebuggableScheduledThreadPoolExecutor.java:122)
>  ~[apache-cassandra-2.2.1.jar:2.2.1]
>   at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) 
> ~[na:1.8.0_51]
>   at java.util.concurrent.FutureTask.runAndReset(Unknown Source) 
> ~[na:1.8.0_51]
>   at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(Unknown
>  Source) ~[na:1.8.0_51]
>   at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(Unknown
>  Source) ~[na:1.8.0_51]
>   at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) 
> [na:1.8.0_51]
>   at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) 
> [na:1.8.0_51]
>   at java.lang.Thread.run(Unknown Source) [na:1.8.0_51]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-9602) EACH_QUORUM READ support needed

2015-10-06 Thread Ariel Weisberg (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-9602?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ariel Weisberg updated CASSANDRA-9602:
--
Reviewer: Ariel Weisberg

> EACH_QUORUM READ support needed
> ---
>
> Key: CASSANDRA-9602
> URL: https://issues.apache.org/jira/browse/CASSANDRA-9602
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Scott Guminy
>Assignee: Carl Yeksigian
>  Labels: client-impacting, doc-impacting
> Fix For: 3.x
>
>
> EACH_QUORUM consistency for READ should be added.
> This bug https://issues.apache.org/jira/browse/CASSANDRA-3272 says it is not 
> needed ever, however I have a use case where I need it.  I think the decision 
> made was incorrect. Here's why...
>  
>  My application has two key pieces:
>  
>  # *End user actions* which add/modify data in the system.  End users 
> typically access the application from only one Data Center and only see their 
> own data
> # *Scheduled business logic tasks* which run from any node in any data 
> center.  These tasks process data added by the end users in an asynchronous 
> way
>  
>  *End user actions must have the highest degree of availability.*  Users must 
> always be able to add data to the system.  The data will be processed later.  
> To support this, end user actions will use *LOCAL_QUORUM Read and Write 
> Consistency*.
>  
>  Scheduled tasks don't need to have a high degree of availability but MUST 
> operate on the most up to date data.  *The tasks will run with EACH_QUORUM* 
> to ensure that no matter how many data centers we have, we always READ the 
> latest data.  This approach allows us some amount of fault tolerance. 
>  
>  The problem is that EACH_QUORUM is not a valid READ consistency level.  This 
> mean I have no alternative but to use ALL.  ALL will work, but is not the 
> best since it offers support for ZERO failures.  I would prefer EACH_QUORUM 
> since it can support some failures in each data center.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10412) Could not initialize class org.apache.cassandra.config.DatabaseDescriptor

2015-10-06 Thread Paulo Motta (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10412?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945671#comment-14945671
 ] 

Paulo Motta commented on CASSANDRA-10412:
-

Overall looks good, but I'd rather not add the {{Config.setEmbeddedService()}} 
method only for this exit check on {{DatabaseDescriptor}} if there are 
alternatives. 

Can't we move the Windows timer period initialization to after 
{{DatabaseDescriptor.forceStaticInitialization()}} on 
{{CassandraDaemon.activate()}} ? The {{DatabaseDescriptor}} static 
initialization will be done anyway before the timer setup, so unless there's 
some nit I'm not aware of with this Windows timer period thing I don't see a 
problem with this approach. We'd still need to throw 
{{ExceptionInInitializerError}} after catching a {{ConfigurationException}} on  
{{DatabaseDescriptor.forceStaticInitialization()}}, so 
{{CassandraDaemon.activate()}} would fail and exit with a nice message after a 
configuration error.

> Could not initialize class org.apache.cassandra.config.DatabaseDescriptor
> -
>
> Key: CASSANDRA-10412
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10412
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
> Environment: OS: Windows 2012 R2
> Java JRE: 1.8.0_51
>Reporter: Eric Simmons
>Assignee: Carl Yeksigian
>Priority: Minor
> Fix For: 2.2.x
>
> Attachments: cassandra-env.ps1, cassandra.yaml
>
>
> We are unable to start version 2.2.1 on our Windows 2012 R2 systems, however 
> we can use the same environment to start version 2.1.2
> I have attached our Cassandra.yaml and cassandra-env.ps1 file for reference.  
> I have also attached the system.log file displaying the error.
> I have also included an excerpt of the log file showing the stack trace of 
> the error.
> ERROR [ScheduledTasks:1] 2015-09-29 07:03:47,482 
> DebuggableThreadPoolExecutor.java:242 - Error in ThreadPoolExecutor
> java.lang.NoClassDefFoundError: Could not initialize class 
> org.apache.cassandra.config.DatabaseDescriptor
>   at 
> org.apache.cassandra.utils.JVMStabilityInspector.inspectThrowable(JVMStabilityInspector.java:57)
>  ~[apache-cassandra-2.2.1.jar:2.2.1]
>   at 
> org.apache.cassandra.concurrent.DebuggableScheduledThreadPoolExecutor$UncomplainingRunnable.run(DebuggableScheduledThreadPoolExecutor.java:122)
>  ~[apache-cassandra-2.2.1.jar:2.2.1]
>   at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) 
> ~[na:1.8.0_51]
>   at java.util.concurrent.FutureTask.runAndReset(Unknown Source) 
> ~[na:1.8.0_51]
>   at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(Unknown
>  Source) ~[na:1.8.0_51]
>   at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(Unknown
>  Source) ~[na:1.8.0_51]
>   at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) 
> [na:1.8.0_51]
>   at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) 
> [na:1.8.0_51]
>   at java.lang.Thread.run(Unknown Source) [na:1.8.0_51]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10456) fix deprecated_repair_test dtests

2015-10-06 Thread Jim Witschey (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10456?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945739#comment-14945739
 ] 

Jim Witschey commented on CASSANDRA-10456:
--

[~yukim] Thanks, and sorry I wasn't clear. I'll rerun the tests; based on 
running it locally, I think that'll fix it.

> fix deprecated_repair_test dtests
> -
>
> Key: CASSANDRA-10456
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10456
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Jim Witschey
> Fix For: 3.0.0 rc2
>
>
> All the deprecated_repair_test dtests are failing on CassCI. I believe the 
> immediate cause is some recent changes to logs and how ccm handles them, like 
> the failures in CASSANDRA-10451 and CASSANDRA-10453. Test failures on CassCI 
> here:
> http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/deprecated_repair_test/
> I'm assigning myself to track this, in case there are other failures to deal 
> with.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10448) "Unknown type 0" Stream failure on Repair

2015-10-06 Thread Paulo Motta (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10448?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945735#comment-14945735
 ] 

Paulo Motta commented on CASSANDRA-10448:
-

[~mroi] can you please attach debug.log of both nodes for the failure period?

> "Unknown type 0" Stream failure on Repair
> -
>
> Key: CASSANDRA-10448
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10448
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
> Environment: Cassandra 2.2.2
> 5 Nodes in Google Compute Engine
> Java 1.8.0_60
>Reporter: Omri Iluz
>Assignee: Yuki Morishita
> Fix For: 2.2.x
>
>
> While running repair after upgrading to 2.2.2 I am getting many stream fail 
> errors:
> {noformat}
> [2015-10-05 23:52:30,353] Repair session 4c181051-6bbb-11e5-acdb-d9a8bbd39330 
> for range (59694553044959221,86389982480621619] failed with error [repair 
> #4c181051-6bbb-11e5-acdb-d9a8bbd39330 on px/acti
> vities, (59694553044959221,86389982480621619]] Sync failed between 
> /10.240.81.104 and /10.240.134.221 (progress: 4%)
> {noformat}
> Logs from both sides of the stream:
> Sides 1 -
> {noformat}
> INFO  [STREAM-INIT-/10.240.81.104:52722] 2015-10-05 23:52:30,063 
> StreamResultFuture.java:111 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550 
> ID#0] Creating new streaming plan for Repair
> INFO  [STREAM-INIT-/10.240.81.104:52722] 2015-10-05 23:52:30,063 
> StreamResultFuture.java:118 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550, 
> ID#0] Received streaming plan for Repair
> INFO  [STREAM-INIT-/10.240.81.104:52723] 2015-10-05 23:52:30,063 
> StreamResultFuture.java:118 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550, 
> ID#0] Received streaming plan for Repair
> INFO  [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,098 
> StreamResultFuture.java:168 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550 
> ID#0] Prepare completed. Receiving 13 files(517391317 bytes), sending 10 
> files(469491729 bytes)
> ERROR [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,234 
> StreamSession.java:524 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Streaming error occurred
> java.lang.IllegalArgumentException: Unknown type 0
>   at 
> org.apache.cassandra.streaming.messages.StreamMessage$Type.get(StreamMessage.java:96)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:57)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at java.lang.Thread.run(Thread.java:745) [na:1.8.0_60]
> INFO  [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,302 
> StreamResultFuture.java:182 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Session with /10.240.81.104 is complete
> WARN  [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,302 
> StreamResultFuture.java:209 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Stream failed
> {noformat}
> Side 2 -
> {noformat}
> INFO  [AntiEntropyStage:1] 2015-10-05 23:52:30,060 StreamResultFuture.java:86 
> - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] Executing streaming plan for 
> Repair
> INFO  [StreamConnectionEstablisher:6] 2015-10-05 23:52:30,061 
> StreamSession.java:232 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Starting streaming to /10.240.134.221
> INFO  [StreamConnectionEstablisher:6] 2015-10-05 23:52:30,063 
> StreamCoordinator.java:213 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550, 
> ID#0] Beginning stream session with /10.240.134.221
> INFO  [STREAM-IN-/10.240.134.221] 2015-10-05 23:52:30,098 
> StreamResultFuture.java:168 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550 
> ID#0] Prepare completed. Receiving 10 files(469491729 bytes), sending 13 
> files(517391317 bytes)
> INFO  [STREAM-IN-/10.240.134.221] 2015-10-05 23:52:30,349 
> StreamResultFuture.java:182 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Session with /10.240.134.221 is complete
> ERROR [STREAM-OUT-/10.240.134.221] 2015-10-05 23:52:30,349 
> StreamSession.java:524 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Streaming error occurred
> org.apache.cassandra.io.FSReadError: java.io.IOException: Broken pipe
>   at 
> org.apache.cassandra.io.util.ChannelProxy.transferTo(ChannelProxy.java:144) 
> ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> org.apache.cassandra.streaming.compress.CompressedStreamWriter$1.apply(CompressedStreamWriter.java:79)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> org.apache.cassandra.streaming.compress.CompressedStreamWriter$1.apply(CompressedStreamWriter.java:76)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> 

[jira] [Commented] (CASSANDRA-9602) EACH_QUORUM READ support needed

2015-10-06 Thread Ariel Weisberg (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-9602?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945782#comment-14945782
 ] 

Ariel Weisberg commented on CASSANDRA-9602:
---

What is the doc impact here? I looked at the [2.2 
docs|http://docs.datastax.com/en/cassandra/2.2/cassandra/dml/dmlConfigConsistency.html]
 and it makes it seem like EACH_QUORUM for reads is already supported, but the 
code makes it look like it would throw InvalidRequestException?

Do we need to have the 2.2 and earlier docs updated?

How is this tested? I looked at consistency_test.py test and I don't think it 
tests combinations with EACH_QUORUM for reads. It looks like ConsistencyLevel 
is unit testable, but I don't see a test for the existing code. Could you add 
tests for the code that changed?

[Is dcsEndpoints something worth caching in some 
way?|https://github.com/apache/cassandra/compare/trunk...carlyeks:ticket/9602#diff-6f46fd4f8a36f42512ae5f848aee033eR225]

[What is the expected load balancing behavior 
here?|https://github.com/apache/cassandra/compare/trunk...carlyeks:ticket/9602#diff-6f46fd4f8a36f42512ae5f848aee033eR244]
 Is it (and should it) hit the same nodes in each DC every time?

> EACH_QUORUM READ support needed
> ---
>
> Key: CASSANDRA-9602
> URL: https://issues.apache.org/jira/browse/CASSANDRA-9602
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Scott Guminy
>Assignee: Carl Yeksigian
>  Labels: client-impacting, doc-impacting
> Fix For: 3.x
>
>
> EACH_QUORUM consistency for READ should be added.
> This bug https://issues.apache.org/jira/browse/CASSANDRA-3272 says it is not 
> needed ever, however I have a use case where I need it.  I think the decision 
> made was incorrect. Here's why...
>  
>  My application has two key pieces:
>  
>  # *End user actions* which add/modify data in the system.  End users 
> typically access the application from only one Data Center and only see their 
> own data
> # *Scheduled business logic tasks* which run from any node in any data 
> center.  These tasks process data added by the end users in an asynchronous 
> way
>  
>  *End user actions must have the highest degree of availability.*  Users must 
> always be able to add data to the system.  The data will be processed later.  
> To support this, end user actions will use *LOCAL_QUORUM Read and Write 
> Consistency*.
>  
>  Scheduled tasks don't need to have a high degree of availability but MUST 
> operate on the most up to date data.  *The tasks will run with EACH_QUORUM* 
> to ensure that no matter how many data centers we have, we always READ the 
> latest data.  This approach allows us some amount of fault tolerance. 
>  
>  The problem is that EACH_QUORUM is not a valid READ consistency level.  This 
> mean I have no alternative but to use ALL.  ALL will work, but is not the 
> best since it offers support for ZERO failures.  I would prefer EACH_QUORUM 
> since it can support some failures in each data center.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-10458) cqlshrc: add option to always use ssl

2015-10-06 Thread Matt Wringe (JIRA)
Matt Wringe created CASSANDRA-10458:
---

 Summary: cqlshrc: add option to always use ssl
 Key: CASSANDRA-10458
 URL: https://issues.apache.org/jira/browse/CASSANDRA-10458
 Project: Cassandra
  Issue Type: Improvement
Reporter: Matt Wringe


I am currently running on a system in which my cassandra cluster is only 
accessible over tls.

The cqlshrc file is used to specify the host, the certificates and other 
configurations, but one option its missing is to always connect over ssl.

I would like to be able to call 'cqlsh' instead of always having to specify 
'cqlsh --ssl'






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-9304) COPY TO improvements

2015-10-06 Thread Tyler Hobbs (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-9304?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945873#comment-14945873
 ] 

Tyler Hobbs commented on CASSANDRA-9304:


Overall the patch is definitely a nice improvement!  Most of my review comments 
are about using more idiomatic python.

Imports:
 * In the imports, StringIO is imported later.  Also, to avoid flake8 warnings, 
add {{# noqa}} to the {{from io import StringIO}} line.

ImportExportConfig:
* The way that checks for unhandled options are performed is kind of strange.  
In perform_csv_import() they're checked immediately, but in 
perform_csv_export() it's deferred to ExportTask.  Additionally, a lot of 
functions seem to modify the options inside an ImportExportConfig instance, 
which is somewhat unexpected behavior and seems likely to result in bugs in the 
future. 
* In Python it's pretty unusual to use a class to only store mutable data, so I 
would replace the ImportExportConfig class with a function that returns a tuple 
(e.g. {{(csv_options, dialect_options, unrecognized_options)}}) or a dict.

perform_csv_import():
 * Why do {{csv_options = config.csv_options}} if it's only used once a couple 
of lines later?
 * As a matter of code style, avoid using tuple unpacking unless you're 
actually unpacking a tuple or the assignments are very trivial (e.g. {{a, b = 
None, None}})
   
ExportTask:
* The class should inherit from object
* get_ranges():
** There's no need to use a lambda sorting key on a list of tuples.  Tuples are 
naturally sorted sanely.
** Instead of {{del hosts\[:\]}}, just say {{hosts = \[\]}}.  You can also 
remove {{hosts = \[hostname\]}} above.
** Instead of doing {{for entry in ring}} and then using {{entry\[0\]}} and 
{{entry\[1\]}}, you can unpack the tuple in the loop: {{for token, replicas in 
ring:}}
** The final {{ranges.append()}} call could use a comment for explanation.  If 
you make the suggested changes to {{hosts}}, just use {{ranges\[-1\]\[1\]}} for 
the hosts (unless it's empty).  This approach is a little more obvious for 
readers.
* In {{send_initial_work()}}, it's idiomatic in python to use {{_}} for the 
name of an unused variable, so the loop should be: {{for _ in 
xrange(num_parallel_jobs)}}.

ExportProcess:
* In {{start_job()}}, it's not clear what would result in an IndexError and why 
we're ignoring it.  Try to limit try/excepts to the minimum number of lines 
they need to contain, and add a comment explaining what's going on there.
* ExportSession should be a separate top-level class instead of a nested one.  
(It's pretty unusual to nest classes.)  Also, it should inherit from object().
* I would rename {{handle_result}} to {{attach_callbacks}} and rename 
{{return_result}} to {{write_rows_to_csv}}.
* The host selection code in {{get_session}} is over-complicated.  How about 
something like:

{code}
new_hosts = [h for h in hosts if h not in self.hosts_to_sessions]
if new_hosts:
host = new_hosts[0]
# open Cluster, etc
else: 
host = min(hosts, key=lambda h: self.hosts_to_sessions[h].jobs)
session = self.hosts_to_sessions[host]
session.jobs += 1
return session
{code}

General:
* Try to add a few more code comments explaining what's going on at a high 
level.  Documenting method return values is also good (since there's no type 
information about that).
* You may want to put the csv-related classes and functions into separate 
modules under {{pylib}}.


> COPY TO improvements
> 
>
> Key: CASSANDRA-9304
> URL: https://issues.apache.org/jira/browse/CASSANDRA-9304
> Project: Cassandra
>  Issue Type: Improvement
>  Components: Core
>Reporter: Jonathan Ellis
>Assignee: Stefania
>Priority: Minor
>  Labels: cqlsh
> Fix For: 2.1.x
>
>
> COPY FROM has gotten a lot of love.  COPY TO not so much.  One obvious 
> improvement could be to parallelize reading and writing (write one page of 
> data while fetching the next).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-7645) cqlsh: show partial trace if incomplete after max_trace_wait

2015-10-06 Thread Paulo Motta (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-7645?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945874#comment-14945874
 ] 

Paulo Motta commented on CASSANDRA-7645:


+1, could you provide 2.2+ patches? apparently cqlsh was moved from bin/cqlsh 
to bin/cqlsh.py, so the patche does not apply to 2.2 and 3.0. Thanks!

> cqlsh: show partial trace if incomplete after max_trace_wait
> 
>
> Key: CASSANDRA-7645
> URL: https://issues.apache.org/jira/browse/CASSANDRA-7645
> Project: Cassandra
>  Issue Type: Improvement
>  Components: Tools
>Reporter: Tyler Hobbs
>Assignee: Carl Yeksigian
>Priority: Trivial
> Fix For: 2.1.x
>
>
> If a trace hasn't completed within {{max_trace_wait}}, cqlsh will say the 
> trace is unavailable and not show anything.  It (and the underlying python 
> driver) determines when the trace is complete by checking if the {{duration}} 
> column in {{system_traces.sessions}} is non-null.  If {{duration}} is null 
> but we still have some trace events when the timeout is hit, cqlsh should 
> print whatever trace events we have along with a warning about it being 
> incomplete.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10412) Could not initialize class org.apache.cassandra.config.DatabaseDescriptor

2015-10-06 Thread Paulo Motta (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10412?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945931#comment-14945931
 ] 

Paulo Motta commented on CASSANDRA-10412:
-

I think we should keep as few {{System.exit}} calls as possible spread around 
the code, and {{CassandraDaemon}} already has a {{runManaged}} field to supress 
{{System.exit()}} on errors, so we would be duplicating that functionality with 
{{DatabaseDescriptor.setEmbeddedService()}}. 

I don't think {{DatabaseDescriptor}} became more difficult to use, we just need 
to make sure it's used after {{DatabaseDescriptor.forceStaticInitialization()}} 
statement on {{CassandraDaemon.activate()}}. Maybe we could rename this method 
or add a notice to {{CassandraDaemon.activate()}} to make this more clear? Do 
you see any situation in which we would need to access {{DatabaseDescriptor}} 
before it's initialization on {{CassandraDaemon.activate()}}?

> Could not initialize class org.apache.cassandra.config.DatabaseDescriptor
> -
>
> Key: CASSANDRA-10412
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10412
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
> Environment: OS: Windows 2012 R2
> Java JRE: 1.8.0_51
>Reporter: Eric Simmons
>Assignee: Carl Yeksigian
>Priority: Minor
> Fix For: 2.2.x
>
> Attachments: cassandra-env.ps1, cassandra.yaml
>
>
> We are unable to start version 2.2.1 on our Windows 2012 R2 systems, however 
> we can use the same environment to start version 2.1.2
> I have attached our Cassandra.yaml and cassandra-env.ps1 file for reference.  
> I have also attached the system.log file displaying the error.
> I have also included an excerpt of the log file showing the stack trace of 
> the error.
> ERROR [ScheduledTasks:1] 2015-09-29 07:03:47,482 
> DebuggableThreadPoolExecutor.java:242 - Error in ThreadPoolExecutor
> java.lang.NoClassDefFoundError: Could not initialize class 
> org.apache.cassandra.config.DatabaseDescriptor
>   at 
> org.apache.cassandra.utils.JVMStabilityInspector.inspectThrowable(JVMStabilityInspector.java:57)
>  ~[apache-cassandra-2.2.1.jar:2.2.1]
>   at 
> org.apache.cassandra.concurrent.DebuggableScheduledThreadPoolExecutor$UncomplainingRunnable.run(DebuggableScheduledThreadPoolExecutor.java:122)
>  ~[apache-cassandra-2.2.1.jar:2.2.1]
>   at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) 
> ~[na:1.8.0_51]
>   at java.util.concurrent.FutureTask.runAndReset(Unknown Source) 
> ~[na:1.8.0_51]
>   at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(Unknown
>  Source) ~[na:1.8.0_51]
>   at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(Unknown
>  Source) ~[na:1.8.0_51]
>   at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) 
> [na:1.8.0_51]
>   at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) 
> [na:1.8.0_51]
>   at java.lang.Thread.run(Unknown Source) [na:1.8.0_51]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10412) Could not initialize class org.apache.cassandra.config.DatabaseDescriptor

2015-10-06 Thread Carl Yeksigian (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10412?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945797#comment-14945797
 ] 

Carl Yeksigian commented on CASSANDRA-10412:


It is an easy fix, but I'd rather fix the prior mistake so that it hopefully 
doesn't happen again. I provided the history behind how this broke to show that 
it's broken a few times and is very fragile. The changes have made 
{{DatabaseDescriptor}} more difficult to use, and it's to support a secondary 
use case whereas for our primary use case, it's become brittle.

> Could not initialize class org.apache.cassandra.config.DatabaseDescriptor
> -
>
> Key: CASSANDRA-10412
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10412
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
> Environment: OS: Windows 2012 R2
> Java JRE: 1.8.0_51
>Reporter: Eric Simmons
>Assignee: Carl Yeksigian
>Priority: Minor
> Fix For: 2.2.x
>
> Attachments: cassandra-env.ps1, cassandra.yaml
>
>
> We are unable to start version 2.2.1 on our Windows 2012 R2 systems, however 
> we can use the same environment to start version 2.1.2
> I have attached our Cassandra.yaml and cassandra-env.ps1 file for reference.  
> I have also attached the system.log file displaying the error.
> I have also included an excerpt of the log file showing the stack trace of 
> the error.
> ERROR [ScheduledTasks:1] 2015-09-29 07:03:47,482 
> DebuggableThreadPoolExecutor.java:242 - Error in ThreadPoolExecutor
> java.lang.NoClassDefFoundError: Could not initialize class 
> org.apache.cassandra.config.DatabaseDescriptor
>   at 
> org.apache.cassandra.utils.JVMStabilityInspector.inspectThrowable(JVMStabilityInspector.java:57)
>  ~[apache-cassandra-2.2.1.jar:2.2.1]
>   at 
> org.apache.cassandra.concurrent.DebuggableScheduledThreadPoolExecutor$UncomplainingRunnable.run(DebuggableScheduledThreadPoolExecutor.java:122)
>  ~[apache-cassandra-2.2.1.jar:2.2.1]
>   at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) 
> ~[na:1.8.0_51]
>   at java.util.concurrent.FutureTask.runAndReset(Unknown Source) 
> ~[na:1.8.0_51]
>   at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(Unknown
>  Source) ~[na:1.8.0_51]
>   at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(Unknown
>  Source) ~[na:1.8.0_51]
>   at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) 
> [na:1.8.0_51]
>   at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) 
> [na:1.8.0_51]
>   at java.lang.Thread.run(Unknown Source) [na:1.8.0_51]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10437) Remove offheap_objects option until 9472 re-introduces them

2015-10-06 Thread Ariel Weisberg (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10437?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945824#comment-14945824
 ] 

Ariel Weisberg commented on CASSANDRA-10437:


[dtests|http://cassci.datastax.com/view/Dev/view/aweisberg/job/aweisberg-C-10437-dtest/1/#showFailuresLink]
 and 
[utests|http://cassci.datastax.com/view/Dev/view/aweisberg/job/aweisberg-C-10437-testall/1/#showFailuresLink]

I went through the results and it seems to match trunk for the most part. None 
of the failures looked related to me. I kicked off another test-all and dtest 
just to see what repeats and what doesn't. Otherwise +1.


> Remove offheap_objects option until 9472 re-introduces them
> ---
>
> Key: CASSANDRA-10437
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10437
> Project: Cassandra
>  Issue Type: Bug
>Reporter: Sylvain Lebresne
>Assignee: Sylvain Lebresne
>Priority: Trivial
> Fix For: 3.0.0 rc2
>
> Attachments: 0001-Remove-offheap_objects-option.txt
>
>
> We need to send a meaningful error message if user try to use 
> {{offheap_objects}} in 3.0 since it's not supported currently (pending 
> CASSANDRA-9472). And document the current removal in the NEWS file.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10437) Remove offheap_objects option until 9472 re-introduces them

2015-10-06 Thread Ariel Weisberg (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10437?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945825#comment-14945825
 ] 

Ariel Weisberg commented on CASSANDRA-10437:


Oh, actually what is the doc impact here? Has that been taken care of?

> Remove offheap_objects option until 9472 re-introduces them
> ---
>
> Key: CASSANDRA-10437
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10437
> Project: Cassandra
>  Issue Type: Bug
>Reporter: Sylvain Lebresne
>Assignee: Sylvain Lebresne
>Priority: Trivial
> Fix For: 3.0.0 rc2
>
> Attachments: 0001-Remove-offheap_objects-option.txt
>
>
> We need to send a meaningful error message if user try to use 
> {{offheap_objects}} in 3.0 since it's not supported currently (pending 
> CASSANDRA-9472). And document the current removal in the NEWS file.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (CASSANDRA-7645) cqlsh: show partial trace if incomplete after max_trace_wait

2015-10-06 Thread Paulo Motta (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-7645?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945874#comment-14945874
 ] 

Paulo Motta edited comment on CASSANDRA-7645 at 10/6/15 9:39 PM:
-

\+1, could you provide 2.2+ patches? apparently cqlsh was moved from bin/cqlsh 
to bin/cqlsh.py, so the patche does not apply to 2.2 and 3.0. Thanks!


was (Author: pauloricardomg):
+1, could you provide 2.2+ patches? apparently cqlsh was moved from bin/cqlsh 
to bin/cqlsh.py, so the patche does not apply to 2.2 and 3.0. Thanks!

> cqlsh: show partial trace if incomplete after max_trace_wait
> 
>
> Key: CASSANDRA-7645
> URL: https://issues.apache.org/jira/browse/CASSANDRA-7645
> Project: Cassandra
>  Issue Type: Improvement
>  Components: Tools
>Reporter: Tyler Hobbs
>Assignee: Carl Yeksigian
>Priority: Trivial
> Fix For: 2.1.x
>
>
> If a trace hasn't completed within {{max_trace_wait}}, cqlsh will say the 
> trace is unavailable and not show anything.  It (and the underlying python 
> driver) determines when the trace is complete by checking if the {{duration}} 
> column in {{system_traces.sessions}} is non-null.  If {{duration}} is null 
> but we still have some trace events when the timeout is hit, cqlsh should 
> print whatever trace events we have along with a warning about it being 
> incomplete.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-10457) fix failing jmx_metrics dtest

2015-10-06 Thread Jim Witschey (JIRA)
Jim Witschey created CASSANDRA-10457:


 Summary: fix failing jmx_metrics dtest
 Key: CASSANDRA-10457
 URL: https://issues.apache.org/jira/browse/CASSANDRA-10457
 Project: Cassandra
  Issue Type: Sub-task
Reporter: Jim Witschey


{{jmxmetrics_test.py:TestJMXMetrics.begin_test}} is failing on CassCI; I've 
reproduced it on OpenStack as well. Here's the failure:

http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/jmxmetrics_test/TestJMXMetrics/begin_test/

My first thought is that a fix may be as simple as changing the parameters to 
stress between collecting the before and after metrics. I haven't debugged this 
further than reproducing it, though.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-10459) Fix largecolumn_test.py:TestLargeColumn.cleanup_test dtest

2015-10-06 Thread Jim Witschey (JIRA)
Jim Witschey created CASSANDRA-10459:


 Summary: Fix largecolumn_test.py:TestLargeColumn.cleanup_test dtest
 Key: CASSANDRA-10459
 URL: https://issues.apache.org/jira/browse/CASSANDRA-10459
 Project: Cassandra
  Issue Type: Sub-task
Reporter: Jim Witschey


This test on large columns is failing:

http://cassci.datastax.com/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/largecolumn_test/TestLargeColumn/cleanup_test/

and it's been failing for a while:

http://cassci.datastax.com/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/junit/largecolumn_test/TestLargeColumn/cleanup_test/history/

I've reproduced the failure on OpenStack, so it's not just a CassCI problem.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-10460) Fix materialized_views_test.py:TestMaterializedViews.complex_mv_select_statements_test

2015-10-06 Thread Jim Witschey (JIRA)
Jim Witschey created CASSANDRA-10460:


 Summary: Fix 
materialized_views_test.py:TestMaterializedViews.complex_mv_select_statements_test
 Key: CASSANDRA-10460
 URL: https://issues.apache.org/jira/browse/CASSANDRA-10460
 Project: Cassandra
  Issue Type: Sub-task
Reporter: Jim Witschey
Assignee: Carl Yeksigian


This test:

http://cassci.datastax.com/view/cassandra-3.0/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/materialized_views_test/TestMaterializedViews/complex_mv_select_statements_test/

fails on CassCI and when I run it manually on OpenStack. It's been failing for 
a while:

http://cassci.datastax.com/view/cassandra-3.0/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/materialized_views_test/TestMaterializedViews/complex_mv_select_statements_test/history/

Assigning to [~carlyeks] for triage; feel free to reassign.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-9126) java.lang.RuntimeException: Last written key DecoratedKey >= current key DecoratedKey

2015-10-06 Thread Marcus Eriksson (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-9126?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14944705#comment-14944705
 ] 

Marcus Eriksson commented on CASSANDRA-9126:


[~mambocab] it is likely this was due to either a corrupt sstable or overlap in 
the LCS leveling, scrubbing and restarting solves both

> java.lang.RuntimeException: Last written key DecoratedKey >= current key 
> DecoratedKey
> -
>
> Key: CASSANDRA-9126
> URL: https://issues.apache.org/jira/browse/CASSANDRA-9126
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
>Reporter: srinivasu gottipati
>Priority: Critical
> Fix For: 2.1.x
>
> Attachments: cassandra-system.log
>
>
> Cassandra V: 2.0.14,
> Getting the following exceptions while trying to compact (I see this issue 
> was raised in earlier versions and marked as closed. However it still appears 
> in 2.0.14). In our case, compaction is not getting succeeded and keep failing 
> with this error.:
> {code}java.lang.RuntimeException: Last written key 
> DecoratedKey(3462767860784856708, 
> 354038323137333038305f3330325f31355f474d4543454f) >= current key 
> DecoratedKey(3462334604624154281, 
> 354036333036353334315f3336315f31355f474d4543454f) writing into {code}
> ...
> Stacktrace:{code}
>   at 
> org.apache.cassandra.io.sstable.SSTableWriter.beforeAppend(SSTableWriter.java:143)
>   at 
> org.apache.cassandra.io.sstable.SSTableWriter.append(SSTableWriter.java:166)
>   at 
> org.apache.cassandra.db.compaction.CompactionTask.runMayThrow(CompactionTask.java:167)
>   at 
> org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28)
>   at 
> org.apache.cassandra.db.compaction.CompactionTask.executeInternal(CompactionTask.java:60)
>   at 
> org.apache.cassandra.db.compaction.AbstractCompactionTask.execute(AbstractCompactionTask.java:59)
>   at 
> org.apache.cassandra.db.compaction.CompactionManager$BackgroundCompactionTask.run(CompactionManager.java:198)
>   at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>   at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745){code}
> Any help is greatly appreciated



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-10453) Fix compression_cql_options_test dtest

2015-10-06 Thread Jim Witschey (JIRA)
Jim Witschey created CASSANDRA-10453:


 Summary: Fix compression_cql_options_test dtest
 Key: CASSANDRA-10453
 URL: https://issues.apache.org/jira/browse/CASSANDRA-10453
 Project: Cassandra
  Issue Type: Sub-task
Reporter: Jim Witschey
 Fix For: 3.0.0 rc2


This has been failing hard for a while:

http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/206/testReport/junit/compression_test/TestCompression/compression_cql_options_test/history/

This regression, like CASSANDRA-10451, started when the recent log changes were 
introduced:

http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/206/changes

I also can't reproduce on OpenStack. Pinging [~pauloricardomg] and [~blerer] 
again.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (CASSANDRA-10453) Fix compression_cql_options_test dtest

2015-10-06 Thread Paulo Motta (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10453?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paulo Motta reassigned CASSANDRA-10453:
---

Assignee: Paulo Motta

> Fix compression_cql_options_test dtest
> --
>
> Key: CASSANDRA-10453
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10453
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Paulo Motta
> Fix For: 3.0.0 rc2
>
>
> This has been failing hard for a while:
> http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/206/testReport/junit/compression_test/TestCompression/compression_cql_options_test/history/
> This regression, like CASSANDRA-10451, started when the recent log changes 
> were introduced:
> http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/206/changes
> I also can't reproduce on OpenStack. Pinging [~pauloricardomg] and [~blerer] 
> again.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10448) "Unknown type 0" Stream failure on Repair

2015-10-06 Thread Omri Iluz (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10448?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Omri Iluz updated CASSANDRA-10448:
--
Attachment: casslogs.txt

System and debug logs attached from both nodes.
This is from another repair session today, after a scrub.

these logs are filtered by the stream id, let me know if the full logs can be 
helpful (but they are big)

> "Unknown type 0" Stream failure on Repair
> -
>
> Key: CASSANDRA-10448
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10448
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
> Environment: Cassandra 2.2.2
> 5 Nodes in Google Compute Engine
> Java 1.8.0_60
>Reporter: Omri Iluz
>Assignee: Yuki Morishita
> Fix For: 2.2.x
>
> Attachments: casslogs.txt
>
>
> While running repair after upgrading to 2.2.2 I am getting many stream fail 
> errors:
> {noformat}
> [2015-10-05 23:52:30,353] Repair session 4c181051-6bbb-11e5-acdb-d9a8bbd39330 
> for range (59694553044959221,86389982480621619] failed with error [repair 
> #4c181051-6bbb-11e5-acdb-d9a8bbd39330 on px/acti
> vities, (59694553044959221,86389982480621619]] Sync failed between 
> /10.240.81.104 and /10.240.134.221 (progress: 4%)
> {noformat}
> Logs from both sides of the stream:
> Sides 1 -
> {noformat}
> INFO  [STREAM-INIT-/10.240.81.104:52722] 2015-10-05 23:52:30,063 
> StreamResultFuture.java:111 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550 
> ID#0] Creating new streaming plan for Repair
> INFO  [STREAM-INIT-/10.240.81.104:52722] 2015-10-05 23:52:30,063 
> StreamResultFuture.java:118 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550, 
> ID#0] Received streaming plan for Repair
> INFO  [STREAM-INIT-/10.240.81.104:52723] 2015-10-05 23:52:30,063 
> StreamResultFuture.java:118 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550, 
> ID#0] Received streaming plan for Repair
> INFO  [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,098 
> StreamResultFuture.java:168 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550 
> ID#0] Prepare completed. Receiving 13 files(517391317 bytes), sending 10 
> files(469491729 bytes)
> ERROR [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,234 
> StreamSession.java:524 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Streaming error occurred
> java.lang.IllegalArgumentException: Unknown type 0
>   at 
> org.apache.cassandra.streaming.messages.StreamMessage$Type.get(StreamMessage.java:96)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:57)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at java.lang.Thread.run(Thread.java:745) [na:1.8.0_60]
> INFO  [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,302 
> StreamResultFuture.java:182 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Session with /10.240.81.104 is complete
> WARN  [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,302 
> StreamResultFuture.java:209 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Stream failed
> {noformat}
> Side 2 -
> {noformat}
> INFO  [AntiEntropyStage:1] 2015-10-05 23:52:30,060 StreamResultFuture.java:86 
> - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] Executing streaming plan for 
> Repair
> INFO  [StreamConnectionEstablisher:6] 2015-10-05 23:52:30,061 
> StreamSession.java:232 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Starting streaming to /10.240.134.221
> INFO  [StreamConnectionEstablisher:6] 2015-10-05 23:52:30,063 
> StreamCoordinator.java:213 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550, 
> ID#0] Beginning stream session with /10.240.134.221
> INFO  [STREAM-IN-/10.240.134.221] 2015-10-05 23:52:30,098 
> StreamResultFuture.java:168 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550 
> ID#0] Prepare completed. Receiving 10 files(469491729 bytes), sending 13 
> files(517391317 bytes)
> INFO  [STREAM-IN-/10.240.134.221] 2015-10-05 23:52:30,349 
> StreamResultFuture.java:182 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Session with /10.240.134.221 is complete
> ERROR [STREAM-OUT-/10.240.134.221] 2015-10-05 23:52:30,349 
> StreamSession.java:524 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Streaming error occurred
> org.apache.cassandra.io.FSReadError: java.io.IOException: Broken pipe
>   at 
> org.apache.cassandra.io.util.ChannelProxy.transferTo(ChannelProxy.java:144) 
> ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> org.apache.cassandra.streaming.compress.CompressedStreamWriter$1.apply(CompressedStreamWriter.java:79)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> 

[jira] [Commented] (CASSANDRA-7392) Abort in-progress queries that time out

2015-10-06 Thread Ariel Weisberg (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-7392?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14946161#comment-14946161
 ] 

Ariel Weisberg commented on CASSANDRA-7392:
---

bq. Thank you for your hard work 
Ditto. +1 then.

> Abort in-progress queries that time out
> ---
>
> Key: CASSANDRA-7392
> URL: https://issues.apache.org/jira/browse/CASSANDRA-7392
> Project: Cassandra
>  Issue Type: New Feature
>  Components: Core
>Reporter: Jonathan Ellis
>Assignee: Stefania
>Priority: Critical
> Fix For: 3.x
>
>
> Currently we drop queries that time out before we get to them (because node 
> is overloaded) but not queries that time out while being processed.  
> (Particularly common for index queries on data that shouldn't be indexed.)  
> Adding the latter and logging when we have to interrupt one gets us a poor 
> man's "slow query log" for free.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-7392) Abort in-progress queries that time out

2015-10-06 Thread Stefania (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-7392?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14946104#comment-14946104
 ] 

Stefania commented on CASSANDRA-7392:
-

Thank you for your hard work Ariel. I have rerun test-all on trunk and it 
completed successfully this time, test results look OK to me. 

> Abort in-progress queries that time out
> ---
>
> Key: CASSANDRA-7392
> URL: https://issues.apache.org/jira/browse/CASSANDRA-7392
> Project: Cassandra
>  Issue Type: New Feature
>  Components: Core
>Reporter: Jonathan Ellis
>Assignee: Stefania
>Priority: Critical
> Fix For: 3.x
>
>
> Currently we drop queries that time out before we get to them (because node 
> is overloaded) but not queries that time out while being processed.  
> (Particularly common for index queries on data that shouldn't be indexed.)  
> Adding the latter and logging when we have to interrupt one gets us a poor 
> man's "slow query log" for free.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-10461) Fix sstableverify_test dtest

2015-10-06 Thread Jim Witschey (JIRA)
Jim Witschey created CASSANDRA-10461:


 Summary: Fix sstableverify_test dtest
 Key: CASSANDRA-10461
 URL: https://issues.apache.org/jira/browse/CASSANDRA-10461
 Project: Cassandra
  Issue Type: Sub-task
Reporter: Jim Witschey


The dtest for sstableverify is failing:

http://cassci.datastax.com/view/cassandra-3.0/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/offline_tools_test/TestOfflineTools/sstableverify_test/

It fails in the same way when I run it on OpenStack, so I don't think it's just 
a CassCI problem.

[~slebresne] Looks like you made changes to this test recently:

https://github.com/riptano/cassandra-dtest/commit/51ab085f21e01cc8e5ad88a277cb4a43abd3f880

Could you have a look at the failure? I'm assigning you for triage, but feel 
free to reassign.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-10462) Fix failing upgrade test

2015-10-06 Thread Jim Witschey (JIRA)
Jim Witschey created CASSANDRA-10462:


 Summary: Fix failing upgrade test
 Key: CASSANDRA-10462
 URL: https://issues.apache.org/jira/browse/CASSANDRA-10462
 Project: Cassandra
  Issue Type: Sub-task
Reporter: Jim Witschey


The 
{{upgrade_tests/paging_test.py:TestPagingWithDeletions.test_failure_threshold_deletions}}
 dtest fails on CassCI:

http://cassci.datastax.com/view/cassandra-3.0/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/upgrade_tests.paging_test/TestPagingWithDeletions/test_failure_threshold_deletions/

and has failed for a while:

http://cassci.datastax.com/view/cassandra-3.0/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/upgrade_tests.paging_test/TestPagingWithDeletions/test_failure_threshold_deletions/history/

It fails identically when I run it manually on OpenStack, so I don't think it's 
a CassCI problem.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-10463) Fix failing compaction tests

2015-10-06 Thread Jim Witschey (JIRA)
Jim Witschey created CASSANDRA-10463:


 Summary: Fix failing compaction tests
 Key: CASSANDRA-10463
 URL: https://issues.apache.org/jira/browse/CASSANDRA-10463
 Project: Cassandra
  Issue Type: Sub-task
Reporter: Jim Witschey
Assignee: Yuki Morishita


A bunch of compaction tests started failing after some changes to how logs are 
handled in ccm:

- 
compaction_test.TestCompaction_with_DateTieredCompactionStrategy.compaction_throughput_test
- 
compaction_test.TestCompaction_with_DateTieredCompactionStrategy.data_size_test
- 
compaction_test.TestCompaction_with_DateTieredCompactionStrategy.disable_autocompaction_alter_and_nodetool_test
- 
compaction_test.TestCompaction_with_DateTieredCompactionStrategy.disable_autocompaction_alter_test
- 
compaction_test.TestCompaction_with_DateTieredCompactionStrategy.disable_autocompaction_nodetool_test
- 
compaction_test.TestCompaction_with_DateTieredCompactionStrategy.disable_autocompaction_schema_test
- 
compaction_test.TestCompaction_with_DateTieredCompactionStrategy.sstable_deletion_test
- 
compaction_test.TestCompaction_with_LeveledCompactionStrategy.compaction_throughput_test
- compaction_test.TestCompaction_with_LeveledCompactionStrategy.data_size_test
- 
compaction_test.TestCompaction_with_LeveledCompactionStrategy.disable_autocompaction_alter_and_nodetool_test
- 
compaction_test.TestCompaction_with_LeveledCompactionStrategy.disable_autocompaction_alter_test
- 
compaction_test.TestCompaction_with_LeveledCompactionStrategy.disable_autocompaction_nodetool_test
- 
compaction_test.TestCompaction_with_LeveledCompactionStrategy.disable_autocompaction_schema_test
- 
compaction_test.TestCompaction_with_LeveledCompactionStrategy.sstable_deletion_test
- 
compaction_test.TestCompaction_with_SizeTieredCompactionStrategy.compaction_throughput_test
- 
compaction_test.TestCompaction_with_SizeTieredCompactionStrategy.data_size_test
- 
compaction_test.TestCompaction_with_SizeTieredCompactionStrategy.disable_autocompaction_alter_and_nodetool_test
- 
compaction_test.TestCompaction_with_SizeTieredCompactionStrategy.disable_autocompaction_alter_test
- 
compaction_test.TestCompaction_with_SizeTieredCompactionStrategy.disable_autocompaction_nodetool_test
- 
compaction_test.TestCompaction_with_SizeTieredCompactionStrategy.disable_autocompaction_schema_test
- 
compaction_test.TestCompaction_with_SizeTieredCompactionStrategy.sstable_deletion_test

I haven't dug into it much, but it looks like the following tests:

- compaction_throughput_test
- data_size_test
- disable_autocompaction_alter_and_nodetool_test
- disable_autocompaction_alter_test
- disable_autocompaction_nodetool_test
- disable_autocompaction_schema_test
- sstable_deletion_test

grep logs to check if operations happened successfully, and that something in 
the new log-handling code broke this.

I'm assigning [~yukim], since I believe you wrote the changes to ccm, but feel 
free to find a new assignee. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (CASSANDRA-10453) Fix compression_cql_options_test dtest

2015-10-06 Thread Jim Witschey (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10453?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jim Witschey resolved CASSANDRA-10453.
--
Resolution: Fixed

> Fix compression_cql_options_test dtest
> --
>
> Key: CASSANDRA-10453
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10453
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Paulo Motta
> Fix For: 3.0.0 rc2
>
>
> This has been failing hard for a while:
> http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/206/testReport/junit/compression_test/TestCompression/compression_cql_options_test/history/
> This regression, like CASSANDRA-10451, started when the recent log changes 
> were introduced:
> http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/206/changes
> I also can't reproduce on OpenStack. Pinging [~pauloricardomg] and [~blerer] 
> again.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10453) Fix compression_cql_options_test dtest

2015-10-06 Thread Jim Witschey (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945986#comment-14945986
 ] 

Jim Witschey commented on CASSANDRA-10453:
--

Sorry to have bugged you on this -- it was already fixed in ccm and passes on 
CassCI now:

http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/226/testReport/junit/compression_test/TestCompression/compression_cql_options_test/

Closing.

> Fix compression_cql_options_test dtest
> --
>
> Key: CASSANDRA-10453
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10453
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Paulo Motta
> Fix For: 3.0.0 rc2
>
>
> This has been failing hard for a while:
> http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/206/testReport/junit/compression_test/TestCompression/compression_cql_options_test/history/
> This regression, like CASSANDRA-10451, started when the recent log changes 
> were introduced:
> http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/206/changes
> I also can't reproduce on OpenStack. Pinging [~pauloricardomg] and [~blerer] 
> again.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (CASSANDRA-10456) fix deprecated_repair_test dtests

2015-10-06 Thread Jim Witschey (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10456?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jim Witschey resolved CASSANDRA-10456.
--
Resolution: Fixed

> fix deprecated_repair_test dtests
> -
>
> Key: CASSANDRA-10456
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10456
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Jim Witschey
> Fix For: 3.0.0 rc2
>
>
> All the deprecated_repair_test dtests are failing on CassCI. I believe the 
> immediate cause is some recent changes to logs and how ccm handles them, like 
> the failures in CASSANDRA-10451 and CASSANDRA-10453. Test failures on CassCI 
> here:
> http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/deprecated_repair_test/
> I'm assigning myself to track this, in case there are other failures to deal 
> with.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10456) fix deprecated_repair_test dtests

2015-10-06 Thread Jim Witschey (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10456?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945987#comment-14945987
 ] 

Jim Witschey commented on CASSANDRA-10456:
--

Yes, this is fixed:

http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/226/testReport/deprecated_repair_test/

Closing.

> fix deprecated_repair_test dtests
> -
>
> Key: CASSANDRA-10456
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10456
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Jim Witschey
> Fix For: 3.0.0 rc2
>
>
> All the deprecated_repair_test dtests are failing on CassCI. I believe the 
> immediate cause is some recent changes to logs and how ccm handles them, like 
> the failures in CASSANDRA-10451 and CASSANDRA-10453. Test failures on CassCI 
> here:
> http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/deprecated_repair_test/
> I'm assigning myself to track this, in case there are other failures to deal 
> with.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10455) Fix pep8 compliance in cqlsh.py

2015-10-06 Thread Stefania (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10455?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14946010#comment-14946010
 ] 

Stefania commented on CASSANDRA-10455:
--

+1

> Fix pep8 compliance in cqlsh.py
> ---
>
> Key: CASSANDRA-10455
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10455
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Philip Thompson
>  Labels: cqlsh
> Fix For: 3.0.0 rc2
>
>
> cqlsh has fallen out of pep8 compliance:
> http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/cqlsh_tests.cqlsh_tests/TestCqlsh/test_pep8_compliance/
> [~philipthompson] has said off-JIRA that he'll fix it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10461) Fix sstableverify_test dtest

2015-10-06 Thread Jim Witschey (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10461?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jim Witschey updated CASSANDRA-10461:
-
Assignee: Sylvain Lebresne

> Fix sstableverify_test dtest
> 
>
> Key: CASSANDRA-10461
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10461
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Sylvain Lebresne
> Fix For: 3.0.0 rc2
>
>
> The dtest for sstableverify is failing:
> http://cassci.datastax.com/view/cassandra-3.0/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/offline_tools_test/TestOfflineTools/sstableverify_test/
> It fails in the same way when I run it on OpenStack, so I don't think it's 
> just a CassCI problem.
> [~slebresne] Looks like you made changes to this test recently:
> https://github.com/riptano/cassandra-dtest/commit/51ab085f21e01cc8e5ad88a277cb4a43abd3f880
> Could you have a look at the failure? I'm assigning you for triage, but feel 
> free to reassign.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10451) Fix batch_test.TestBatch.logged_batch_gcgs_below_threshold_{multi,single}_table_test on CassCI

2015-10-06 Thread Jim Witschey (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945982#comment-14945982
 ] 

Jim Witschey commented on CASSANDRA-10451:
--

Fixed by the changes to ccm:

http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/batch_test/TestBatch/logged_batch_gcgs_below_threshold_single_table_test/history/

Closing.

> Fix 
> batch_test.TestBatch.logged_batch_gcgs_below_threshold_{multi,single}_table_test
>  on CassCI
> --
>
> Key: CASSANDRA-10451
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10451
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Paulo Motta
> Fix For: 3.0.0 rc2
>
>
> This is a recent regression:
> http://cassci.datastax.com/view/trunk/job/trunk_dtest/lastCompletedBuild/testReport/batch_test/TestBatch/logged_batch_gcgs_below_threshold_multi_table_test/history/
> Based on [when it started 
> failing|http://cassci.datastax.com/view/trunk/job/trunk_dtest/623/changes], 
> it looks like the regression was introduced by [the recent logging 
> changes|https://github.com/apache/cassandra/commit/4a849efeb7c7c1a54bc12094d2d6a9f3f008a2fa].
>  Pinging [~blerer] and [~pauloricardomg] to have a look.
> My initial guess was that the logs the test depends on were moved to the new 
> debug log. However, the tests passes when I run it manually on openstack, so 
> it could just be a configuration issue on CassCI.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (CASSANDRA-10451) Fix batch_test.TestBatch.logged_batch_gcgs_below_threshold_{multi,single}_table_test on CassCI

2015-10-06 Thread Jim Witschey (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10451?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jim Witschey resolved CASSANDRA-10451.
--
Resolution: Fixed

> Fix 
> batch_test.TestBatch.logged_batch_gcgs_below_threshold_{multi,single}_table_test
>  on CassCI
> --
>
> Key: CASSANDRA-10451
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10451
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Paulo Motta
> Fix For: 3.0.0 rc2
>
>
> This is a recent regression:
> http://cassci.datastax.com/view/trunk/job/trunk_dtest/lastCompletedBuild/testReport/batch_test/TestBatch/logged_batch_gcgs_below_threshold_multi_table_test/history/
> Based on [when it started 
> failing|http://cassci.datastax.com/view/trunk/job/trunk_dtest/623/changes], 
> it looks like the regression was introduced by [the recent logging 
> changes|https://github.com/apache/cassandra/commit/4a849efeb7c7c1a54bc12094d2d6a9f3f008a2fa].
>  Pinging [~blerer] and [~pauloricardomg] to have a look.
> My initial guess was that the logs the test depends on were moved to the new 
> debug log. However, the tests passes when I run it manually on openstack, so 
> it could just be a configuration issue on CassCI.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10449) OOM on bootstrap due to long GC pause

2015-10-06 Thread Robbie Strickland (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10449?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robbie Strickland updated CASSANDRA-10449:
--
Attachment: system.log.10-05

> OOM on bootstrap due to long GC pause
> -
>
> Key: CASSANDRA-10449
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10449
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
> Environment: Ubuntu 14.04, AWS
>Reporter: Robbie Strickland
>  Labels: gc
> Fix For: 2.1.x
>
> Attachments: system.log.10-05
>
>
> I have a 20-node cluster (i2.4xlarge) with vnodes (default of 256) and 
> 500-700GB per node.  SSTable counts are <10 per table.  I am attempting to 
> provision additional nodes, but bootstrapping OOMs every time after about 10 
> hours with a sudden long GC pause:
> {noformat}
> INFO  [Service Thread] 2015-10-05 23:33:33,373 GCInspector.java:252 - G1 Old 
> Generation GC in 1586126ms.  G1 Old Gen: 49213756976 -> 49072277176;
> ...
> ERROR [MemtableFlushWriter:454] 2015-10-05 23:33:33,380 
> CassandraDaemon.java:223 - Exception in thread 
> Thread[MemtableFlushWriter:454,5,main]
> java.lang.OutOfMemoryError: Java heap space
> {noformat}
> I have tried increasing max heap to 48G just to get through the bootstrap, to 
> no avail.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10448) "Unknown type 0" Stream failure on Repair

2015-10-06 Thread Yuki Morishita (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10448?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945077#comment-14945077
 ] 

Yuki Morishita commented on CASSANDRA-10448:


This looks like network error.
We are working on to introduce auto reconnect in the future (CASSANDRA-8621).

> "Unknown type 0" Stream failure on Repair
> -
>
> Key: CASSANDRA-10448
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10448
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
> Environment: Cassandra 2.2.2
> 5 Nodes in Google Compute Engine
> Java 1.8.0_60
>Reporter: Omri Iluz
>Assignee: Yuki Morishita
> Fix For: 2.2.x
>
>
> While running repair after upgrading to 2.2.2 I am getting many stream fail 
> errors:
> {noformat}
> [2015-10-05 23:52:30,353] Repair session 4c181051-6bbb-11e5-acdb-d9a8bbd39330 
> for range (59694553044959221,86389982480621619] failed with error [repair 
> #4c181051-6bbb-11e5-acdb-d9a8bbd39330 on px/acti
> vities, (59694553044959221,86389982480621619]] Sync failed between 
> /10.240.81.104 and /10.240.134.221 (progress: 4%)
> {noformat}
> Logs from both sides of the stream:
> Sides 1 -
> {noformat}
> INFO  [STREAM-INIT-/10.240.81.104:52722] 2015-10-05 23:52:30,063 
> StreamResultFuture.java:111 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550 
> ID#0] Creating new streaming plan for Repair
> INFO  [STREAM-INIT-/10.240.81.104:52722] 2015-10-05 23:52:30,063 
> StreamResultFuture.java:118 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550, 
> ID#0] Received streaming plan for Repair
> INFO  [STREAM-INIT-/10.240.81.104:52723] 2015-10-05 23:52:30,063 
> StreamResultFuture.java:118 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550, 
> ID#0] Received streaming plan for Repair
> INFO  [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,098 
> StreamResultFuture.java:168 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550 
> ID#0] Prepare completed. Receiving 13 files(517391317 bytes), sending 10 
> files(469491729 bytes)
> ERROR [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,234 
> StreamSession.java:524 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Streaming error occurred
> java.lang.IllegalArgumentException: Unknown type 0
>   at 
> org.apache.cassandra.streaming.messages.StreamMessage$Type.get(StreamMessage.java:96)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:57)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at java.lang.Thread.run(Thread.java:745) [na:1.8.0_60]
> INFO  [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,302 
> StreamResultFuture.java:182 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Session with /10.240.81.104 is complete
> WARN  [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,302 
> StreamResultFuture.java:209 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Stream failed
> {noformat}
> Side 2 -
> {noformat}
> INFO  [AntiEntropyStage:1] 2015-10-05 23:52:30,060 StreamResultFuture.java:86 
> - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] Executing streaming plan for 
> Repair
> INFO  [StreamConnectionEstablisher:6] 2015-10-05 23:52:30,061 
> StreamSession.java:232 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Starting streaming to /10.240.134.221
> INFO  [StreamConnectionEstablisher:6] 2015-10-05 23:52:30,063 
> StreamCoordinator.java:213 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550, 
> ID#0] Beginning stream session with /10.240.134.221
> INFO  [STREAM-IN-/10.240.134.221] 2015-10-05 23:52:30,098 
> StreamResultFuture.java:168 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550 
> ID#0] Prepare completed. Receiving 10 files(469491729 bytes), sending 13 
> files(517391317 bytes)
> INFO  [STREAM-IN-/10.240.134.221] 2015-10-05 23:52:30,349 
> StreamResultFuture.java:182 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Session with /10.240.134.221 is complete
> ERROR [STREAM-OUT-/10.240.134.221] 2015-10-05 23:52:30,349 
> StreamSession.java:524 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] 
> Streaming error occurred
> org.apache.cassandra.io.FSReadError: java.io.IOException: Broken pipe
>   at 
> org.apache.cassandra.io.util.ChannelProxy.transferTo(ChannelProxy.java:144) 
> ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> org.apache.cassandra.streaming.compress.CompressedStreamWriter$1.apply(CompressedStreamWriter.java:79)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> org.apache.cassandra.streaming.compress.CompressedStreamWriter$1.apply(CompressedStreamWriter.java:76)
>  ~[apache-cassandra-2.2.2.jar:2.2.2]
>   at 
> 

[jira] [Assigned] (CASSANDRA-10413) Replaying materialized view updates from commitlog after node decommission crashes Cassandra

2015-10-06 Thread Joel Knighton (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10413?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joel Knighton reassigned CASSANDRA-10413:
-

Assignee: Joel Knighton  (was: T Jake Luciani)

> Replaying materialized view updates from commitlog after node decommission 
> crashes Cassandra
> 
>
> Key: CASSANDRA-10413
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10413
> Project: Cassandra
>  Issue Type: Bug
>Reporter: Joel Knighton
>Assignee: Joel Knighton
>Priority: Critical
> Fix For: 3.0.0 rc2
>
> Attachments: n1.log, n2.log, n3.log, n4.log, n5.log
>
>
> This issue is reproducible through a Jepsen test, runnable as
> {code}
> lein with-profile +trunk test :only 
> cassandra.mv-test/mv-crash-subset-decommission
> {code}
> This test crashes/restarts nodes while decommissioning nodes. These actions 
> are not coordinated.
> In [10164|https://issues.apache.org/jira/browse/CASSANDRA-10164], we 
> introduced a change to re-apply materialized view updates on commitlog replay.
> Some nodes, upon restart, will crash in commitlog replay. They throw the 
> "Trying to get the view natural endpoint on a non-data replica" runtime 
> exception in getViewNaturalEndpoint. I added logging to 
> getViewNaturalEndpoint to show the results of 
> replicationStrategy.getNaturalEndpoints for the baseToken and viewToken.
> It can be seen that these problems occur when the baseEndpoints and 
> viewEndpoints are identical but do not contain the broadcast address of the 
> local node.
> For example, a node at 10.0.0.5 crashes on replay of a write whose base token 
> and view token replicas are both [10.0.0.2, 10.0.0.4, 10.0.0.6]. It seems we 
> try to guard against this by considering pendingEndpoints for the viewToken, 
> but this does not appear to be sufficient.
> I've attached the system.logs for a test run with added logging. In the 
> attached logs, n1 is at 10.0.0.2, n2 is at 10.0.0.3, and so on. 10.0.0.6/n5 
> is the decommissioned node.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10443) CQLSStableWriter example fails on 3.0rc1

2015-10-06 Thread Yuki Morishita (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10443?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yuki Morishita updated CASSANDRA-10443:
---
Reviewer: Yuki Morishita

> CQLSStableWriter example fails on 3.0rc1
> 
>
> Key: CASSANDRA-10443
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10443
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core, Tools
>Reporter: Jonathan Shook
>Assignee: Carl Yeksigian
> Fix For: 3.0.0 rc2
>
>
> CQLSSTableWriter which works with 2.2.1 does not work with 3.0rc1.
> Something like https://github.com/yukim/cassandra-bulkload-example should be 
> added to the test suite.
> Exception in thread "main" java.lang.RuntimeException: 
> java.lang.ExceptionInInitializerError
>   at 
> org.apache.cassandra.io.sstable.SSTableSimpleUnsortedWriter.close(SSTableSimpleUnsortedWriter.java:136)
>   at 
> org.apache.cassandra.io.sstable.CQLSSTableWriter.close(CQLSSTableWriter.java:274)
>   at com.metawiring.sandbox.BulkLoadExample.main(BulkLoadExample.java:160)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:497)
>   at com.intellij.rt.execution.application.AppMain.main(AppMain.java:140)
> Caused by: java.lang.ExceptionInInitializerError
>   at org.apache.cassandra.db.Keyspace.initCf(Keyspace.java:372)
>   at org.apache.cassandra.db.Keyspace.(Keyspace.java:309)
>   at org.apache.cassandra.db.Keyspace.open(Keyspace.java:133)
>   at org.apache.cassandra.db.Keyspace.open(Keyspace.java:110)
>   at 
> org.apache.cassandra.io.sstable.SSTableTxnWriter.create(SSTableTxnWriter.java:97)
>   at 
> org.apache.cassandra.io.sstable.AbstractSSTableSimpleWriter.createWriter(AbstractSSTableSimpleWriter.java:63)
>   at 
> org.apache.cassandra.io.sstable.SSTableSimpleUnsortedWriter$DiskWriter.run(SSTableSimpleUnsortedWriter.java:206)
> Caused by: java.lang.NullPointerException
>   at 
> org.apache.cassandra.config.DatabaseDescriptor.getFlushWriters(DatabaseDescriptor.java:1153)
>   at 
> org.apache.cassandra.db.ColumnFamilyStore.(ColumnFamilyStore.java:116)
>   ... 7 more



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10365) Consider storing types by their CQL names in schema tables instead of fully-qualified internal class names

2015-10-06 Thread Bulat Shakirzyanov (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10365?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945353#comment-14945353
 ] 

Bulat Shakirzyanov commented on CASSANDRA-10365:


I think this is a very useful metadata to include. My only suggestion is to 
keep the old representation (fqcn) as well. Currently drivers use this 
representation to detect type information. Parsing CQL-only representation 
might not be robust for the reasons mentioned in 
https://issues.apache.org/jira/browse/CASSANDRA-6717#comment-14791788.


> Consider storing types by their CQL names in schema tables instead of 
> fully-qualified internal class names
> --
>
> Key: CASSANDRA-10365
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10365
> Project: Cassandra
>  Issue Type: Improvement
>Reporter: Aleksey Yeschenko
>Assignee: Aleksey Yeschenko
>  Labels: client-impacting
> Fix For: 3.0.0 rc2
>
>
> Consider saving CQL types names for column, UDF/UDA arguments and return 
> types, and UDT components.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10413) Replaying materialized view updates from commitlog after node decommission crashes Cassandra

2015-10-06 Thread Joshua McKenzie (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10413?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joshua McKenzie updated CASSANDRA-10413:

Reviewer: Joel Knighton

> Replaying materialized view updates from commitlog after node decommission 
> crashes Cassandra
> 
>
> Key: CASSANDRA-10413
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10413
> Project: Cassandra
>  Issue Type: Bug
>Reporter: Joel Knighton
>Assignee: T Jake Luciani
>Priority: Critical
> Fix For: 3.0.0 rc2
>
> Attachments: n1.log, n2.log, n3.log, n4.log, n5.log
>
>
> This issue is reproducible through a Jepsen test, runnable as
> {code}
> lein with-profile +trunk test :only 
> cassandra.mv-test/mv-crash-subset-decommission
> {code}
> This test crashes/restarts nodes while decommissioning nodes. These actions 
> are not coordinated.
> In [10164|https://issues.apache.org/jira/browse/CASSANDRA-10164], we 
> introduced a change to re-apply materialized view updates on commitlog replay.
> Some nodes, upon restart, will crash in commitlog replay. They throw the 
> "Trying to get the view natural endpoint on a non-data replica" runtime 
> exception in getViewNaturalEndpoint. I added logging to 
> getViewNaturalEndpoint to show the results of 
> replicationStrategy.getNaturalEndpoints for the baseToken and viewToken.
> It can be seen that these problems occur when the baseEndpoints and 
> viewEndpoints are identical but do not contain the broadcast address of the 
> local node.
> For example, a node at 10.0.0.5 crashes on replay of a write whose base token 
> and view token replicas are both [10.0.0.2, 10.0.0.4, 10.0.0.6]. It seems we 
> try to guard against this by considering pendingEndpoints for the viewToken, 
> but this does not appear to be sufficient.
> I've attached the system.logs for a test run with added logging. In the 
> attached logs, n1 is at 10.0.0.2, n2 is at 10.0.0.3, and so on. 10.0.0.6/n5 
> is the decommissioned node.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10413) Replaying materialized view updates from commitlog after node decommission crashes Cassandra

2015-10-06 Thread Joel Knighton (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10413?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joel Knighton updated CASSANDRA-10413:
--
Reviewer:   (was: Joel Knighton)

> Replaying materialized view updates from commitlog after node decommission 
> crashes Cassandra
> 
>
> Key: CASSANDRA-10413
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10413
> Project: Cassandra
>  Issue Type: Bug
>Reporter: Joel Knighton
>Assignee: Joel Knighton
>Priority: Critical
> Fix For: 3.0.0 rc2
>
> Attachments: n1.log, n2.log, n3.log, n4.log, n5.log
>
>
> This issue is reproducible through a Jepsen test, runnable as
> {code}
> lein with-profile +trunk test :only 
> cassandra.mv-test/mv-crash-subset-decommission
> {code}
> This test crashes/restarts nodes while decommissioning nodes. These actions 
> are not coordinated.
> In [10164|https://issues.apache.org/jira/browse/CASSANDRA-10164], we 
> introduced a change to re-apply materialized view updates on commitlog replay.
> Some nodes, upon restart, will crash in commitlog replay. They throw the 
> "Trying to get the view natural endpoint on a non-data replica" runtime 
> exception in getViewNaturalEndpoint. I added logging to 
> getViewNaturalEndpoint to show the results of 
> replicationStrategy.getNaturalEndpoints for the baseToken and viewToken.
> It can be seen that these problems occur when the baseEndpoints and 
> viewEndpoints are identical but do not contain the broadcast address of the 
> local node.
> For example, a node at 10.0.0.5 crashes on replay of a write whose base token 
> and view token replicas are both [10.0.0.2, 10.0.0.4, 10.0.0.6]. It seems we 
> try to guard against this by considering pendingEndpoints for the viewToken, 
> but this does not appear to be sufficient.
> I've attached the system.logs for a test run with added logging. In the 
> attached logs, n1 is at 10.0.0.2, n2 is at 10.0.0.3, and so on. 10.0.0.6/n5 
> is the decommissioned node.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10413) Replaying materialized view updates from commitlog after node decommission crashes Cassandra

2015-10-06 Thread Joel Knighton (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10413?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joel Knighton updated CASSANDRA-10413:
--
Reviewer: T Jake Luciani

> Replaying materialized view updates from commitlog after node decommission 
> crashes Cassandra
> 
>
> Key: CASSANDRA-10413
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10413
> Project: Cassandra
>  Issue Type: Bug
>Reporter: Joel Knighton
>Assignee: Joel Knighton
>Priority: Critical
> Fix For: 3.0.0 rc2
>
> Attachments: n1.log, n2.log, n3.log, n4.log, n5.log
>
>
> This issue is reproducible through a Jepsen test, runnable as
> {code}
> lein with-profile +trunk test :only 
> cassandra.mv-test/mv-crash-subset-decommission
> {code}
> This test crashes/restarts nodes while decommissioning nodes. These actions 
> are not coordinated.
> In [10164|https://issues.apache.org/jira/browse/CASSANDRA-10164], we 
> introduced a change to re-apply materialized view updates on commitlog replay.
> Some nodes, upon restart, will crash in commitlog replay. They throw the 
> "Trying to get the view natural endpoint on a non-data replica" runtime 
> exception in getViewNaturalEndpoint. I added logging to 
> getViewNaturalEndpoint to show the results of 
> replicationStrategy.getNaturalEndpoints for the baseToken and viewToken.
> It can be seen that these problems occur when the baseEndpoints and 
> viewEndpoints are identical but do not contain the broadcast address of the 
> local node.
> For example, a node at 10.0.0.5 crashes on replay of a write whose base token 
> and view token replicas are both [10.0.0.2, 10.0.0.4, 10.0.0.6]. It seems we 
> try to guard against this by considering pendingEndpoints for the viewToken, 
> but this does not appear to be sufficient.
> I've attached the system.logs for a test run with added logging. In the 
> attached logs, n1 is at 10.0.0.2, n2 is at 10.0.0.3, and so on. 10.0.0.6/n5 
> is the decommissioned node.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10447) Async logging configuration doesn't result in data flushing when shutdown hook runs

2015-10-06 Thread Ariel Weisberg (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10447?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945348#comment-14945348
 ] 

Ariel Weisberg commented on CASSANDRA-10447:


Oh thanks Paulo, I didn't have the source for logbook classic loaded so I 
didn't see that LoggerContext overrides stop(). I was very confused.

It's easy to reproduce. On trunk run "ant test -Dtest.name=RoleOptionsTest" and 
then look at the log file in build/test/logs. For me the log file is empty.

It definitely happens in the test config and it's very obvious. Whether it 
happens in the production config is harder to suss out since the production 
config doesn't buffer heavily like the test config and drains as fast as it 
can. My suspicion is that it should have the same issue not stopping appenders.

If you are free you are welcome to look at it. I am doing two code reviews, but 
I should be able to get to both for 3.0.0rc2.

> Async logging configuration doesn't result in data flushing when shutdown 
> hook runs
> ---
>
> Key: CASSANDRA-10447
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10447
> Project: Cassandra
>  Issue Type: Bug
>Reporter: Ariel Weisberg
>Assignee: Ariel Weisberg
> Fix For: 3.0.0 rc2
>
>
> Stefania discovered that tests that don't produce a lot of log output end up 
> producing 0 debug output to files because the data is not flushed as part of 
> the shutdown hook. I traced through and it looks like the shutdown hook 
> doesn't actually invoke code that does anything useful. It shuts down an 
> executor service in the logging context but doesn't call stop on any 
> appenders.
> A hackish thing we can do is use a status listener to collect all the 
> appenders and then stop them when the shutdown hook runs. Even adding a small 
> delay to the shutdown hook (no code changes on our part) would in let the 
> async appender flush in 90% of cases.
> We still need to fix it for test which uses a different config file and for 
> which a small delay is not desirable.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-10452) Fix flapping resummable_bootstrap_test

2015-10-06 Thread Jim Witschey (JIRA)
Jim Witschey created CASSANDRA-10452:


 Summary: Fix flapping resummable_bootstrap_test
 Key: CASSANDRA-10452
 URL: https://issues.apache.org/jira/browse/CASSANDRA-10452
 Project: Cassandra
  Issue Type: Sub-task
Reporter: Jim Witschey
Assignee: Yuki Morishita
 Fix For: 3.0.0 rc2


{{bootstrap_test.py:TestBootstrap.resumable_bootstrap_test}} has been flapping 
on CassCI lately:

http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/bootstrap_test/TestBootstrap/resumable_bootstrap_test/history/

I have not been able to reproduce on OpenStack. I'm assigning [~yukim] for now, 
but feel free to reassign.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10412) Could not initialize class org.apache.cassandra.config.DatabaseDescriptor

2015-10-06 Thread Carl Yeksigian (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10412?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Yeksigian updated CASSANDRA-10412:
---
Fix Version/s: 2.2.x

> Could not initialize class org.apache.cassandra.config.DatabaseDescriptor
> -
>
> Key: CASSANDRA-10412
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10412
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
> Environment: OS: Windows 2012 R2
> Java JRE: 1.8.0_51
>Reporter: Eric Simmons
>Assignee: Carl Yeksigian
>Priority: Minor
> Fix For: 2.2.x
>
> Attachments: cassandra-env.ps1, cassandra.yaml
>
>
> We are unable to start version 2.2.1 on our Windows 2012 R2 systems, however 
> we can use the same environment to start version 2.1.2
> I have attached our Cassandra.yaml and cassandra-env.ps1 file for reference.  
> I have also attached the system.log file displaying the error.
> I have also included an excerpt of the log file showing the stack trace of 
> the error.
> ERROR [ScheduledTasks:1] 2015-09-29 07:03:47,482 
> DebuggableThreadPoolExecutor.java:242 - Error in ThreadPoolExecutor
> java.lang.NoClassDefFoundError: Could not initialize class 
> org.apache.cassandra.config.DatabaseDescriptor
>   at 
> org.apache.cassandra.utils.JVMStabilityInspector.inspectThrowable(JVMStabilityInspector.java:57)
>  ~[apache-cassandra-2.2.1.jar:2.2.1]
>   at 
> org.apache.cassandra.concurrent.DebuggableScheduledThreadPoolExecutor$UncomplainingRunnable.run(DebuggableScheduledThreadPoolExecutor.java:122)
>  ~[apache-cassandra-2.2.1.jar:2.2.1]
>   at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) 
> ~[na:1.8.0_51]
>   at java.util.concurrent.FutureTask.runAndReset(Unknown Source) 
> ~[na:1.8.0_51]
>   at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(Unknown
>  Source) ~[na:1.8.0_51]
>   at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(Unknown
>  Source) ~[na:1.8.0_51]
>   at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) 
> [na:1.8.0_51]
>   at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) 
> [na:1.8.0_51]
>   at java.lang.Thread.run(Unknown Source) [na:1.8.0_51]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10447) Async logging configuration doesn't result in data flushing when shutdown hook runs

2015-10-06 Thread Paulo Motta (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10447?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945255#comment-14945255
 ] 

Paulo Motta commented on CASSANDRA-10447:
-

The shutdown hook seems to be implemented correctly on 
[DelayingShutdownHook|https://github.com/qos-ch/logback/blob/master/logback-core/src/main/java/ch/qos/logback/core/hook/DelayingShutdownHook.java],
 that calls {{ContextBase.stop()}}. Even though the 
[ContextBase|https://github.com/qos-ch/logback/blob/master/logback-core/src/main/java/ch/qos/logback/core/ContextBase.java]
 {{stop()}} implementation only shutdowns an executor service, it seems the 
[LoggerContext|https://github.com/qos-ch/logback/blob/master/logback-classic/src/main/java/ch/qos/logback/classic/LoggerContext.java],
 which overrides {{ContextBase.stop()}}, closes all appenders when invoked by 
{{DelayingShutdownHook}}.

In what situations is the problem occurring, is there an easy way to reproduce? 
Maybe shutdown hooks are never triggered during unit tests? We need first to 
clarify if this is specific to the test configuration or if it also happens in 
the production configuration too. I'm happy to review and help with this.

> Async logging configuration doesn't result in data flushing when shutdown 
> hook runs
> ---
>
> Key: CASSANDRA-10447
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10447
> Project: Cassandra
>  Issue Type: Bug
>Reporter: Ariel Weisberg
>Assignee: Ariel Weisberg
> Fix For: 3.0.0 rc2
>
>
> Stefania discovered that tests that don't produce a lot of log output end up 
> producing 0 debug output to files because the data is not flushed as part of 
> the shutdown hook. I traced through and it looks like the shutdown hook 
> doesn't actually invoke code that does anything useful. It shuts down an 
> executor service in the logging context but doesn't call stop on any 
> appenders.
> A hackish thing we can do is use a status listener to collect all the 
> appenders and then stop them when the shutdown hook runs. Even adding a small 
> delay to the shutdown hook (no code changes on our part) would in let the 
> async appender flush in 90% of cases.
> We still need to fix it for test which uses a different config file and for 
> which a small delay is not desirable.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10205) decommissioned_wiped_node_can_join_test fails on Jenkins

2015-10-06 Thread Jim Witschey (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jim Witschey updated CASSANDRA-10205:
-
Fix Version/s: 3.0.0 rc2

> decommissioned_wiped_node_can_join_test fails on Jenkins
> 
>
> Key: CASSANDRA-10205
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10205
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Stefania
>Assignee: Stefania
> Fix For: 3.0.0 rc2
>
> Attachments: decommissioned_wiped_node_can_join_test.tar.gz
>
>
> This test passes locally but reliably fails on Jenkins. It seems after we 
> restart node4, it is unable to Gossip with other nodes:
> {code}
> INFO  [HANDSHAKE-/127.0.0.2] 2015-08-27 06:50:42,778 
> OutboundTcpConnection.java:494 - Handshaking version with /127.0.0.2
> INFO  [HANDSHAKE-/127.0.0.1] 2015-08-27 06:50:42,778 
> OutboundTcpConnection.java:494 - Handshaking version with /127.0.0.1
> INFO  [HANDSHAKE-/127.0.0.3] 2015-08-27 06:50:42,778 
> OutboundTcpConnection.java:494 - Handshaking version with /127.0.0.3
> ERROR [main] 2015-08-27 06:51:13,785 CassandraDaemon.java:635 - Exception 
> encountered during startup
> java.lang.RuntimeException: Unable to gossip with any seeds
> at 
> org.apache.cassandra.gms.Gossiper.doShadowRound(Gossiper.java:1342) 
> ~[main/:na]
> at 
> org.apache.cassandra.service.StorageService.checkForEndpointCollision(StorageService.java:518)
>  ~[main/:na]
> at 
> org.apache.cassandra.service.StorageService.prepareToJoin(StorageService.java:763)
>  ~[main/:na]
> at 
> org.apache.cassandra.service.StorageService.initServer(StorageService.java:687)
>  ~[main/:na]
> at 
> org.apache.cassandra.service.StorageService.initServer(StorageService.java:570)
>  ~[main/:na]
> at 
> org.apache.cassandra.service.CassandraDaemon.setup(CassandraDaemon.java:320) 
> [main/:na]
> at 
> org.apache.cassandra.service.CassandraDaemon.activate(CassandraDaemon.java:516)
>  [main/:na]
> at 
> org.apache.cassandra.service.CassandraDaemon.main(CassandraDaemon.java:622) 
> [main/:na]
> WARN  [StorageServiceShutdownHook] 2015-08-27 06:51:13,799 Gossiper.java:1453 
> - No local state or state is in silent shutdown, not announcing shutdown
> {code}
> It seems both the addresses and port number of the seeds are correct so I 
> don't think the problem is the Amazon private addresses but I might be wrong. 
> It's also worth noting that the first time the node starts up without 
> problems. The problem only occurs during a restart.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10205) decommissioned_wiped_node_can_join_test fails on Jenkins

2015-10-06 Thread Jim Witschey (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jim Witschey updated CASSANDRA-10205:
-
Issue Type: Sub-task  (was: Test)
Parent: CASSANDRA-10166

> decommissioned_wiped_node_can_join_test fails on Jenkins
> 
>
> Key: CASSANDRA-10205
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10205
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Stefania
>Assignee: Stefania
> Fix For: 3.0.0 rc2
>
> Attachments: decommissioned_wiped_node_can_join_test.tar.gz
>
>
> This test passes locally but reliably fails on Jenkins. It seems after we 
> restart node4, it is unable to Gossip with other nodes:
> {code}
> INFO  [HANDSHAKE-/127.0.0.2] 2015-08-27 06:50:42,778 
> OutboundTcpConnection.java:494 - Handshaking version with /127.0.0.2
> INFO  [HANDSHAKE-/127.0.0.1] 2015-08-27 06:50:42,778 
> OutboundTcpConnection.java:494 - Handshaking version with /127.0.0.1
> INFO  [HANDSHAKE-/127.0.0.3] 2015-08-27 06:50:42,778 
> OutboundTcpConnection.java:494 - Handshaking version with /127.0.0.3
> ERROR [main] 2015-08-27 06:51:13,785 CassandraDaemon.java:635 - Exception 
> encountered during startup
> java.lang.RuntimeException: Unable to gossip with any seeds
> at 
> org.apache.cassandra.gms.Gossiper.doShadowRound(Gossiper.java:1342) 
> ~[main/:na]
> at 
> org.apache.cassandra.service.StorageService.checkForEndpointCollision(StorageService.java:518)
>  ~[main/:na]
> at 
> org.apache.cassandra.service.StorageService.prepareToJoin(StorageService.java:763)
>  ~[main/:na]
> at 
> org.apache.cassandra.service.StorageService.initServer(StorageService.java:687)
>  ~[main/:na]
> at 
> org.apache.cassandra.service.StorageService.initServer(StorageService.java:570)
>  ~[main/:na]
> at 
> org.apache.cassandra.service.CassandraDaemon.setup(CassandraDaemon.java:320) 
> [main/:na]
> at 
> org.apache.cassandra.service.CassandraDaemon.activate(CassandraDaemon.java:516)
>  [main/:na]
> at 
> org.apache.cassandra.service.CassandraDaemon.main(CassandraDaemon.java:622) 
> [main/:na]
> WARN  [StorageServiceShutdownHook] 2015-08-27 06:51:13,799 Gossiper.java:1453 
> - No local state or state is in silent shutdown, not announcing shutdown
> {code}
> It seems both the addresses and port number of the seeds are correct so I 
> don't think the problem is the Amazon private addresses but I might be wrong. 
> It's also worth noting that the first time the node starts up without 
> problems. The problem only occurs during a restart.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10451) Fix batch_test.TestBatch.logged_batch_gcgs_below_threshold_multi_table_test on CassCI

2015-10-06 Thread Jim Witschey (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945325#comment-14945325
 ] 

Jim Witschey commented on CASSANDRA-10451:
--

Sorry, I was looking at trunk instead of the 3.0 tests. However, the same 
pattern of failures shows up on the cassandra-3.0 branch:

http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/batch_test/TestBatch/logged_batch_gcgs_below_threshold_multi_table_test/history/

The same pattern also shows on the 
{{logged_batch_gcgs_below_threshold_single_table_test}} tests:

http://cassci.datastax.com/view/trunk/job/cassandra-3.0_dtest/lastCompletedBuild/testReport/batch_test/TestBatch/logged_batch_gcgs_below_threshold_single_table_test/history/

I'll edit the title to reflect that.

> Fix batch_test.TestBatch.logged_batch_gcgs_below_threshold_multi_table_test 
> on CassCI
> -
>
> Key: CASSANDRA-10451
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10451
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Paulo Motta
> Fix For: 3.0.0 rc2
>
>
> This is a recent regression:
> http://cassci.datastax.com/view/trunk/job/trunk_dtest/lastCompletedBuild/testReport/batch_test/TestBatch/logged_batch_gcgs_below_threshold_multi_table_test/history/
> Based on [when it started 
> failing|http://cassci.datastax.com/view/trunk/job/trunk_dtest/623/changes], 
> it looks like the regression was introduced by [the recent logging 
> changes|https://github.com/apache/cassandra/commit/4a849efeb7c7c1a54bc12094d2d6a9f3f008a2fa].
>  Pinging [~blerer] and [~pauloricardomg] to have a look.
> My initial guess was that the logs the test depends on were moved to the new 
> debug log. However, the tests passes when I run it manually on openstack, so 
> it could just be a configuration issue on CassCI.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10447) Async logging configuration doesn't result in data flushing when shutdown hook runs

2015-10-06 Thread Paulo Motta (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paulo Motta updated CASSANDRA-10447:

Reviewer: Paulo Motta

> Async logging configuration doesn't result in data flushing when shutdown 
> hook runs
> ---
>
> Key: CASSANDRA-10447
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10447
> Project: Cassandra
>  Issue Type: Bug
>Reporter: Ariel Weisberg
>Assignee: Ariel Weisberg
> Fix For: 3.0.0 rc2
>
>
> Stefania discovered that tests that don't produce a lot of log output end up 
> producing 0 debug output to files because the data is not flushed as part of 
> the shutdown hook. I traced through and it looks like the shutdown hook 
> doesn't actually invoke code that does anything useful. It shuts down an 
> executor service in the logging context but doesn't call stop on any 
> appenders.
> A hackish thing we can do is use a status listener to collect all the 
> appenders and then stop them when the shutdown hook runs. Even adding a small 
> delay to the shutdown hook (no code changes on our part) would in let the 
> async appender flush in 90% of cases.
> We still need to fix it for test which uses a different config file and for 
> which a small delay is not desirable.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-10450) upgrade_tests/paging_test.py:TestPagingWithDeletions.test_single_cell_deletions flapped on CassCI

2015-10-06 Thread Jim Witschey (JIRA)
Jim Witschey created CASSANDRA-10450:


 Summary: 
upgrade_tests/paging_test.py:TestPagingWithDeletions.test_single_cell_deletions 
flapped on CassCI
 Key: CASSANDRA-10450
 URL: https://issues.apache.org/jira/browse/CASSANDRA-10450
 Project: Cassandra
  Issue Type: Sub-task
Reporter: Jim Witschey
Priority: Minor
 Fix For: 3.0.0 rc2


This test has failed with a single flap:

http://cassci.datastax.com/view/trunk/job/trunk_dtest/lastCompletedBuild/testReport/upgrade_tests.paging_test/TestPagingWithDeletions/test_single_cell_deletions/history/

This may not be worth blocking release on; I haven't been able to reproduce on 
openstack, and it's only flapped once.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10233) IndexOutOfBoundsException in HintedHandOffManager

2015-10-06 Thread Joshua McKenzie (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10233?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joshua McKenzie updated CASSANDRA-10233:

Assignee: J.P. Eiti Kimura  (was: Paulo Motta)

> IndexOutOfBoundsException in HintedHandOffManager
> -
>
> Key: CASSANDRA-10233
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10233
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
> Environment: Cassandra 2.2.0
>Reporter: Omri Iluz
>Assignee: J.P. Eiti Kimura
> Attachments: cassandra-2.1.8-10233-v2.txt, 
> cassandra-2.1.8-10233-v3.txt, cassandra-2.1.8-10233-v4.txt, 
> cassandra-2.2.1-10233.txt
>
>
> After upgrading our cluster to 2.2.0, the following error started showing 
> exectly every 10 minutes on every server in the cluster:
> {noformat}
> INFO  [CompactionExecutor:1381] 2015-08-31 18:31:55,506 
> CompactionTask.java:142 - Compacting (8e7e1520-500e-11e5-b1e3-e95897ba4d20) 
> [/cassandra/data/system/hints-2666e20573ef38b390fefecf96e8f0c7/la-540-big-Data.db:level=0,
>  ]
> INFO  [CompactionExecutor:1381] 2015-08-31 18:31:55,599 
> CompactionTask.java:224 - Compacted (8e7e1520-500e-11e5-b1e3-e95897ba4d20) 1 
> sstables to 
> [/cassandra/data/system/hints-2666e20573ef38b390fefecf96e8f0c7/la-541-big,] 
> to level=0.  1,544,495 bytes to 1,544,495 (~100% of original) in 93ms = 
> 15.838121MB/s.  0 total partitions merged to 4.  Partition merge counts were 
> {1:4, }
> ERROR [HintedHandoff:1] 2015-08-31 18:31:55,600 CassandraDaemon.java:182 - 
> Exception in thread Thread[HintedHandoff:1,1,main]
> java.lang.IndexOutOfBoundsException: null
>   at java.nio.Buffer.checkIndex(Buffer.java:538) ~[na:1.7.0_79]
>   at java.nio.HeapByteBuffer.getLong(HeapByteBuffer.java:410) 
> ~[na:1.7.0_79]
>   at org.apache.cassandra.utils.UUIDGen.getUUID(UUIDGen.java:106) 
> ~[apache-cassandra-2.2.0.jar:2.2.0]
>   at 
> org.apache.cassandra.db.HintedHandOffManager.scheduleAllDeliveries(HintedHandOffManager.java:515)
>  ~[apache-cassandra-2.2.0.jar:2.2.0]
>   at 
> org.apache.cassandra.db.HintedHandOffManager.access$000(HintedHandOffManager.java:88)
>  ~[apache-cassandra-2.2.0.jar:2.2.0]
>   at 
> org.apache.cassandra.db.HintedHandOffManager$1.run(HintedHandOffManager.java:168)
>  ~[apache-cassandra-2.2.0.jar:2.2.0]
>   at 
> org.apache.cassandra.concurrent.DebuggableScheduledThreadPoolExecutor$UncomplainingRunnable.run(DebuggableScheduledThreadPoolExecutor.java:118)
>  ~[apache-cassandra-2.2.0.jar:2.2.0]
>   at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) 
> [na:1.7.0_79]
>   at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:304) 
> [na:1.7.0_79]
>   at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:178)
>  [na:1.7.0_79]
>   at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
>  [na:1.7.0_79]
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>  [na:1.7.0_79]
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>  [na:1.7.0_79]
>   at java.lang.Thread.run(Thread.java:745) [na:1.7.0_79]
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-10451) Fix batch_test.TestBatch.logged_batch_gcgs_below_threshold_multi_table_test on CassCI

2015-10-06 Thread Jim Witschey (JIRA)
Jim Witschey created CASSANDRA-10451:


 Summary: Fix 
batch_test.TestBatch.logged_batch_gcgs_below_threshold_multi_table_test on 
CassCI
 Key: CASSANDRA-10451
 URL: https://issues.apache.org/jira/browse/CASSANDRA-10451
 Project: Cassandra
  Issue Type: Sub-task
Reporter: Jim Witschey
 Fix For: 3.0.0 rc2


This is a recent regression:

http://cassci.datastax.com/view/trunk/job/trunk_dtest/lastCompletedBuild/testReport/batch_test/TestBatch/logged_batch_gcgs_below_threshold_multi_table_test/history/

Based on [when it started 
failing|http://cassci.datastax.com/view/trunk/job/trunk_dtest/623/changes], it 
looks like the regression was introduced by [the recent logging 
changes|https://github.com/apache/cassandra/commit/4a849efeb7c7c1a54bc12094d2d6a9f3f008a2fa].
 Pinging [~blerer] and [~pauloricardomg] to have a look.

My initial guess was that the logs the test depends on were moved to the new 
debug log. However, the tests passes when I run it manually on openstack, so it 
could just be a configuration issue on CassCI.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10444) Create an option to forcibly disable tracing

2015-10-06 Thread Joshua McKenzie (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10444?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joshua McKenzie updated CASSANDRA-10444:

Assignee: (was: Joshua McKenzie)

> Create an option to forcibly disable tracing
> 
>
> Key: CASSANDRA-10444
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10444
> Project: Cassandra
>  Issue Type: Improvement
>  Components: Core
>Reporter: Brandon Williams
>Priority: Minor
> Fix For: 2.1.x
>
>
> Sometimes people will experience dropped TRACE messages.  Ostensibly, trace 
> is disabled on the server and we know it's from some client, somewhere.  With 
> an inability to locate exactly where client code is causing this, it would be 
> useful to just be able to kill it entirely on the server side.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (CASSANDRA-10451) Fix batch_test.TestBatch.logged_batch_gcgs_below_threshold_multi_table_test on CassCI

2015-10-06 Thread Paulo Motta (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10451?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paulo Motta reassigned CASSANDRA-10451:
---

Assignee: Paulo Motta

> Fix batch_test.TestBatch.logged_batch_gcgs_below_threshold_multi_table_test 
> on CassCI
> -
>
> Key: CASSANDRA-10451
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10451
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Paulo Motta
> Fix For: 3.0.0 rc2
>
>
> This is a recent regression:
> http://cassci.datastax.com/view/trunk/job/trunk_dtest/lastCompletedBuild/testReport/batch_test/TestBatch/logged_batch_gcgs_below_threshold_multi_table_test/history/
> Based on [when it started 
> failing|http://cassci.datastax.com/view/trunk/job/trunk_dtest/623/changes], 
> it looks like the regression was introduced by [the recent logging 
> changes|https://github.com/apache/cassandra/commit/4a849efeb7c7c1a54bc12094d2d6a9f3f008a2fa].
>  Pinging [~blerer] and [~pauloricardomg] to have a look.
> My initial guess was that the logs the test depends on were moved to the new 
> debug log. However, the tests passes when I run it manually on openstack, so 
> it could just be a configuration issue on CassCI.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-10451) Fix batch_test.TestBatch.logged_batch_gcgs_below_threshold_{multi,single}_table_test on CassCI

2015-10-06 Thread Jim Witschey (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10451?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jim Witschey updated CASSANDRA-10451:
-
Summary: Fix 
batch_test.TestBatch.logged_batch_gcgs_below_threshold_{multi,single}_table_test
 on CassCI  (was: Fix 
batch_test.TestBatch.logged_batch_gcgs_below_threshold_multi_table_test on 
CassCI)

> Fix 
> batch_test.TestBatch.logged_batch_gcgs_below_threshold_{multi,single}_table_test
>  on CassCI
> --
>
> Key: CASSANDRA-10451
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10451
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Jim Witschey
>Assignee: Paulo Motta
> Fix For: 3.0.0 rc2
>
>
> This is a recent regression:
> http://cassci.datastax.com/view/trunk/job/trunk_dtest/lastCompletedBuild/testReport/batch_test/TestBatch/logged_batch_gcgs_below_threshold_multi_table_test/history/
> Based on [when it started 
> failing|http://cassci.datastax.com/view/trunk/job/trunk_dtest/623/changes], 
> it looks like the regression was introduced by [the recent logging 
> changes|https://github.com/apache/cassandra/commit/4a849efeb7c7c1a54bc12094d2d6a9f3f008a2fa].
>  Pinging [~blerer] and [~pauloricardomg] to have a look.
> My initial guess was that the logs the test depends on were moved to the new 
> debug log. However, the tests passes when I run it manually on openstack, so 
> it could just be a configuration issue on CassCI.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-7392) Abort in-progress queries that time out

2015-10-06 Thread Ariel Weisberg (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-7392?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945386#comment-14945386
 ] 

Ariel Weisberg commented on CASSANDRA-7392:
---

+1 on the code. I can't really tell what test-all on trunk is doing. The last 
build timed out and the build before that did something weird. dtests look like 
dtests. test-all on 3.0 looked OK to me.

> Abort in-progress queries that time out
> ---
>
> Key: CASSANDRA-7392
> URL: https://issues.apache.org/jira/browse/CASSANDRA-7392
> Project: Cassandra
>  Issue Type: New Feature
>  Components: Core
>Reporter: Jonathan Ellis
>Assignee: Stefania
>Priority: Critical
> Fix For: 3.x
>
>
> Currently we drop queries that time out before we get to them (because node 
> is overloaded) but not queries that time out while being processed.  
> (Particularly common for index queries on data that shouldn't be indexed.)  
> Adding the latter and logging when we have to interrupt one gets us a poor 
> man's "slow query log" for free.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-10437) Remove offheap_objects option until 9472 re-introduces them

2015-10-06 Thread Ariel Weisberg (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-10437?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945432#comment-14945432
 ] 

Ariel Weisberg commented on CASSANDRA-10437:


The code changes look fine, but is there a branch where this is being tested by 
cassci? Are there dtests requesting this option?

> Remove offheap_objects option until 9472 re-introduces them
> ---
>
> Key: CASSANDRA-10437
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10437
> Project: Cassandra
>  Issue Type: Bug
>Reporter: Sylvain Lebresne
>Assignee: Sylvain Lebresne
>Priority: Trivial
> Fix For: 3.0.0 rc2
>
> Attachments: 0001-Remove-offheap_objects-option.txt
>
>
> We need to send a meaningful error message if user try to use 
> {{offheap_objects}} in 3.0 since it's not supported currently (pending 
> CASSANDRA-9472). And document the current removal in the NEWS file.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-9126) java.lang.RuntimeException: Last written key DecoratedKey >= current key DecoratedKey

2015-10-06 Thread Jim Witschey (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-9126?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14945115#comment-14945115
 ] 

Jim Witschey commented on CASSANDRA-9126:
-

[~krummas] Sure -- could we add that advice to the error messages these users 
received? Or are these errors raised in too many situations for that to be 
appropriate?

> java.lang.RuntimeException: Last written key DecoratedKey >= current key 
> DecoratedKey
> -
>
> Key: CASSANDRA-9126
> URL: https://issues.apache.org/jira/browse/CASSANDRA-9126
> Project: Cassandra
>  Issue Type: Bug
>  Components: Core
>Reporter: srinivasu gottipati
>Priority: Critical
> Fix For: 2.1.x
>
> Attachments: cassandra-system.log
>
>
> Cassandra V: 2.0.14,
> Getting the following exceptions while trying to compact (I see this issue 
> was raised in earlier versions and marked as closed. However it still appears 
> in 2.0.14). In our case, compaction is not getting succeeded and keep failing 
> with this error.:
> {code}java.lang.RuntimeException: Last written key 
> DecoratedKey(3462767860784856708, 
> 354038323137333038305f3330325f31355f474d4543454f) >= current key 
> DecoratedKey(3462334604624154281, 
> 354036333036353334315f3336315f31355f474d4543454f) writing into {code}
> ...
> Stacktrace:{code}
>   at 
> org.apache.cassandra.io.sstable.SSTableWriter.beforeAppend(SSTableWriter.java:143)
>   at 
> org.apache.cassandra.io.sstable.SSTableWriter.append(SSTableWriter.java:166)
>   at 
> org.apache.cassandra.db.compaction.CompactionTask.runMayThrow(CompactionTask.java:167)
>   at 
> org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28)
>   at 
> org.apache.cassandra.db.compaction.CompactionTask.executeInternal(CompactionTask.java:60)
>   at 
> org.apache.cassandra.db.compaction.AbstractCompactionTask.execute(AbstractCompactionTask.java:59)
>   at 
> org.apache.cassandra.db.compaction.CompactionManager$BackgroundCompactionTask.run(CompactionManager.java:198)
>   at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>   at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745){code}
> Any help is greatly appreciated



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (CASSANDRA-10447) Async logging configuration doesn't result in data flushing when shutdown hook runs

2015-10-06 Thread Ariel Weisberg (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-10447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ariel Weisberg reassigned CASSANDRA-10447:
--

Assignee: Ariel Weisberg

> Async logging configuration doesn't result in data flushing when shutdown 
> hook runs
> ---
>
> Key: CASSANDRA-10447
> URL: https://issues.apache.org/jira/browse/CASSANDRA-10447
> Project: Cassandra
>  Issue Type: Bug
>Reporter: Ariel Weisberg
>Assignee: Ariel Weisberg
> Fix For: 3.0.0 rc2
>
>
> Stefania discovered that tests that don't produce a lot of log output end up 
> producing 0 debug output to files because the data is not flushed as part of 
> the shutdown hook. I traced through and it looks like the shutdown hook 
> doesn't actually invoke code that does anything useful. It shuts down an 
> executor service in the logging context but doesn't call stop on any 
> appenders.
> A hackish thing we can do is use a status listener to collect all the 
> appenders and then stop them when the shutdown hook runs. Even adding a small 
> delay to the shutdown hook (no code changes on our part) would in let the 
> async appender flush in 90% of cases.
> We still need to fix it for test which uses a different config file and for 
> which a small delay is not desirable.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)