Re: Long GC due to promotion failures

2014-01-22 Thread John Watson
old gen collection in some cases. The flag's low overhead, and very effective if that's your problem too. Cheers, Lee On Tue, Jan 21, 2014 at 12:02 AM, John Watson j...@disqus.com wrote: Pretty reliable, at some point, nodes will have super long GCs. Followed by https://issues.apache.org

Re: Long GC due to promotion failures

2014-01-22 Thread John Watson
really understood fully. Anyone care to explain using OP attached cfhistogram ? Taking a wild shot, perhaps trying different build, oracle jdk 1.6u25 perhaps? HTH Jason On Tue, Jan 21, 2014 at 4:02 PM, John Watson j...@disqus.com wrote: Pretty reliable, at some point, nodes will have

Long GC due to promotion failures

2014-01-21 Thread John Watson
Pretty reliable, at some point, nodes will have super long GCs. Followed by https://issues.apache.org/jira/browse/CASSANDRA-6592 Lovely log messages: 9030.798: [ParNew (0: promotion failure size = 4194306) (2: promotion failure size = 4194306) (4: promotion failure size = 4194306)

hot sstables evicted from page cache on compaction causing high latency

2013-07-12 Thread John Watson
Having a real issue where at the completion of large compactions, it will evict hot sstables from the kernel page cache causing huge read latency while it is backfilled. https://dl.dropboxusercontent.com/s/149h7ssru0dapkg/Screen%20Shot%202013-07-12%20at%201.46.19%20PM.png Blue line - page cache

Compaction causing OutOfHeap

2013-05-26 Thread John Watson
Having (2) 1.2.5 nodes constantly crashing due to OutOfHeap errors. It always happens when the same large compaction is about to finish (they re-run the same compaction after restarting.) An indicator is CMS GC time of 3-5s (and the many related problems felt throughout the rest of the cluster)

Re: cassandra-shuffle time to completion and required disk space

2013-04-29 Thread John Watson
bootstrap/decommission. You would set num_tokens on the existing hosts (and restart them) so that they split their ranges, then bootstrap in N new hosts, then decommission the old ones. On 28 April 2013 22:21, John Watson j...@disqus.com wrote: The amount of time/space cassandra-shuffle

Re: setcompactionthroughput and setstreamthroughput have no effect

2013-04-29 Thread John Watson
Same behavior on 1.1.3, 1.1.5 and 1.1.9. Currently: 1.2.3 On Mon, Apr 29, 2013 at 11:43 AM, Robert Coli rc...@eventbrite.com wrote: On Sun, Apr 28, 2013 at 2:28 PM, John Watson j...@disqus.com wrote: Running these 2 commands are noop IO wise: nodetool setcompactionthroughput 0

Re: Adding nodes in 1.2 with vnodes requires huge disks

2013-04-29 Thread John Watson
nodetool status so we can get a feel for the ring? Can you include the logs from one of the nodes that failed to join ? Thanks - Aaron Morton Freelance Cassandra Consultant New Zealand @aaronmorton http://www.thelastpickle.com On 29/04/2013, at 10:01 AM, John Watson j

cassandra-shuffle time to completion and required disk space

2013-04-28 Thread John Watson
The amount of time/space cassandra-shuffle requires when upgrading to using vnodes should really be apparent in documentation (when some is made). Only semi-noticeable remark about the exorbitant amount of time is a bullet point in: http://wiki.apache.org/cassandra/VirtualNodes/Balance Shuffling

setcompactionthroughput and setstreamthroughput have no effect

2013-04-28 Thread John Watson
Running these 2 commands are noop IO wise: nodetool setcompactionthroughput 0 nodetool setstreamtrhoughput 0 If trying to recover or rebuild nodes, it would be super helpful to get more than ~120mbit/s of streaming throughput (per session or ~500mbit total) and ~5% IO utilization in (8) 15k

Re: setcompactionthroughput and setstreamthroughput have no effect

2013-04-28 Thread John Watson
at 5:28 PM, John Watson j...@disqus.com wrote: Running these 2 commands are noop IO wise: nodetool setcompactionthroughput 0 nodetool setstreamtrhoughput 0 If trying to recover or rebuild nodes, it would be super helpful to get more than ~120mbit/s of streaming throughput (per session

Re: Adding nodes in 1.2 with vnodes requires huge disks

2013-04-28 Thread John Watson
://www.thelastpickle.com On 27/04/2013, at 9:35 AM, Bryan Talbot btal...@aeriagames.com wrote: I believe that nodetool rebuild is used to add a new datacenter, not just a new host to an existing cluster. Is that what you ran to add the node? -Bryan On Fri, Apr 26, 2013 at 1:27 PM, John Watson

Re: cassandra-shuffle time to completion and required disk space

2013-04-28 Thread John Watson
Zealand @aaronmorton http://www.thelastpickle.com On 29/04/2013, at 9:21 AM, John Watson j...@disqus.com wrote: The amount of time/space cassandra-shuffle requires when upgrading to using vnodes should really be apparent in documentation (when some is made). Only semi-noticeable remark about

Re: Adding nodes in 1.2 with vnodes requires huge disks

2013-04-26 Thread John Watson
was the same. The problem only seemed to disappear when nodetool repair was applied to all nodes. Regards, Francisco Sobral. On Apr 25, 2013, at 4:57 PM, John Watson j...@disqus.com wrote: After finally upgrading to 1.2.3 from 1.1.9, enabling vnodes, and running upgradesstables, I

Re: 1.1.9 to 1.2.3 upgrade issue

2013-04-18 Thread John Watson
/2013, at 9:20 AM, John Watson j...@disqus.com wrote: Started doing a rolling upgrade of nodes from 1.1.9 to 1.2.3 and nodes on 1.1.9 started flooding this error: Exception in thread Thread[RequestResponseStage:19496,5,main] java.io.IOError: java.io.EOFException

1.1.9 to 1.2.3 upgrade issue

2013-04-14 Thread John Watson
Started doing a rolling upgrade of nodes from 1.1.9 to 1.2.3 and nodes on 1.1.9 started flooding this error: Exception in thread Thread[RequestResponseStage:19496,5,main] java.io.IOError: java.io.EOFException at