Thanks a lot for the help Graham and Robert! Will try increasing heap and
see how it goes.
Here are my gc settings, if they're still helpful (they're mostly the
defaults):
-Xms6G -Xmx6G -Xmn400M -XX:+UseParNewGC -XX:+UseConcMarkSweepGC
-XX:+CMSParallelRemarkEnabled -XX:SurvivorRatio=8
-XX:MaxTenu
ack and many thanks for the tips and help..
jason
On Wed, Dec 3, 2014 at 4:49 AM, Robert Coli wrote:
> On Mon, Dec 1, 2014 at 11:07 PM, Jason Wee wrote:
>
>> Hi Rob, any recommended documentation on describing
>> explanation/configuration of the JVM heap and permanent generation ? We
>> stucke
On Mon, Dec 1, 2014 at 11:07 PM, Jason Wee wrote:
> Hi Rob, any recommended documentation on describing
> explanation/configuration of the JVM heap and permanent generation ? We
> stucked in this same situation too. :(
>
The archives of this list are chock full of explorations of various cases.
Hi Rob, any recommended documentation on describing
explanation/configuration of the JVM heap and permanent generation ? We
stucked in this same situation too. :(
Jason
On Tue, Dec 2, 2014 at 3:42 AM, Robert Coli wrote:
> On Fri, Nov 28, 2014 at 12:55 PM, Paulo Ricardo Motta Gomes <
> paulo.mo.
On Fri, Nov 28, 2014 at 12:55 PM, Paulo Ricardo Motta Gomes <
paulo.mo...@chaordicsystems.com> wrote:
> We restart the whole cluster every 1 or 2 months, to avoid machines
> getting into this crazy state. We tried tuning GC size and parameters,
> different cassandra versions (1.1, 1.2, 2.0), but t
I should note that the young gen size is just a tuning suggestion, not directly
related to your problem at hand.
You might want to make sure you don’t have issues with key/row cache.
Also, I’m assuming that your extra load isn’t hitting tables that you wouldn’t
normally be hitting.
> On Nov 28
Your GC settings would be helpful, though you can see guesstimate by eyeballing
(assuming settings are the same across all 4 images)
Bursty load can be a big cause of old gen fragmentation (as small working set
objects tends to get spilled (promoted) along with memtable slabs which aren’t
flush
Hello,
This is a recurrent behavior of JVM GC in Cassandra that I never completely
understood: when a node is UP for many days (or even months), or receives a
very high load spike (3x-5x normal load), CMS GC pauses start becoming very
frequent and slow, causing periodic timeouts in Cassandra. Tryi
On Wed, Aug 21, 2013 at 10:47 AM, Robert Coli wrote:
> On Tue, Aug 20, 2013 at 11:35 PM, Keith Wright wrote:
>
>> Still looking for help! We have stopped almost ALL traffic to the
>> cluster and still some nodes are showing almost 1000% CPU for cassandra
>> with no iostat activity. We were run
On Tue, Aug 20, 2013 at 11:35 PM, Keith Wright wrote:
> Still looking for help! We have stopped almost ALL traffic to the cluster
> and still some nodes are showing almost 1000% CPU for cassandra with no
> iostat activity. We were running cleanup on one of the nodes that was not
> showing load
m.java:82)
> at java.io.BufferedOutputStream.write(BufferedOutputStream.java:126)
> at org.xerial.snappy.SnappyOutputStream.dump(SnappyOutputStream.java:297)
> at
> org.xerial.snappy.SnappyOutputStream.rawWrite(SnappyOutputStream.java:244)
> at org.xerial.snappy.SnappyOutputStream.write(Sn
...@nanigans.com>>
Reply-To: "user@cassandra.apache.org<mailto:user@cassandra.apache.org>"
mailto:user@cassandra.apache.org>>
Date: Tuesday, August 20, 2013 8:32 PM
To: "user@cassandra.apache.org<mailto:user@cassandra.apache.org>"
mailto:user@cassandra.apache.org>
Hi all,
We are using C* 1.2.4 with Vnodes and SSD. We have seen behavior recently
where 3 of our nodes get locked up in high load in what appears to be a GC
spiral while the rest of the cluster (7 total nodes) appears fine. When I run
a tpstats, I see the following (assuming tpstats retur
13 matches
Mail list logo