Yes, we are doing a lot of inserts.

But how can CASSANDRA-1042 cause an OutOfMemory?
And we are using multigetSlice(). We are not doing any get_range_slice() at all.

2010-07-23 







发件人: Jonathan Ellis 
发送时间: 2010-07-21  21:17:21 
收件人: user 
抄送: 
主题: Re: Re: What is consuming the heap? 
 
On Tue, Jul 20, 2010 at 11:33 PM, Peter Schuller
<peter.schul...@infidyne.com> wrote:
>>  INFO [GC inspection] 2010-07-21 01:01:49,661 GCInspector.java (line 110) GC 
>> for ConcurrentMarkSweep: 11748 ms, 413673472 reclaimed leaving 9779542600 
>> used; max is 10873667584
>> ERROR [Thread-35] 2010-07-21 01:02:10,941 CassandraDaemon.java (line 78) 
>> Fatal exception in thread Thread[Thread-35,5,main]
>> java.lang.OutOfMemoryError: Java heap space
>>  INFO [GC inspection] 2010-07-21 01:02:10,958 GCInspector.java (line 110) GC 
>> for ConcurrentMarkSweep: 10043 ms, 259576 reclaimed leaving 10172790816 
>> used; max is 10873667584
>
> So that confirms a "legitimate" out-of-memory condition in the sense
> that CMS is reclaiming extremely little and the live set after a
> concurrent mark/sweep is indeed around the 10 gig.
Are you doing a lot of inserts?  You might be hitting
https://issues.apache.org/jira/browse/CASSANDRA-1042
-- 
Jonathan Ellis
Project Chair, Apache Cassandra
co-founder of Riptano, the source for professional Cassandra support
http://riptano.com

Reply via email to