Upping to 1GB, memory usage seems to level off at 750MB, but there's a problem in there somewhere. I'm getting a failure message, and the marvel dashboard isn't able to fetch.
C:\elasticsearch-1.1.1\bin>elasticsearch Picked up _JAVA_OPTIONS: -Djava.net.preferIPv4Stack=true [2014-07-30 17:33:27,138][INFO ][node ] [Mondo] version[1.1.1], pid[10864], build[f1585f0/2014-04-16 T14:27:12Z] [2014-07-30 17:33:27,139][INFO ][node ] [Mondo] initializing ... [2014-07-30 17:33:27,163][INFO ][plugins ] [Mondo] loaded [ldap-river, marvel], sites [marvel] [2014-07-30 17:33:30,731][INFO ][node ] [Mondo] initialized [2014-07-30 17:33:30,731][INFO ][node ] [Mondo] starting ... [2014-07-30 17:33:31,027][INFO ][transport ] [Mondo] bound_address {inet[/0.0.0.0:9300]}, publish_address {inet[/192.168.0.6:9300]} [2014-07-30 17:33:34,202][INFO ][cluster.service ] [Mondo] new_master [Mondo][liyNQAHAS0-8f-qDDqa5Rg][twilson-T HINK][inet[/192.168.0.6:9300]], reason: zen-disco-join (elected_as_master) [2014-07-30 17:33:34,239][INFO ][discovery ] [Mondo] elasticsearch/liyNQAHAS0-8f-qDDqa5Rg [2014-07-30 17:33:34,600][INFO ][http ] [Mondo] bound_address {inet[/0.0.0.0:9200]}, publish_address {inet[/192.168.0.6:9200]} [2014-07-30 17:33:35,799][INFO ][gateway ] [Mondo] recovered [66] indices into cluster_state [2014-07-30 17:33:35,815][INFO ][node ] [Mondo] started [2014-07-30 17:33:39,823][DEBUG][action.search.type ] [Mondo] All shards failed for phase: [query_fetch] [2014-07-30 17:33:39,830][DEBUG][action.search.type ] [Mondo] All shards failed for phase: [query_fetch] [2014-07-30 17:33:39,837][DEBUG][action.search.type ] [Mondo] All shards failed for phase: [query_fetch] [2014-07-30 17:33:39,838][DEBUG][action.search.type ] [Mondo] All shards failed for phase: [query_fetch] [2014-07-30 17:33:43,973][DEBUG][action.search.type ] [Mondo] All shards failed for phase: [query_fetch] [2014-07-30 17:33:44,212][DEBUG][action.search.type ] [Mondo] All shards failed for phase: [query_fetch] [2014-07-30 17:33:44,357][DEBUG][action.search.type ] [Mondo] All shards failed for phase: [query_fetch] [2014-07-30 17:33:44,501][DEBUG][action.search.type ] [Mondo] All shards failed for phase: [query_fetch] [2014-07-30 17:33:53,294][DEBUG][action.search.type ] [Mondo] All shards failed for phase: [query_fetch] [2014-07-30 17:33:53,309][DEBUG][action.search.type ] [Mondo] All shards failed for phase: [query_fetch] [2014-07-30 17:33:53,310][DEBUG][action.search.type ] [Mondo] All shards failed for phase: [query_fetch] [2014-07-30 17:33:53,310][DEBUG][action.search.type ] [Mondo] All shards failed for phase: [query_fetch] [2014-07-30 17:34:03,281][DEBUG][action.search.type ] [Mondo] All shards failed for phase: [query_fetch] [2014-07-30 17:34:03,283][DEBUG][action.search.type ] [Mondo] All shards failed for phase: [query_fetch] [2014-07-30 17:34:03,286][DEBUG][action.search.type ] [Mondo] All shards failed for phase: [query_fetch] [2014-07-30 17:34:45,662][ERROR][marvel.agent.exporter ] [Mondo] create failure (index:[.marvel-2014.07.31] type: [no de_stats]): UnavailableShardsException[[.marvel-2014.07.31][0] [2] shardIt, [0] active : Timeout waiting for [1m], reque st: org.elasticsearch.action.bulk.BulkShardRequest@39b65640] On Wednesday, July 30, 2014 5:30:29 PM UTC-7, Mark Walkom wrote: > > Up that to 1GB and see if it starts. > 512MB is pretty tiny, you're better off starting at 1/2GB if you can. > > Regards, > Mark Walkom > > Infrastructure Engineer > Campaign Monitor > email: ma...@campaignmonitor.com <javascript:> > web: www.campaignmonitor.com > > > On 31 July 2014 10:28, Tom Wilson <twils...@gmail.com <javascript:>> > wrote: > >> JDK 1.7.0_51 >> >> It has 512MB of heap, which was enough -- I've been running it like that >> for the past few months, and I only have two indexes and around 300-400 >> documents. This is a development instance I'm running on my local machine. >> This only happened when I started it today. >> >> -tom >> >> >> On Wednesday, July 30, 2014 5:16:11 PM UTC-7, Mark Walkom wrote: >> >>> What java version? How much heap have you allocated and how much RAM on >>> the server? >>> >>> Basically you have too much data for the heap size, so increasing it >>> will help. >>> >>> Regards, >>> Mark Walkom >>> >>> Infrastructure Engineer >>> Campaign Monitor >>> email: ma...@campaignmonitor.com >>> web: www.campaignmonitor.com >>> >>> >>> On 31 July 2014 10:11, Tom Wilson <twils...@gmail.com> wrote: >>> >>>> Help! Elasticsearch was working fine, but now it's using up all its >>>> heap space in the matter of a few minutes. I uninstalled the river and am >>>> performing no queries. How do I diagnose the problem? 2-3 minutes after >>>> starting, it runs out of heap space, and I'm not sure how to find out why. >>>> >>>> Here is the profile of memory usage: >>>> >>>> >>>> <https://lh6.googleusercontent.com/-La0i_IrQBLA/U9mIyZZDYLI/AAAAAAAAFx0/SfnYVdKvFAw/s1600/elasticsearch-memory.png> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> And here is the console output. You can see it takes less than a >>>> minute after starting to run out of memory. This isn't even enough time to >>>> examine the indices through marvel. >>>> >>>> C:\elasticsearch-1.1.1\bin>elasticsearch >>>> Picked up _JAVA_OPTIONS: -Djava.net.preferIPv4Stack=true >>>> [2014-07-30 16:59:02,579][INFO ][node ] [Texas >>>> Twister] version[1.1.1], pid[8572], build[f1585f0/201 >>>> 4-04-16T14:27:12Z] >>>> [2014-07-30 16:59:02,580][INFO ][node ] [Texas >>>> Twister] initializing ... >>>> [2014-07-30 16:59:02,600][INFO ][plugins ] [Texas >>>> Twister] loaded [marvel], sites [marvel] >>>> [2014-07-30 16:59:06,437][INFO ][node ] [Texas >>>> Twister] initialized >>>> [2014-07-30 16:59:06,437][INFO ][node ] [Texas >>>> Twister] starting ... >>>> [2014-07-30 16:59:06,691][INFO ][transport ] [Texas >>>> Twister] bound_address {inet[/0.0.0.0:9300]}, publish >>>> _address {inet[/192.168.0.6:9300]} >>>> [2014-07-30 16:59:09,862][INFO ][cluster.service ] [Texas >>>> Twister] new_master [Texas Twister][ShQRhZRFQnuZMTRCu >>>> vY9XQ][twilson-THINK][inet[/192.168.0.6:9300]], reason: zen-disco-join >>>> (elected_as_master) >>>> [2014-07-30 16:59:09,902][INFO ][discovery ] [Texas >>>> Twister] elasticsearch/ShQRhZRFQnuZMTRCuvY9XQ >>>> [2014-07-30 16:59:10,213][INFO ][http ] [Texas >>>> Twister] bound_address {inet[/0.0.0.0:9200]}, publish >>>> _address {inet[/192.168.0.6:9200]} >>>> [2014-07-30 16:59:11,631][INFO ][gateway ] [Texas >>>> Twister] recovered [65] indices into cluster_state >>>> java.lang.OutOfMemoryError: Java heap space >>>> Dumping heap to java_pid8572.hprof ... >>>> Heap dump file created [814218130 bytes in 14.202 secs] >>>> Exception in thread "elasticsearch[Texas Twister][generic][T#2]" >>>> java.lang.OutOfMemoryError: Java heap space >>>> at java.lang.Class.getDeclaredFields0(Native Method) >>>> at java.lang.Class.privateGetDeclaredFields(Class.java:2397) >>>> at java.lang.Class.getDeclaredFields(Class.java:1806) >>>> at org.apache.lucene.util.RamUsageEstimator. >>>> shallowSizeOfInstance(RamUsageEstimator.java:388) >>>> at org.apache.lucene.codecs.lucene42. >>>> Lucene42DocValuesProducer.<init>(Lucene42DocValuesProducer.java:101) >>>> at org.apache.lucene.codecs.lucene42.Lucene42NormsFormat. >>>> normsProducer(Lucene42NormsFormat.java:75) >>>> at org.apache.lucene.index.SegmentCoreReaders.<init>( >>>> SegmentCoreReaders.java:123) >>>> at org.apache.lucene.index.SegmentReader.<init>( >>>> SegmentReader.java:96) >>>> at org.apache.lucene.index.ReadersAndUpdates.getReader( >>>> ReadersAndUpdates.java:141) >>>> at org.apache.lucene.index.ReadersAndUpdates.getReadOnlyClone( >>>> ReadersAndUpdates.java:235) >>>> at org.apache.lucene.index.StandardDirectoryReader.open( >>>> StandardDirectoryReader.java:101) >>>> at org.apache.lucene.index.IndexWriter.getReader( >>>> IndexWriter.java:382) >>>> at org.apache.lucene.index.DirectoryReader.open( >>>> DirectoryReader.java:111) >>>> at org.apache.lucene.search.SearcherManager.<init>( >>>> SearcherManager.java:89) >>>> at org.elasticsearch.index.engine.internal.InternalEngine. >>>> buildSearchManager(InternalEngine.java:1471) >>>> at org.elasticsearch.index.engine.internal. >>>> InternalEngine.start(InternalEngine.java:279) >>>> at org.elasticsearch.index.shard.service.InternalIndexShard. >>>> performRecoveryPrepareForTranslog(InternalIndexShard >>>> .java:699) >>>> at org.elasticsearch.index.gateway.local. >>>> LocalIndexShardGateway.recover(LocalIndexShardGateway.java:205) >>>> at org.elasticsearch.index.gateway.IndexShardGatewayService$1. >>>> run(IndexShardGatewayService.java:197) >>>> at java.util.concurrent.ThreadPoolExecutor.runWorker( >>>> ThreadPoolExecutor.java:1145) >>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >>>> ThreadPoolExecutor.java:615) >>>> at java.lang.Thread.run(Thread.java:744) >>>> >>>> -- >>>> You received this message because you are subscribed to the Google >>>> Groups "elasticsearch" group. >>>> To unsubscribe from this group and stop receiving emails from it, send >>>> an email to elasticsearc...@googlegroups.com. >>>> >>>> To view this discussion on the web visit https://groups.google.com/d/ >>>> msgid/elasticsearch/21e9cc63-0c5d-4ea0-96a2-78d817b89236% >>>> 40googlegroups.com >>>> <https://groups.google.com/d/msgid/elasticsearch/21e9cc63-0c5d-4ea0-96a2-78d817b89236%40googlegroups.com?utm_medium=email&utm_source=footer> >>>> . >>>> For more options, visit https://groups.google.com/d/optout. >>>> >>> >>> -- >> You received this message because you are subscribed to the Google Groups >> "elasticsearch" group. >> To unsubscribe from this group and stop receiving emails from it, send an >> email to elasticsearc...@googlegroups.com <javascript:>. >> To view this discussion on the web visit >> https://groups.google.com/d/msgid/elasticsearch/662fdcc9-0ed0-4547-aaf1-26f12efaef91%40googlegroups.com >> >> <https://groups.google.com/d/msgid/elasticsearch/662fdcc9-0ed0-4547-aaf1-26f12efaef91%40googlegroups.com?utm_medium=email&utm_source=footer> >> . >> >> For more options, visit https://groups.google.com/d/optout. >> > > -- You received this message because you are subscribed to the Google Groups "elasticsearch" group. To unsubscribe from this group and stop receiving emails from it, send an email to elasticsearch+unsubscr...@googlegroups.com. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/31aa9efb-7a5a-47b3-86f6-2ac9de61edb1%40googlegroups.com. For more options, visit https://groups.google.com/d/optout.