Look into the curator, which should help:
https://github.com/elasticsearch/curator

If you have just a single development instance, perhaps Marvel is an
overkill. Do you need historical metrics? If not, just use some other
plugin such as head/bigdesk/hq.

Cheers,

Ivan




On Thu, Jul 31, 2014 at 10:52 AM, Tom Wilson <twilson...@gmail.com> wrote:

> What exactly do I need to delete and how do I do it?
>
>
> On Wednesday, July 30, 2014 5:45:03 PM UTC-7, Mark Walkom wrote:
>
>> Unless you are attached to the stats you have in the marvel index for
>> today it might be easier to delete them than try to recover the unavailable
>> shards.
>>
>> Regards,
>> Mark Walkom
>>
>> Infrastructure Engineer
>> Campaign Monitor
>> email: ma...@campaignmonitor.com
>> web: www.campaignmonitor.com
>>
>>
>> On 31 July 2014 10:36, Tom Wilson <twils...@gmail.com> wrote:
>>
>>> Upping to 1GB, memory usage seems to level off at 750MB, but there's a
>>> problem in there somewhere. I'm getting a failure message, and the marvel
>>> dashboard isn't able to fetch.
>>>
>>>
>>> C:\elasticsearch-1.1.1\bin>elasticsearch
>>> Picked up _JAVA_OPTIONS: -Djava.net.preferIPv4Stack=true
>>> [2014-07-30 17:33:27,138][INFO ][node                     ] [Mondo]
>>> version[1.1.1], pid[10864], build[f1585f0/2014-04-16
>>> T14:27:12Z]
>>> [2014-07-30 17:33:27,139][INFO ][node                     ] [Mondo]
>>> initializing ...
>>> [2014-07-30 17:33:27,163][INFO ][plugins                  ] [Mondo]
>>> loaded [ldap-river, marvel], sites [marvel]
>>> [2014-07-30 17:33:30,731][INFO ][node                     ] [Mondo]
>>> initialized
>>> [2014-07-30 17:33:30,731][INFO ][node                     ] [Mondo]
>>> starting ...
>>> [2014-07-30 17:33:31,027][INFO ][transport                ] [Mondo]
>>> bound_address {inet[/0.0.0.0:9300]}, publish_address
>>>  {inet[/192.168.0.6:9300]}
>>> [2014-07-30 17:33:34,202][INFO ][cluster.service          ] [Mondo]
>>> new_master [Mondo][liyNQAHAS0-8f-qDDqa5Rg][twilson-T
>>> HINK][inet[/192.168.0.6:9300]], reason: zen-disco-join
>>> (elected_as_master)
>>> [2014-07-30 17:33:34,239][INFO ][discovery                ] [Mondo]
>>> elasticsearch/liyNQAHAS0-8f-qDDqa5Rg
>>>  [2014-07-30 17:33:34,600][INFO ][http                     ] [Mondo]
>>> bound_address {inet[/0.0.0.0:9200]}, publish_address
>>>  {inet[/192.168.0.6:9200]}
>>> [2014-07-30 17:33:35,799][INFO ][gateway                  ] [Mondo]
>>> recovered [66] indices into cluster_state
>>> [2014-07-30 17:33:35,815][INFO ][node                     ] [Mondo]
>>> started
>>> [2014-07-30 17:33:39,823][DEBUG][action.search.type       ] [Mondo] All
>>> shards failed for phase: [query_fetch]
>>> [2014-07-30 17:33:39,830][DEBUG][action.search.type       ] [Mondo] All
>>> shards failed for phase: [query_fetch]
>>> [2014-07-30 17:33:39,837][DEBUG][action.search.type       ] [Mondo] All
>>> shards failed for phase: [query_fetch]
>>> [2014-07-30 17:33:39,838][DEBUG][action.search.type       ] [Mondo] All
>>> shards failed for phase: [query_fetch]
>>> [2014-07-30 17:33:43,973][DEBUG][action.search.type       ] [Mondo] All
>>> shards failed for phase: [query_fetch]
>>> [2014-07-30 17:33:44,212][DEBUG][action.search.type       ] [Mondo] All
>>> shards failed for phase: [query_fetch]
>>> [2014-07-30 17:33:44,357][DEBUG][action.search.type       ] [Mondo] All
>>> shards failed for phase: [query_fetch]
>>> [2014-07-30 17:33:44,501][DEBUG][action.search.type       ] [Mondo] All
>>> shards failed for phase: [query_fetch]
>>> [2014-07-30 17:33:53,294][DEBUG][action.search.type       ] [Mondo] All
>>> shards failed for phase: [query_fetch]
>>> [2014-07-30 17:33:53,309][DEBUG][action.search.type       ] [Mondo] All
>>> shards failed for phase: [query_fetch]
>>> [2014-07-30 17:33:53,310][DEBUG][action.search.type       ] [Mondo] All
>>> shards failed for phase: [query_fetch]
>>> [2014-07-30 17:33:53,310][DEBUG][action.search.type       ] [Mondo] All
>>> shards failed for phase: [query_fetch]
>>> [2014-07-30 17:34:03,281][DEBUG][action.search.type       ] [Mondo] All
>>> shards failed for phase: [query_fetch]
>>> [2014-07-30 17:34:03,283][DEBUG][action.search.type       ] [Mondo] All
>>> shards failed for phase: [query_fetch]
>>> [2014-07-30 17:34:03,286][DEBUG][action.search.type       ] [Mondo] All
>>> shards failed for phase: [query_fetch]
>>> [2014-07-30 17:34:45,662][ERROR][marvel.agent.exporter    ] [Mondo]
>>> create failure (index:[.marvel-2014.07.31] type: [no
>>> de_stats]): UnavailableShardsException[[.marvel-2014.07.31][0] [2]
>>> shardIt, [0] active : Timeout waiting for [1m], reque
>>> st: org.elasticsearch.action.bulk.BulkShardRequest@39b65640]
>>>
>>>
>>>
>>> On Wednesday, July 30, 2014 5:30:29 PM UTC-7, Mark Walkom wrote:
>>>
>>>> Up that to 1GB and see if it starts.
>>>> 512MB is pretty tiny, you're better off starting at 1/2GB if you can.
>>>>
>>>> Regards,
>>>> Mark Walkom
>>>>
>>>> Infrastructure Engineer
>>>> Campaign Monitor
>>>> email: ma...@campaignmonitor.com
>>>> web: www.campaignmonitor.com
>>>>
>>>>
>>>> On 31 July 2014 10:28, Tom Wilson <twils...@gmail.com> wrote:
>>>>
>>>>>  JDK 1.7.0_51
>>>>>
>>>>> It has 512MB of heap, which was enough -- I've been running it like
>>>>> that for the past few months, and I only have two indexes and around
>>>>> 300-400 documents. This is a development instance I'm running on my local
>>>>> machine. This only happened when I started it today.
>>>>>
>>>>> -tom
>>>>>
>>>>>
>>>>> On Wednesday, July 30, 2014 5:16:11 PM UTC-7, Mark Walkom wrote:
>>>>>
>>>>>> What java version? How much heap have you allocated and how much RAM
>>>>>> on the server?
>>>>>>
>>>>>> Basically you have too much data for the heap size, so increasing it
>>>>>> will help.
>>>>>>
>>>>>> Regards,
>>>>>> Mark Walkom
>>>>>>
>>>>>> Infrastructure Engineer
>>>>>> Campaign Monitor
>>>>>> email: ma...@campaignmonitor.com
>>>>>> web: www.campaignmonitor.com
>>>>>>
>>>>>>
>>>>>> On 31 July 2014 10:11, Tom Wilson <twils...@gmail.com> wrote:
>>>>>>
>>>>>>>  Help! Elasticsearch was working fine, but now it's using up all
>>>>>>> its heap space in the matter of a few minutes. I uninstalled the river 
>>>>>>> and
>>>>>>> am performing no queries. How do I diagnose the problem? 2-3 minutes 
>>>>>>> after
>>>>>>> starting, it runs out of heap space, and I'm not sure how to find out 
>>>>>>> why.
>>>>>>>
>>>>>>> Here is the profile of memory usage:
>>>>>>>
>>>>>>>
>>>>>>> <https://lh6.googleusercontent.com/-La0i_IrQBLA/U9mIyZZDYLI/AAAAAAAAFx0/SfnYVdKvFAw/s1600/elasticsearch-memory.png>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> And  here is the console output. You can see it takes less than a
>>>>>>> minute after starting to run out of memory. This isn't even enough time 
>>>>>>> to
>>>>>>> examine the indices through marvel.
>>>>>>>
>>>>>>> C:\elasticsearch-1.1.1\bin>elasticsearch
>>>>>>> Picked up _JAVA_OPTIONS: -Djava.net.preferIPv4Stack=true
>>>>>>> [2014-07-30 16:59:02,579][INFO ][node                     ] [Texas
>>>>>>> Twister] version[1.1.1], pid[8572], build[f1585f0/201
>>>>>>> 4-04-16T14:27:12Z]
>>>>>>> [2014-07-30 16:59:02,580][INFO ][node                     ] [Texas
>>>>>>> Twister] initializing ...
>>>>>>> [2014-07-30 16:59:02,600][INFO ][plugins                  ] [Texas
>>>>>>> Twister] loaded [marvel], sites [marvel]
>>>>>>> [2014-07-30 16:59:06,437][INFO ][node                     ] [Texas
>>>>>>> Twister] initialized
>>>>>>> [2014-07-30 16:59:06,437][INFO ][node                     ] [Texas
>>>>>>> Twister] starting ...
>>>>>>> [2014-07-30 16:59:06,691][INFO ][transport                ] [Texas
>>>>>>> Twister] bound_address {inet[/0.0.0.0:9300]}, publish
>>>>>>> _address {inet[/192.168.0.6:9300]}
>>>>>>> [2014-07-30 16:59:09,862][INFO ][cluster.service          ] [Texas
>>>>>>> Twister] new_master [Texas Twister][ShQRhZRFQnuZMTRCu
>>>>>>> vY9XQ][twilson-THINK][inet[/192.168.0.6:9300]], reason:
>>>>>>> zen-disco-join (elected_as_master)
>>>>>>> [2014-07-30 16:59:09,902][INFO ][discovery                ] [Texas
>>>>>>> Twister] elasticsearch/ShQRhZRFQnuZMTRCuvY9XQ
>>>>>>> [2014-07-30 16:59:10,213][INFO ][http                     ] [Texas
>>>>>>> Twister] bound_address {inet[/0.0.0.0:9200]}, publish
>>>>>>> _address {inet[/192.168.0.6:9200]}
>>>>>>> [2014-07-30 16:59:11,631][INFO ][gateway                  ] [Texas
>>>>>>> Twister] recovered [65] indices into cluster_state
>>>>>>> java.lang.OutOfMemoryError: Java heap space
>>>>>>> Dumping heap to java_pid8572.hprof ...
>>>>>>> Heap dump file created [814218130 bytes in 14.202 secs]
>>>>>>> Exception in thread "elasticsearch[Texas Twister][generic][T#2]"
>>>>>>> java.lang.OutOfMemoryError: Java heap space
>>>>>>>         at java.lang.Class.getDeclaredFields0(Native Method)
>>>>>>>         at java.lang.Class.privateGetDeclaredFields(Class.java:2397)
>>>>>>>         at java.lang.Class.getDeclaredFields(Class.java:1806)
>>>>>>>         at org.apache.lucene.util.RamUsageEstimator.
>>>>>>> shallowSizeOfInstance(RamUsageEstimator.java:388)
>>>>>>>         at org.apache.lucene.codecs.lucene42.
>>>>>>> Lucene42DocValuesProducer.<init>(Lucene42DocValuesProducer.java:101)
>>>>>>>         at org.apache.lucene.codecs.lucene42.Lucene42NormsFormat.
>>>>>>> normsProducer(Lucene42NormsFormat.java:75)
>>>>>>>         at org.apache.lucene.index.SegmentCoreReaders.<init>(
>>>>>>> SegmentCoreReaders.java:123)
>>>>>>>         at org.apache.lucene.index.SegmentReader.<init>(
>>>>>>> SegmentReader.java:96)
>>>>>>>         at org.apache.lucene.index.ReadersAndUpdates.getReader(
>>>>>>> ReadersAndUpdates.java:141)
>>>>>>>         at org.apache.lucene.index.ReadersAndUpdates.
>>>>>>> getReadOnlyClone(ReadersAndUpdates.java:235)
>>>>>>>         at org.apache.lucene.index.StandardDirectoryReader.open(
>>>>>>> StandardDirectoryReader.java:101)
>>>>>>>          at org.apache.lucene.index.IndexWriter.getReader(
>>>>>>> IndexWriter.java:382)
>>>>>>>         at org.apache.lucene.index.DirectoryReader.open(
>>>>>>> DirectoryReader.java:111)
>>>>>>>         at org.apache.lucene.search.SearcherManager.<init>(
>>>>>>> SearcherManager.java:89)
>>>>>>>         at org.elasticsearch.index.engine.internal.InternalEngine.
>>>>>>> buildSearchManager(InternalEngine.java:1471)
>>>>>>>         at org.elasticsearch.index.engine.internal.InternalEngine.
>>>>>>> start(InternalEngine.java:279)
>>>>>>>         at org.elasticsearch.index.shard.service.InternalIndexShard.
>>>>>>> performRecoveryPrepareForTranslog(InternalIndexShard
>>>>>>> .java:699)
>>>>>>>         at org.elasticsearch.index.gateway.local.
>>>>>>> LocalIndexShardGateway.recover(LocalIndexShardGateway.java:205)
>>>>>>>         at org.elasticsearch.index.gateway.
>>>>>>> IndexShardGatewayService$1.run(IndexShardGatewayService.java:197)
>>>>>>>         at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>>>> ThreadPoolExecutor.java:1145)
>>>>>>>          at java.util.concurrent.ThreadPoolExecutor$Worker.run(
>>>>>>> ThreadPoolExecutor.java:615)
>>>>>>>         at java.lang.Thread.run(Thread.java:744)
>>>>>>>
>>>>>>> --
>>>>>>> You received this message because you are subscribed to the Google
>>>>>>> Groups "elasticsearch" group.
>>>>>>> To unsubscribe from this group and stop receiving emails from it,
>>>>>>> send an email to elasticsearc...@googlegroups.com.
>>>>>>>
>>>>>>> To view this discussion on the web visit
>>>>>>> https://groups.google.com/d/msgid/elasticsearch/21e9cc63-0c5
>>>>>>> d-4ea0-96a2-78d817b89236%40googlegroups.com
>>>>>>> <https://groups.google.com/d/msgid/elasticsearch/21e9cc63-0c5d-4ea0-96a2-78d817b89236%40googlegroups.com?utm_medium=email&utm_source=footer>
>>>>>>> .
>>>>>>> For more options, visit https://groups.google.com/d/optout.
>>>>>>>
>>>>>>
>>>>>>  --
>>>>> You received this message because you are subscribed to the Google
>>>>> Groups "elasticsearch" group.
>>>>> To unsubscribe from this group and stop receiving emails from it, send
>>>>> an email to elasticsearc...@googlegroups.com.
>>>>> To view this discussion on the web visit https://groups.google.com/d/
>>>>> msgid/elasticsearch/662fdcc9-0ed0-4547-aaf1-26f12efaef91%40goo
>>>>> glegroups.com
>>>>> <https://groups.google.com/d/msgid/elasticsearch/662fdcc9-0ed0-4547-aaf1-26f12efaef91%40googlegroups.com?utm_medium=email&utm_source=footer>
>>>>> .
>>>>>
>>>>> For more options, visit https://groups.google.com/d/optout.
>>>>>
>>>>
>>>>  --
>>> You received this message because you are subscribed to the Google
>>> Groups "elasticsearch" group.
>>> To unsubscribe from this group and stop receiving emails from it, send
>>> an email to elasticsearc...@googlegroups.com.
>>> To view this discussion on the web visit https://groups.google.com/d/
>>> msgid/elasticsearch/31aa9efb-7a5a-47b3-86f6-2ac9de61edb1%
>>> 40googlegroups.com
>>> <https://groups.google.com/d/msgid/elasticsearch/31aa9efb-7a5a-47b3-86f6-2ac9de61edb1%40googlegroups.com?utm_medium=email&utm_source=footer>
>>> .
>>>
>>> For more options, visit https://groups.google.com/d/optout.
>>>
>>
>>  --
> You received this message because you are subscribed to the Google Groups
> "elasticsearch" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to elasticsearch+unsubscr...@googlegroups.com.
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/elasticsearch/5caa9794-8105-417b-b84e-5cb28dcaa488%40googlegroups.com
> <https://groups.google.com/d/msgid/elasticsearch/5caa9794-8105-417b-b84e-5cb28dcaa488%40googlegroups.com?utm_medium=email&utm_source=footer>
> .
>
> For more options, visit https://groups.google.com/d/optout.
>

-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to elasticsearch+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/CALY%3DcQBQbkgtD%3DSUoxZr-k-fPC2YmgXc2LOSjZ6J8R7w%2B2dTkg%40mail.gmail.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to