We're seeing the same thing. ES 1.1.0, JDK 7u55 on Ubuntu 12.04, 5 data 
nodes, 3 separate masters, all are 15GB hosts with 7.5GB Heaps, storage is 
SSD. Data set is ~1.6TB according to Marvel.

Our daily indices are roughly 33GB in size, with 5 shards and 2 replicas. 
I'm still investigating what happened yesterday, but I do see in Marvel a 
large spike in the "Indices Current Merges" graph just before the node 
dies, and a corresponding increase in JVM Heap. When Heap hits 99% 
everything grinds to a halt. Restarting the node "fixes" the issue, but 
this is third or fourth time it's happened.

I'm still researching how to deal with this, but a couple of things I am 
looking at are:

   - increase the number of shards so that the segment merges stay smaller 
   (is that even a legitimate sentence?) I'm still reading through this page 
the 
   Index Module Merge page 
   
<http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/index-modules-merge.html>
 for 
   more details.
   - look at store level throttling 
   
<http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/index-modules-store.html#store-throttling>
   .

I would love to get some feedback on my ramblings. If I find anything more 
I'll update this thread.

cheers
mike




On Thursday, June 19, 2014 4:05:54 PM UTC-4, Bruce Ritchie wrote:
>
> Java 8 with G1GC perhaps? It'll have more overhead but perhaps it'll be 
> more consistent wrt pauses.
>
>
>
> On Wednesday, June 18, 2014 2:02:24 PM UTC-4, Eric Brandes wrote:
>>
>> I'd just like to chime in with a "me too".  Is the answer just more 
>> nodes?  In my case this is happening every week or so.
>>
>> On Monday, April 21, 2014 9:04:33 PM UTC-5, Brian Flad wrote:
>>
>> My dataset currently is 100GB across a few "daily" indices (~5-6GB and 15 
>> shards each). Data nodes are 12 CPU, 12GB RAM (6GB heap).
>>
>>
>> On Mon, Apr 21, 2014 at 6:33 PM, Mark Walkom <ma...@campaignmonitor.com> 
>> wrote:
>>
>> How big are your data sets? How big are your nodes?
>>
>> Regards,
>> Mark Walkom
>>
>> Infrastructure Engineer
>> Campaign Monitor
>> email: ma...@campaignmonitor.com
>> web: www.campaignmonitor.com
>>
>>
>> On 22 April 2014 00:32, Brian Flad <bfla...@gmail.com> wrote:
>>
>> We're seeing the same behavior with 1.1.1, JDK 7u55, 3 master nodes (2 
>> min master), and 5 data nodes. Interestingly, we see the repeated young GCs 
>> only on a node or two at a time. Cluster operations (such as recovering 
>> unassigned shards) grinds to a halt. After restarting a GCing node, 
>> everything returns to normal operation in the cluster.
>>
>> Brian F
>>
>>
>> On Wed, Apr 16, 2014 at 8:00 PM, Mark Walkom <ma...@campaignmonitor.com> 
>> wrote:
>>
>> In both your instances, if you can, have 3 master eligible nodes as it 
>> will reduce the likelihood of a split cluster as you will always have a 
>> majority quorum. Also look at discovery.zen.minimum_master_nodes to go with 
>> that.
>> However you may just be reaching the limit of your nodes, which means the 
>> best option is to add another node (which also neatly solves your split 
>> brain!).
>>
>> Ankush it would help if you can update java, most people recommend u25 
>> but we run u51 with no problems.
>>
>>
>>
>> Regards,
>> Mark Walkom
>>
>> Infrastructure Engineer
>> Campaign Monitor
>> email: ma...@campaignmonitor.com
>> web: www.campaignmonitor.com
>>
>>
>> On 17 April 2014 07:31, Dominiek ter Heide <domin...@gmail.com> wrote:
>>
>> We are seeing the same issue here. 
>>
>> Our environment:
>>
>> - 2 nodes
>> - 30GB Heap allocated to ES
>> - ~140GB of data
>> - 639 indices, 10 shards per index
>> - ~48M documents
>>
>> After starting ES everything is good, but after a couple of hours we see 
>> the Heap build up towards 96% on one node and 80% on the other. We then see 
>> the GC take very long on the 96% node:
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> TOuKgmlzaVaFVA][elasticsearch1.trend1.bottlenose.com][inet[/192.99.45.125
>> :9300]]])
>>
>> [2014-04-16 12:04:27,845][INFO ][discovery                ] 
>> [elasticsearch2.trend1] trend1/I3EHG_XjSayz2OsHyZpeZA
>>
>> [2014-04-16 12:04:27,850][INFO ][http                     ] [
>> elasticsearch2.trend1] bound_address {inet[/0.0.0.0:9200]}, 
>> publish_address {inet[/192.99.45.126:9200]}
>>
>> [2014-04-16 12:04:27,851][INFO ][node                     ] 
>> [elasticsearch2.trend1] started
>>
>> [2014-04-16 12:04:32,669][INFO ][indices.store            ] 
>> [elasticsearch2.trend1] updating indices.store.throttle.max_bytes_per_sec 
>> from [20mb] to [1gb], note, type is [MERGE]
>>
>> [2014-04-16 12:04:32,669][INFO ][cluster.routing.allocation.decider] 
>> [elasticsearch2.trend1] updating 
>> [cluster.routing.allocation.node_initial_primaries_recoveries] from [4] 
>> to [50]
>>
>> [2014-04-16 12:04:32,670][INFO ][indices.recovery         ] 
>> [elasticsearch2.trend1] updating [indices.recovery.max_bytes_per_sec] 
>> from [200mb] to [2gb]
>>
>> [2014-04-16 12:04:32,670][INFO ][cluster.routing.allocation.decider] 
>> [elasticsearch2.trend1] updating 
>> [cluster.routing.allocation.node_initial_primaries_recoveries] from [4] 
>> to [50]
>>
>> [2014-04-16 12:04:32,670][INFO ][cluster.routing.allocation.decider] 
>> [elasticsearch2.trend1] updating 
>> [cluster.routing.allocation.node_initial_primaries_recoveries] from [4] 
>> to [50]
>>
>> [2014-04-16 15:25:21,409][WARN ][monitor.jvm              ] 
>> [elasticsearch2.trend1] [gc][old][11876][106] duration [1.1m], 
>> collections [1]/[1.1m], total [1.1m]/[1.4m], memory [28.7gb]->[22gb]/[
>> 29.9gb], all_pools {[young] [67.9mb]->[268.9mb]/[665.6mb]}{[survivor] [
>> 60.5mb]->[0b]/[83.1mb]}{[old] [28.6gb]->[21.8gb]/[29.1gb]}
>>
>> [2014-04-16 16:02:32,523][WARN ][monitor.jvm              ] [
>> elasticsearch2.trend1] [gc][old][13996][144] duration [1.4m], 
>> collections [1]/[1.4m], total [1.4m]/[3m], memory [28.8gb]->[23.5gb]/[
>> 29.9gb], all_pools {[young] [21.8mb]->[238.2mb]/[665.6mb]}{[survivor] [
>> 82.4mb]->[0b]/[83.1mb]}{[old] [28.7gb]->[23.3gb]/[29.1gb]}
>>
>> [2014-04-16 16:14:12,386][WARN ][monitor.jvm              ] [
>> elasticsearch2.trend1] [gc][old][14603][155] duration [1.3m], 
>> collections [2]/[1.3m], total [1.3m]/[4.4m], memory [29.2gb]->[23.9gb]/[
>> 29.9gb], all_pools {[young] [289mb]->[161.3mb]/[665.6mb]}{[survivor] [
>> 58.3mb]->[0b]/[83.1mb]}{[old] [28.8gb]->[23.8gb]/[29.1gb]}
>>
>> [2014-04-16 16:17:55,480][WARN ][monitor.jvm              ] [
>> elasticsearch2.trend1] [gc][old][14745][158] duration [1.3m], 
>> collections [1]/[1.3m], total [1.3m]/[5.7m], memory [29.7gb]->[24.1gb]/[
>> 29.9gb], all_pools {[young] [633.8mb]->[149.7mb]/[665.6mb]}{[survivor] [
>> 68.6mb]->[0b]/[83.1mb]}{[old] [29gb]->[24gb]/[29.1gb]}
>>
>> [2014-04-16 16:21:17,950][WARN ][monitor.jvm              ] [
>> elasticsearch2.trend1] [gc][old][14857][161] duration [1.4m], 
>> collections [1]/[1.4m], total [1.4m]/[7.2m], memory [28.6gb]->[24.5gb]/[
>> 29.9gb], all_pools {[young] [27.7mb]->[154.8mb]/[665.6mb]}{[survivor] [
>> 83.1mb]->[0b]/[83.1mb]}{[old] [28.5gb]->[24.3gb]/[29.1gb]}
>>
>> [2014-04-16 16:24:48,776][WARN ][monitor.jvm              ] [
>> elasticsearch2.trend1] [gc][old][14978][164] duration [1.4m], 
>> collections [1]/[1.4m], total [1.4m]/[8.6m], memory [29.4gb]->[24.7gb]/[
>> 29.9gb], all_pools {[young] [475.5mb]->[125.1mb]/[665.6mb]}{[survivor] [
>> 68.9mb]->[0b]/[83.1mb]}{[old] [28.9gb]->[24.6gb]/[29.1gb]}
>>
>> [2014-04-16 16:26:54,801][WARN ][monitor.jvm              ] [
>> elasticsearch2.trend1] [gc][old][15021][165] duration [1.3m], 
>> collections [1]/[1.3m], total [1.3m]/[9.9m], memory [29.3gb]->[24.8gb]/[
>> 29.9gb], all_pools {[young] [391.8mb]->[151.1mb]/[665.6mb]}{[survivor] [
>> 62.4mb]->[0b]/[83.1mb]}{[old] [28.9gb]->[24.6gb]/[29.1gb]}
>>
>> [2014-04-16 16:30:45,393][WARN ][monitor.jvm              ] [
>> elasticsearch2.trend1] [gc][old][15170][168] duration [1.3m], 
>> collections [1]/[1.3m], total [1.3m]/[11.3m], memory [29.4gb]->[24.6gb]/[
>> 29.9gb], all_pools {[young] [320.3mb]->[186.7mb<span style="col
>>
>> ...
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to elasticsearch+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/5bf568eb-103e-4fee-8bd2-ba2b5bc76178%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to