It seems there are more than one process trying to create the index, it
that possible?

Jörg

On Fri, Jan 9, 2015 at 7:16 PM, Tomas Andres Rossi <tro...@despegar.com>
wrote:

> We enlarged our cluster to 5 nodes and now the QUORUM error message seems
> to have disappeared.
> "failed to process cluster event (acquire index lock) within 1s" kind of
> messages are still happening though.
>
> :(
>
> Tom;
>
> On Fri, Jan 9, 2015 at 3:11 PM, Tomas Andres Rossi <tro...@despegar.com>
> wrote:
>
>> We enlarged our cluster to 5 nodes and now the QUORUM error message seems
>> to have disappeared.
>> "failed to process cluster event (acquire index lock) within 1s" kind of
>> messages are still happening though.
>>
>> :(
>>
>> Tom;
>>
>> On Fri, Jan 9, 2015 at 1:25 PM, joergpra...@gmail.com <
>> joergpra...@gmail.com> wrote:
>>
>>> Exactly, with 3 nodes, the error will be gone.
>>>
>>> Please, always use an odd number of data nodes, in particular with
>>> replica > 0, in order not to confuse ES quorum formula, and also to avoid
>>> split brains with minimun_master_nodes
>>>
>>> Jörg
>>>
>>> On Fri, Jan 9, 2015 at 3:17 PM, Tom <tarossi.despe...@gmail.com> wrote:
>>>
>>>> Also, we have another cluster (for different purposes) that has 3 nodes
>>>> but we didn't experience such errors with it (for this ES we create indices
>>>> on a daily basis).
>>>>
>>>> El jueves, 8 de enero de 2015, 16:23:12 (UTC-3), Tom escribió:
>>>>
>>>>> 4
>>>>>
>>>>> El jueves, 8 de enero de 2015 16:19:50 UTC-3, Jörg Prante escribió:
>>>>>>
>>>>>> How many nodes do you have in the cluster?
>>>>>>
>>>>>> Jörg
>>>>>>
>>>>>> On Thu, Jan 8, 2015 at 6:57 PM, Tom <tarossi....@gmail.com> wrote:
>>>>>>
>>>>>>> Hi, we'd been using ES for a while now. Specifically version 0.90.3.
>>>>>>> A couple of months ago we decided to migrate to the latest version which
>>>>>>> was finally frozen to be 1.4.1. No data migration was necessary because 
>>>>>>> we
>>>>>>> have a redundant MongoDB, but yesterday we enabled data writing to the 
>>>>>>> new
>>>>>>> ES cluster. All was running smoothly when we noticed that at o'clock 
>>>>>>> times
>>>>>>> there were bursts of four or five log messages of the following kinds:
>>>>>>>
>>>>>>> Error indexing None into index ind-analytics-2015.01.08. Total
>>>>>>> elapsed time: 1065 ms. org.elasticsearch.cluster.metadata.
>>>>>>> ProcessClusterEventTimeoutException: failed to process cluster
>>>>>>> event (acquire index lock) within 1s
>>>>>>> at org.elasticsearch.cluster.metadata.MetaDataCreateIndexService$1.
>>>>>>> run(MetaDataCreateIndexService.java:148) ~[org.elasticsearch.
>>>>>>> elasticsearch-1.4.1.jar:na]
>>>>>>> at 
>>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>>>>> ~[na:1.7.0_17]
>>>>>>> at 
>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>>>>> ~[na:1.7.0_17]
>>>>>>> at java.lang.Thread.run(Thread.java:722) ~[na:1.7.0_17]
>>>>>>>
>>>>>>> [ForkJoinPool-2-worker-15] c.d.i.p.ActorScatterGatherStrategy -
>>>>>>> Scattering to failed in 1043ms 
>>>>>>> org.elasticsearch.action.UnavailableShardsException:
>>>>>>> [ind-2015.01.08.00][0] Not enough active copies to meet write 
>>>>>>> consistency
>>>>>>> of [QUORUM] (have 1, needed 2). Timeout: [1s], request: index
>>>>>>> {[ind-2015.01.08.00][search][...]}
>>>>>>> at org.elasticsearch.action.support.replication.
>>>>>>> TransportShardReplicationOperationAction$AsyncShardOperationAction.
>>>>>>> retryBecauseUnavailable(TransportShardReplicationOperationAction.java:784)
>>>>>>> ~[org.elasticsearch.elasticsearch-1.4.1.jar:na]
>>>>>>> at org.elasticsearch.action.support.replication.
>>>>>>> TransportShardReplicationOperationAction$AsyncShardOperationAction.
>>>>>>> raiseFailureIfHaveNotEnoughActiveShardCopies(
>>>>>>> TransportShardReplicationOperationAction.java:776)
>>>>>>> ~[org.elasticsearch.elasticsearch-1.4.1.jar:na]
>>>>>>> at org.elasticsearch.action.support.replication.
>>>>>>> TransportShardReplicationOperationAction$AsyncShardOperationAction.
>>>>>>> performOnPrimary(TransportShardReplicationOperationAction.java:507)
>>>>>>> ~[org.elasticsearch.elasticsearch-1.4.1.jar:na]
>>>>>>> at org.elasticsearch.action.support.replication.
>>>>>>> TransportShardReplicationOperationAction$
>>>>>>> AsyncShardOperationAction$1.run(TransportShardReplicationOperationAction.java:419)
>>>>>>> ~[org.elasticsearch.elasticsearch-1.4.1.jar:na]
>>>>>>> at 
>>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>>>>> ~[na:1.7.0_17]
>>>>>>> at 
>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>>>>> ~[na:1.7.0_17]
>>>>>>> at java.lang.Thread.run(Thread.java:722) ~[na:1.7.0_17]
>>>>>>>
>>>>>>> This occurs at o'clock times because we write over hour-based
>>>>>>> indices. For example, all writes from 18:00:00 to 18:59:59 of 01/08 
>>>>>>> goes to
>>>>>>> ind-2015.01.08.18. At 19:00:00 all writes will go to ind-2015.01.08.19, 
>>>>>>> and
>>>>>>> so on.
>>>>>>>
>>>>>>> With 0.90.3 version of ES, automatic index creation was working
>>>>>>> flawlessly (with no complaints) but the new version doesn't seem to 
>>>>>>> handle
>>>>>>> that feature very well. It looks like, when all those concurrent writes
>>>>>>> competes to be the first to create the index, all but one fails. Of 
>>>>>>> course
>>>>>>> we could just create such indices manually to avoid this situation
>>>>>>> altogether, but this would only be a workaround for a feature that
>>>>>>> previously worked.
>>>>>>>
>>>>>>> Also, we use ES through the native Java client and the configuration
>>>>>>> for all our indices is
>>>>>>>
>>>>>>> settings = {
>>>>>>>   number_of_shards = 5,
>>>>>>>   number_of_replicas = 2
>>>>>>> }
>>>>>>>
>>>>>>> Any ideas?
>>>>>>>
>>>>>>> Thanks in advance,
>>>>>>> Tom;
>>>>>>>
>>>>>>> --
>>>>>>> You received this message because you are subscribed to the Google
>>>>>>> Groups "elasticsearch" group.
>>>>>>> To unsubscribe from this group and stop receiving emails from it,
>>>>>>> send an email to elasticsearc...@googlegroups.com.
>>>>>>> To view this discussion on the web visit
>>>>>>> https://groups.google.com/d/msgid/elasticsearch/4deefb09-
>>>>>>> bed1-499a-b9fc-3ed4d78fc4c0%40googlegroups.com
>>>>>>> <https://groups.google.com/d/msgid/elasticsearch/4deefb09-bed1-499a-b9fc-3ed4d78fc4c0%40googlegroups.com?utm_medium=email&utm_source=footer>
>>>>>>> .
>>>>>>> For more options, visit https://groups.google.com/d/optout.
>>>>>>>
>>>>>>
>>>>>>  --
>>>> You received this message because you are subscribed to the Google
>>>> Groups "elasticsearch" group.
>>>> To unsubscribe from this group and stop receiving emails from it, send
>>>> an email to elasticsearch+unsubscr...@googlegroups.com.
>>>> To view this discussion on the web visit
>>>> https://groups.google.com/d/msgid/elasticsearch/4b052ab5-ab02-49bb-ad79-8e47f249e755%40googlegroups.com
>>>> <https://groups.google.com/d/msgid/elasticsearch/4b052ab5-ab02-49bb-ad79-8e47f249e755%40googlegroups.com?utm_medium=email&utm_source=footer>
>>>> .
>>>>
>>>> For more options, visit https://groups.google.com/d/optout.
>>>>
>>>
>>>  --
>>> You received this message because you are subscribed to a topic in the
>>> Google Groups "elasticsearch" group.
>>> To unsubscribe from this topic, visit
>>> https://groups.google.com/d/topic/elasticsearch/-H-sNVTSYbQ/unsubscribe.
>>> To unsubscribe from this group and all its topics, send an email to
>>> elasticsearch+unsubscr...@googlegroups.com.
>>> To view this discussion on the web visit
>>> https://groups.google.com/d/msgid/elasticsearch/CAKdsXoFaU96UN8YaguRs%2BMqD%2BtgypEWd6LP0CrisyFhh%2BTzjKw%40mail.gmail.com
>>> <https://groups.google.com/d/msgid/elasticsearch/CAKdsXoFaU96UN8YaguRs%2BMqD%2BtgypEWd6LP0CrisyFhh%2BTzjKw%40mail.gmail.com?utm_medium=email&utm_source=footer>
>>> .
>>>
>>> For more options, visit https://groups.google.com/d/optout.
>>>
>>
>>
>>
>> --
>> Tom;
>>
>
>
>
> --
> Tom;
>
> --
> You received this message because you are subscribed to the Google Groups
> "elasticsearch" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to elasticsearch+unsubscr...@googlegroups.com.
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/elasticsearch/CAOs2X_cjcJNOt-aRZQc_3u7Xj0Knev%3D66Z_6fxc43zKpRUNg_g%40mail.gmail.com
> <https://groups.google.com/d/msgid/elasticsearch/CAOs2X_cjcJNOt-aRZQc_3u7Xj0Knev%3D66Z_6fxc43zKpRUNg_g%40mail.gmail.com?utm_medium=email&utm_source=footer>
> .
>
> For more options, visit https://groups.google.com/d/optout.
>

-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to elasticsearch+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/CAKdsXoG9QZ14j%2Ba96pfP8BJvLqJADRxTyYGgnqM5DAwS%3Dbt0nQ%40mail.gmail.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to