¿So you're saying that those indexing operations associated with the
unsuccessful create-index requests will however succeed (i.e. all data will
be stored)?

On Sat, Jan 10, 2015 at 8:15 PM, joergpra...@gmail.com <
joergpra...@gmail.com> wrote:

> I think you can safely ignore "failed to process cluster event (acquire
> index lock) within 1s" in that case. These messages come from index
> creation requests that are submitted concurrently - only one request can
> succeed, the other will get stuck.
>
> Jörg
>
> On Fri, Jan 9, 2015 at 8:16 PM, Tom <tro...@despegar.com> wrote:
>
>> Sorry, didn't mean to say "the same indexing operation" but multiple
>> indexing operations (distinct data) on the same non-existent index.
>>
>> El viernes, 9 de enero de 2015, 16:13:52 (UTC-3), Tom escribió:
>>
>>> Well yes. We also have a cluster for the app where each node talks to
>>> the elastic cluster independently.
>>>
>>> Remember that we are not creating the index manually. Each app node
>>> issues an index operation on an index that may yet not exist and we expect
>>> ES to take care of the index creation on demand. Many processes may issue
>>> the same indexing operation on the ES cluster "simultaneously" and only one
>>> of them must succeed in triggering the index creation.
>>>
>>> Tom;
>>>
>>> El viernes, 9 de enero de 2015, 15:53:01 (UTC-3), Jörg Prante escribió:
>>>>
>>>> It seems there are more than one process trying to create the index, it
>>>> that possible?
>>>>
>>>> Jörg
>>>>
>>>> On Fri, Jan 9, 2015 at 7:16 PM, Tomas Andres Rossi <tro...@despegar.com
>>>> > wrote:
>>>>
>>>>> We enlarged our cluster to 5 nodes and now the QUORUM error message
>>>>> seems to have disappeared.
>>>>> "failed to process cluster event (acquire index lock) within 1s" kind
>>>>> of messages are still happening though.
>>>>>
>>>>> :(
>>>>>
>>>>> Tom;
>>>>>
>>>>> On Fri, Jan 9, 2015 at 3:11 PM, Tomas Andres Rossi <
>>>>> tro...@despegar.com> wrote:
>>>>>
>>>>>> We enlarged our cluster to 5 nodes and now the QUORUM error message
>>>>>> seems to have disappeared.
>>>>>> "failed to process cluster event (acquire index lock) within 1s"
>>>>>> kind of messages are still happening though.
>>>>>>
>>>>>> :(
>>>>>>
>>>>>> Tom;
>>>>>>
>>>>>> On Fri, Jan 9, 2015 at 1:25 PM, joerg...@gmail.com <
>>>>>> joerg...@gmail.com> wrote:
>>>>>>
>>>>>>> Exactly, with 3 nodes, the error will be gone.
>>>>>>>
>>>>>>> Please, always use an odd number of data nodes, in particular with
>>>>>>> replica > 0, in order not to confuse ES quorum formula, and also to 
>>>>>>> avoid
>>>>>>> split brains with minimun_master_nodes
>>>>>>>
>>>>>>> Jörg
>>>>>>>
>>>>>>> On Fri, Jan 9, 2015 at 3:17 PM, Tom <tarossi....@gmail.com> wrote:
>>>>>>>
>>>>>>>> Also, we have another cluster (for different purposes) that has 3
>>>>>>>> nodes but we didn't experience such errors with it (for this ES we 
>>>>>>>> create
>>>>>>>> indices on a daily basis).
>>>>>>>>
>>>>>>>> El jueves, 8 de enero de 2015, 16:23:12 (UTC-3), Tom escribió:
>>>>>>>>
>>>>>>>>> 4
>>>>>>>>>
>>>>>>>>> El jueves, 8 de enero de 2015 16:19:50 UTC-3, Jörg Prante escribió:
>>>>>>>>>>
>>>>>>>>>> How many nodes do you have in the cluster?
>>>>>>>>>>
>>>>>>>>>> Jörg
>>>>>>>>>>
>>>>>>>>>> On Thu, Jan 8, 2015 at 6:57 PM, Tom <tarossi....@gmail.com>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi, we'd been using ES for a while now. Specifically version
>>>>>>>>>>> 0.90.3. A couple of months ago we decided to migrate to the latest 
>>>>>>>>>>> version
>>>>>>>>>>> which was finally frozen to be 1.4.1. No data migration was 
>>>>>>>>>>> necessary
>>>>>>>>>>> because we have a redundant MongoDB, but yesterday we enabled data 
>>>>>>>>>>> writing
>>>>>>>>>>> to the new ES cluster. All was running smoothly when we noticed 
>>>>>>>>>>> that at
>>>>>>>>>>> o'clock times there were bursts of four or five log messages of the
>>>>>>>>>>> following kinds:
>>>>>>>>>>>
>>>>>>>>>>> Error indexing None into index ind-analytics-2015.01.08. Total
>>>>>>>>>>> elapsed time: 1065 ms. org.elasticsearch.cluster.metadata.
>>>>>>>>>>> ProcessClusterEventTimeoutException: failed to process cluster
>>>>>>>>>>> event (acquire index lock) within 1s
>>>>>>>>>>> at org.elasticsearch.cluster.metadata.MetaDataCreateIndexServic
>>>>>>>>>>> e$1.run(MetaDataCreateIndexService.java:148)
>>>>>>>>>>> ~[org.elasticsearch.elasticsearch-1.4.1.jar:na]
>>>>>>>>>>> at 
>>>>>>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>>>>>>>>> ~[na:1.7.0_17]
>>>>>>>>>>> at 
>>>>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>>>>>>>>> ~[na:1.7.0_17]
>>>>>>>>>>> at java.lang.Thread.run(Thread.java:722) ~[na:1.7.0_17]
>>>>>>>>>>>
>>>>>>>>>>> [ForkJoinPool-2-worker-15] c.d.i.p.ActorScatterGatherStrategy -
>>>>>>>>>>> Scattering to failed in 1043ms 
>>>>>>>>>>> org.elasticsearch.action.UnavailableShardsException:
>>>>>>>>>>> [ind-2015.01.08.00][0] Not enough active copies to meet write 
>>>>>>>>>>> consistency
>>>>>>>>>>> of [QUORUM] (have 1, needed 2). Timeout: [1s], request: index
>>>>>>>>>>> {[ind-2015.01.08.00][search][...]}
>>>>>>>>>>> at org.elasticsearch.action.support.replication.TransportShardR
>>>>>>>>>>> eplicationOperationAction$AsyncShardOperationAction.retryBec
>>>>>>>>>>> auseUnavailable(TransportShardReplicationOperationAction.java:784)
>>>>>>>>>>> ~[org.elasticsearch.elasticsearch-1.4.1.jar:na]
>>>>>>>>>>> at org.elasticsearch.action.support.replication.TransportShardR
>>>>>>>>>>> eplicationOperationAction$AsyncShardOperationAction.raiseFai
>>>>>>>>>>> lureIfHaveNotEnoughActiveShardCopies(TransportShardReplicati
>>>>>>>>>>> onOperationAction.java:776) ~[org.elasticsearch.elasticsea
>>>>>>>>>>> rch-1.4.1.jar:na]
>>>>>>>>>>> at org.elasticsearch.action.support.replication.TransportShardR
>>>>>>>>>>> eplicationOperationAction$AsyncShardOperationAction.performO
>>>>>>>>>>> nPrimary(TransportShardReplicationOperationAction.java:507)
>>>>>>>>>>> ~[org.elasticsearch.elasticsearch-1.4.1.jar:na]
>>>>>>>>>>> at org.elasticsearch.action.support.replication.TransportShardR
>>>>>>>>>>> eplicationOperationAction$AsyncShardOperationAction$1.run(Tr
>>>>>>>>>>> ansportShardReplicationOperationAction.java:419)
>>>>>>>>>>> ~[org.elasticsearch.elasticsearch-1.4.1.jar:na]
>>>>>>>>>>> at 
>>>>>>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>>>>>>>>> ~[na:1.7.0_17]
>>>>>>>>>>> at 
>>>>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>>>>>>>>> ~[na:1.7.0_17]
>>>>>>>>>>> at java.lang.Thread.run(Thread.java:722) ~[na:1.7.0_17]
>>>>>>>>>>>
>>>>>>>>>>> This occurs at o'clock times because we write over hour-based
>>>>>>>>>>> indices. For example, all writes from 18:00:00 to 18:59:59 of 01/08 
>>>>>>>>>>> goes to
>>>>>>>>>>> ind-2015.01.08.18. At 19:00:00 all writes will go to 
>>>>>>>>>>> ind-2015.01.08.19, and
>>>>>>>>>>> so on.
>>>>>>>>>>>
>>>>>>>>>>> With 0.90.3 version of ES, automatic index creation was working
>>>>>>>>>>> flawlessly (with no complaints) but the new version doesn't seem to 
>>>>>>>>>>> handle
>>>>>>>>>>> that feature very well. It looks like, when all those concurrent 
>>>>>>>>>>> writes
>>>>>>>>>>> competes to be the first to create the index, all but one fails. Of 
>>>>>>>>>>> course
>>>>>>>>>>> we could just create such indices manually to avoid this situation
>>>>>>>>>>> altogether, but this would only be a workaround for a feature that
>>>>>>>>>>> previously worked.
>>>>>>>>>>>
>>>>>>>>>>> Also, we use ES through the native Java client and the
>>>>>>>>>>> configuration for all our indices is
>>>>>>>>>>>
>>>>>>>>>>> settings = {
>>>>>>>>>>>   number_of_shards = 5,
>>>>>>>>>>>   number_of_replicas = 2
>>>>>>>>>>> }
>>>>>>>>>>>
>>>>>>>>>>> Any ideas?
>>>>>>>>>>>
>>>>>>>>>>> Thanks in advance,
>>>>>>>>>>> Tom;
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> You received this message because you are subscribed to the
>>>>>>>>>>> Google Groups "elasticsearch" group.
>>>>>>>>>>> To unsubscribe from this group and stop receiving emails from
>>>>>>>>>>> it, send an email to elasticsearc...@googlegroups.com.
>>>>>>>>>>> To view this discussion on the web visit
>>>>>>>>>>> https://groups.google.com/d/msgid/elasticsearch/4deefb09-bed
>>>>>>>>>>> 1-499a-b9fc-3ed4d78fc4c0%40googlegroups.com
>>>>>>>>>>> <https://groups.google.com/d/msgid/elasticsearch/4deefb09-bed1-499a-b9fc-3ed4d78fc4c0%40googlegroups.com?utm_medium=email&utm_source=footer>
>>>>>>>>>>> .
>>>>>>>>>>> For more options, visit https://groups.google.com/d/optout.
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>  --
>>>>>>>> You received this message because you are subscribed to the Google
>>>>>>>> Groups "elasticsearch" group.
>>>>>>>> To unsubscribe from this group and stop receiving emails from it,
>>>>>>>> send an email to elasticsearc...@googlegroups.com.
>>>>>>>> To view this discussion on the web visit
>>>>>>>> https://groups.google.com/d/msgid/elasticsearch/4b052ab5-
>>>>>>>> ab02-49bb-ad79-8e47f249e755%40googlegroups.com
>>>>>>>> <https://groups.google.com/d/msgid/elasticsearch/4b052ab5-ab02-49bb-ad79-8e47f249e755%40googlegroups.com?utm_medium=email&utm_source=footer>
>>>>>>>> .
>>>>>>>>
>>>>>>>> For more options, visit https://groups.google.com/d/optout.
>>>>>>>>
>>>>>>>
>>>>>>>  --
>>>>>>> You received this message because you are subscribed to a topic in
>>>>>>> the Google Groups "elasticsearch" group.
>>>>>>> To unsubscribe from this topic, visit https://groups.google.com/d/
>>>>>>> topic/elasticsearch/-H-sNVTSYbQ/unsubscribe.
>>>>>>> To unsubscribe from this group and all its topics, send an email to
>>>>>>> elasticsearc...@googlegroups.com.
>>>>>>> To view this discussion on the web visit
>>>>>>> https://groups.google.com/d/msgid/elasticsearch/
>>>>>>> CAKdsXoFaU96UN8YaguRs%2BMqD%2BtgypEWd6LP0CrisyFhh%2BTzjKw%
>>>>>>> 40mail.gmail.com
>>>>>>> <https://groups.google.com/d/msgid/elasticsearch/CAKdsXoFaU96UN8YaguRs%2BMqD%2BtgypEWd6LP0CrisyFhh%2BTzjKw%40mail.gmail.com?utm_medium=email&utm_source=footer>
>>>>>>> .
>>>>>>>
>>>>>>> For more options, visit https://groups.google.com/d/optout.
>>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Tom;
>>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Tom;
>>>>>
>>>>> --
>>>>> You received this message because you are subscribed to the Google
>>>>> Groups "elasticsearch" group.
>>>>> To unsubscribe from this group and stop receiving emails from it, send
>>>>> an email to elasticsearc...@googlegroups.com.
>>>>> To view this discussion on the web visit https://groups.google.com/d/
>>>>> msgid/elasticsearch/CAOs2X_cjcJNOt-aRZQc_3u7Xj0Knev%
>>>>> 3D66Z_6fxc43zKpRUNg_g%40mail.gmail.com
>>>>> <https://groups.google.com/d/msgid/elasticsearch/CAOs2X_cjcJNOt-aRZQc_3u7Xj0Knev%3D66Z_6fxc43zKpRUNg_g%40mail.gmail.com?utm_medium=email&utm_source=footer>
>>>>> .
>>>>>
>>>>> For more options, visit https://groups.google.com/d/optout.
>>>>>
>>>>
>>>>  --
>> You received this message because you are subscribed to the Google Groups
>> "elasticsearch" group.
>> To unsubscribe from this group and stop receiving emails from it, send an
>> email to elasticsearch+unsubscr...@googlegroups.com.
>> To view this discussion on the web visit
>> https://groups.google.com/d/msgid/elasticsearch/b6f30f8b-d258-4356-b7ef-ee6a0f9edf3c%40googlegroups.com
>> <https://groups.google.com/d/msgid/elasticsearch/b6f30f8b-d258-4356-b7ef-ee6a0f9edf3c%40googlegroups.com?utm_medium=email&utm_source=footer>
>> .
>>
>> For more options, visit https://groups.google.com/d/optout.
>>
>
>  --
> You received this message because you are subscribed to a topic in the
> Google Groups "elasticsearch" group.
> To unsubscribe from this topic, visit
> https://groups.google.com/d/topic/elasticsearch/-H-sNVTSYbQ/unsubscribe.
> To unsubscribe from this group and all its topics, send an email to
> elasticsearch+unsubscr...@googlegroups.com.
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/elasticsearch/CAKdsXoEh2%2BzGP_iSRSpa9wXaq0mzaRh74cmf9CkZMakk7KuHrg%40mail.gmail.com
> <https://groups.google.com/d/msgid/elasticsearch/CAKdsXoEh2%2BzGP_iSRSpa9wXaq0mzaRh74cmf9CkZMakk7KuHrg%40mail.gmail.com?utm_medium=email&utm_source=footer>
> .
>
> For more options, visit https://groups.google.com/d/optout.
>



-- 
Tom;

-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to elasticsearch+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/CAOs2X_fQjEbOFwkB237vDE1qErKcs1zmzUrWp93L5_5kLw83zA%40mail.gmail.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to