Thanks Alex.

About the context, we use spark on mesos and marathon to launch some 
elastisearch,

I kill each leader one-by-one.

By the way as I said, we are on a config Mesos-Master on ubuntu 12, and 
mesos-slave on ubuntu 14, to reproduce this comportement.

When I deploy only on Ubuntu 14 master+slave, the issue disappear …

Fred






On 09 Dec 2015, at 16:30, Alex Rukletsov 
<a...@mesosphere.com<mailto:a...@mesosphere.com>> wrote:

Frederic,

I have skimmed through the logs and they are do not seem to be complete 
(especially for master1). Could you please say what task has been killed (id) 
and which master failover triggered that? I see at least three failovers in the 
logs : ). Also, could you please share some background about your setup? I 
believe you're on systemd, do you use docker tasks?

To connect our conversation to particular events, let me post here the chain of 
(potentially) interesting events and some info I mined from the logs.
master1: 192.168.37.59 ?
master2: 192.168.37.58
master3: 192.168.37.104

timestamp   observed by   event
13:48:38     master1          master1 killed by sigterm
13:48:48     master2,3       new leader elected (192.168.37.104), id=5
13:49:25     master2          master2 killed by sigterm
13:50:44     master2,3       new leader elected (192.168.37.59), id=7
14:23:34     master1          master1 killed by sigterm
14:23:44     master2,3       new leader elected (192.168.37.58), id=8

One interesting thing I cannot understand is why master3 did not commit suicide 
when it lost leadership?


On Mon, Dec 7, 2015 at 4:08 PM, Frederic LE BRIS 
<fleb...@pagesjaunes.fr<mailto:fleb...@pagesjaunes.fr>> wrote:
With the context .. sorry



Reply via email to