If you're using the mesos-init-wrapper you can write the IP to
/etc/mesos-master/ip and that flag will be set. This goes for all the
flags, and can be done for the slave as well in /etc/mesos-slave.


On 26 August 2014 10:18, Vinod Kone <vinodk...@gmail.com> wrote:

> From the logs, it looks like master is binding to its loopback address
> (127.0.0.1) and publishing that to ZK. So the slave is trying to reach the
> master on its loopback interface, which is failing.
>
> Start the master with "--ip" flag set to its visible ip (10.1.100.116).
> Mesosphere probably has a file (/etc/defaults/mesos-master?) to set these
> flags.
>
>
> On Mon, Aug 25, 2014 at 3:26 PM, Frank Hinek <frank.hi...@gmail.com>
> wrote:
>
>> Logs attached from master, slave, and zookeeper after a reboot of both
>> nodes.
>>
>>
>>
>>
>> On August 25, 2014 at 1:14:07 PM, Vinod Kone (vinodk...@gmail.com) wrote:
>>
>> what do the master and slave logs say?
>>
>>
>> On Mon, Aug 25, 2014 at 9:03 AM, Frank Hinek <frank.hi...@gmail.com>
>> wrote:
>>
>>>  I was able to get a single node environment setup on Ubuntu 14.04.1
>>> following this guide: http://mesosphere.io/learn/install_ubuntu_debian/
>>>
>>>  The single slave registered with the master via the local Zookeeper and
>>> I could run basic commands by posting to Marathon.
>>>
>>>  I then tried to build a multi node cluster following this guide:
>>> http://mesosphere.io/docs/mesosphere/getting-started/cloud-install/
>>>
>>>  The guide walks you through using the Mesosphere packages to install
>>> Mesos, Marathon, and Zookeeper one one node that will be the master and on
>>> the slave just Mesos.  You then disable automatic start of: mesos-slave on
>>> the master, mesos-master on the slave, and zookeeper on the slave.  It ends
>>> up looking like:
>>>
>>>  NODE 1 (MASTER):
>>>  - IP Address: 10.1.100.116
>>>  - mesos-master
>>>  - marathon
>>>  - zookeeper
>>>
>>>  NODE 2 (SLAVE):
>>>  - IP Address: 10.1.100.117
>>>  - mesos-slave
>>>
>>>  The issue I’m running into is that the slave rarely is able to register
>>> with the master using the Zookeeper.  I can never run any jobs from
>>> marathon (just trying a simple sleep 5 command).  Even when the slave does
>>> register the Mesos UI shows 1 “Deactivated” slave — it never goes active.
>>>
>>>  Here are the values I have for /etc/mesos/zk:
>>>
>>>  MASTER: zk://10.1.100.116:2181/mesos
>>>  SLAVE: zk://10.1.100.116:2181/mesos
>>>
>>>  Any ideas of what to troubleshoot?  Would greatly appreciate pointers.
>>>
>>>  Environment details:
>>>  - Ubuntu Server 14.04.1 running as VMs on ESXi 5.5U1
>>>  - Mesos: 0.20.0
>>>  - Marathon 0.6.1
>>>
>>>  There are no apparent connectivity issues, and I’m not having any
>>> problems with other VMs on the ESXi host.  All VM to VM communication is on
>>> the same VLAN and within the same host.
>>>
>>>  Zookeeper log on master (slave briefly registered so I tried to run a
>>> sleep 5 command from marathon and then the slave disconnected):
>>>
>>>  2014-08-25 11:50:34,976 - INFO  [NIOServerCxn.Factory:
>>> 0.0.0.0/0.0.0.0:2181:NIOServerCnxnFactory@197] - Accepted socket
>>> connection from /10.1.100.117:45778
>>> 2014-08-25 11:50:34,977 - WARN  [NIOServerCxn.Factory:
>>> 0.0.0.0/0.0.0.0:2181:ZooKeeperServer@793] - Connection request from old
>>> client /10.1.100.117:45778; will be dropped if server is in r-o mode
>>> 2014-08-25 11:50:34,977 - INFO  [NIOServerCxn.Factory:
>>> 0.0.0.0/0.0.0.0:2181:ZooKeeperServer@839] - Client attempting to
>>> establish new session at /10.1.100.117:45778
>>> 2014-08-25 11:50:34,978 - INFO  [SyncThread:0:ZooKeeperServer@595] -
>>> Established session 0x1480b22f7f0000c with negotiated timeout 10000 for
>>> client /10.1.100.117:45778
>>> 2014-08-25 11:51:05,724 - INFO  [ProcessThread(sid:0
>>> cport:-1)::PrepRequestProcessor@627] - Got user-level KeeperException
>>> when processing sessionid:0x1480b22f7f00001 type:create cxid:0x53faafa9
>>> zxid:0x49 txntype:-1 reqpath:n/a Error Path:/marathon Error:KeeperErrorCode
>>> = NodeExists for /marathon
>>> 2014-08-25 11:51:05,724 - INFO  [ProcessThread(sid:0
>>> cport:-1)::PrepRequestProcessor@627] - Got user-level KeeperException
>>> when processing sessionid:0x1480b22f7f00001 type:create cxid:0x53faafaa
>>> zxid:0x4a txntype:-1 reqpath:n/a Error Path:/marathon/state
>>> Error:KeeperErrorCode = NodeExists for /marathon/state
>>> 2014-08-25 11:51:09,145 - INFO  [ProcessThread(sid:0
>>> cport:-1)::PrepRequestProcessor@627] - Got user-level KeeperException
>>> when processing sessionid:0x1480b22f7f00001 type:create cxid:0x53faafb5
>>> zxid:0x4d txntype:-1 reqpath:n/a Error Path:/marathon Error:KeeperErrorCode
>>> = NodeExists for /marathon
>>> 2014-08-25 11:51:09,146 - INFO  [ProcessThread(sid:0
>>> cport:-1)::PrepRequestProcessor@627] - Got user-level KeeperException
>>> when processing sessionid:0x1480b22f7f00001 type:create cxid:0x53faafb6
>>> zxid:0x4e txntype:-1 reqpath:n/a Error Path:/marathon/state
>>> Error:KeeperErrorCode = NodeExists for /marathon/state
>>>
>>>
>>
>

Reply via email to