On 11/06/2017 9:23 PM, Atin Mukherjee wrote:
Until and unless server side quorum is not enabled that's not correct. I/O path should be active even though management plane is down. We can still get this done by one node after another with out bringing down all glusterd instances at one go but just wanted to ensure the workaround is safe and clean.

Not quite sure of your wording here but I

 * brought down all glusterd with "systemctl stop
   glusterfs-server.service"  on each node
 * rm /var/lib/glusterd/peers/de673495-8cb2-4328-ba00-0419357c03d7 on
   each node
 * systemctl start glusterfs-server.service"  on each node


Several hundred shards needed to be healed after that, but all done now with no split-brain. And:

   root@vng:~# gluster peer status
   Number of Peers: 2

   Hostname: vnh.proxmox.softlog
   Uuid: 9eb54c33-7f79-4a75-bc2b-67111bf3eae7
   State: Peer in Cluster (Connected)

   Hostname: vnb.proxmox.softlog
   Uuid: 43a1bf8c-3e69-4581-8e16-f2e1462cfc36
   State: Peer in Cluster (Connected)


Which is good. Not in a position to test quorum by rebooting a node right now though :) but I'm going to assume its all good, probably test next weekend.

Thanks for all the help, much appreciated.

--
Lindsay Mathieson

_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to