On 09/05/17 19:18, hvjunk wrote:

On 03 May 2017, at 07:49 , Jiffin Tony Thottan <[email protected] <mailto:[email protected]>> wrote:



On 02/05/17 15:27, hvjunk wrote:
Good day,

I’m busy setting up/testing NFS-HA with GlusterFS storage across VMs running Debian 8. GlusterFS volume to be "replica 3 arbiter 1"

In the NFS-ganesha information I’ve gleamed thus far, it mentions the "gluster volume set all cluster.enable-shared-storage enable”.

My first question is this: is that shared volume that gets created/setup, suppose to be resilient across reboots? It appears to not be the case in my test setup thus far, that that mount doesn’t get recreated/remounted after a reboot.

Following is the script which creates shared storage and mount it in the node, plus an entry will be added to /etc/fstab
https://github.com/gluster/glusterfs/blob/master/extras/hook-scripts/set/post/S32gluster_enable_shared_storage.sh

But there is a possibility such that, if glusterd(I hope u have enabled enabled glusterd service) is not started before
systemd tries mount the shared storage then it will fail.



Thanks for the systemd helper script
--
Jiffin

Thank Jiffin,

I since found that (1) you need to wait a bit for the cluster to “settle” with that script having executed, before you reboot the cluster (As you might see in my bitbucket ansible scripts in https://bitbucket.org/dismyne/gluster-ansibles/src ) … something to add in the manuals perhaps to warn people to wait for that script to finish before rebooting node/vm/server(s)?

(2) the default configuration, can’t bootstrap the /gluster_shared_storage volume/directory reliably from a clean shutdown-reboot of the whole cluster!!!

The problem: SystemD and it’s wanting to have the control over /etc/fstab and the mounting, and and and…. (and I’ll not empty my mind about L.P. based on his remarks in: https://github.com/systemd/systemd/issues/4468#issuecomment-255711912 after my struggling with this issue)


To have a reliably bootstrapped (from all nodes down booting up) I'm using the following SystemD service and helper script(s) to have the gluster cluster node mount their local mounts (like /gluster_shared_storage) reliably:

https://bitbucket.org/dismyne/gluster-ansibles/src/24b62dcc858364ee3744d351993de0e8e35c2680/ansible/files/glusterfsmounts.service-centos?at=master


https://bitbucket.org/dismyne/gluster-ansibles/src/24b62dcc858364ee3744d351993de0e8e35c2680/ansible/files/test-mounts.sh?at=master






--
Jiffin

If the mount is not resilient, ie. not recreated/mounted by glusterfs and neither added to the /etc/fstab by glusterfs, why the initial auto mount by glusterfs and not afterwards with a reboot?

The biggest “issue” I have found with glusterfs is the interaction with SystemD and mounts that fails and don’t get properly retried later (Will email separately on that issue) during bootstrapping of the cluster, and that is why I need to confirm the reasoning/etc. on this initial auto-mounting, but then the need to manually add it into the /etc/fstab

Thank you
Hendrik

_______________________________________________
Gluster-users mailing list
[email protected] <mailto:[email protected]>
http://lists.gluster.org/mailman/listinfo/gluster-users



_______________________________________________
Gluster-users mailing list
[email protected]
http://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to