Hello guys, I have 3 questions if you please. I have a HA NFS cluster - Centos 6.2, pacemaker, corosync, two NFS nodes plus 1 quorum node, in semi Active-Active configuration. By "semi", I mean that both NFS nodes are active and each of them is under normal circumstances exclusively responsible for one (out of two) Volume Group - using the ocf:heartbeat:LVM RA. Each LVM volume group lives on a dedicated multipath iscsi device, exported from a shared SAN.
I'm exporting a NFSv3/v4 export (/srv/nfs/software_repos directory). I need to make it available for 2 separate /21 networks as read-only, and for 3 different servers as read-write. I'm using the ocf:heartbeat:exportfs RA and it seems to me I have to use the ocf:heartbeat:exportfs RA 5 times. The configuration (only IP addresses changed) is here: http://pastebin.com/eHkgUv64 1) is there a way how to export this directory 5 times without defining 5 ocf:heartbeat:exportfs primitives? It's a lot of duplications... I search all the forums and I fear the ocf:heartbeat:exportfs simply supports only one host / network range. But maybe someone has been working on a patch? 2) while using the ocf:heartbeat:exportfs 5 times for the same directory, do I have to use the _same_ FSID (201 in my config) for all these 5 primitives (as I"m exporting the _same_ filesystem / directory)? I'm getting this warning when doing so WARNING: Resources p_exportfs_software_repos_ae1,p_exportfs_software_repos_ae2,p_exportfs_software_repos_buller,p_exportfs_software_repos_iap-mgmt,p_exportfs_software_repos_youyangs violate uniqueness for parameter "fsid": "201" Do you still want to commit? 3) wait_for_leasetime_on_stop - I believe this must be set to true when exporting NFSv4 with ocf:heartbeat:exportfs. http://www.linux-ha.org/doc/man-pages/re-ra-exportfs.html My 5 exportfs primitives reside in the same group: group g_nas02 p_lvm02 p_exportfs_software_repos_youyangs p_exportfs_software_repos_buller p_fs_software_repos p_exportfs_software_repos_ae1 p_exportfs_software_repos_ae2 p_exportfs_software_repos_iap-mgmt p_ip02 \ meta resource-stickiness="101" Even though I have the /proc/fs/nfsd/nfsv4gracetime set to 10 seconds, a failover of the NFS group from one NFS node to the second node would take more than 50 seconds, as it will be waiting for each ocf:heartbeat:exportfs resource sleeping 10 seconds 5 times. Is there any way of making them fail over / sleeping in parallel, instead of sequential? I workarounded this by setting wait_for_leasetime_on_stop="true" for only one of these (which I believe is safe and does the job it is expected to do - please correct me if I'm wrong). Thank you for your valuable comments. My Pacemaker configuration: http://pastebin.com/eHkgUv64 [root@irvine ~]# facter | egrep 'lsbdistid|lsbdistrelease' lsbdistid => CentOS lsbdistrelease => 6.2 [root@irvine ~]# rpm -qa | egrep 'pacemaker|corosync|agents' corosync-1.4.1-4.el6_2.2.x86_64 pacemaker-cli-1.1.6-3.el6.x86_64 pacemaker-libs-1.1.6-3.el6.x86_64 corosynclib-1.4.1-4.el6_2.2.x86_64 pacemaker-cluster-libs-1.1.6-3.el6.x86_64 pacemaker-1.1.6-3.el6.x86_64 fence-agents-3.1.5-10.el6_2.2.x86_64 resource-agents-3.9.2-7.el6.x86_64 with /usr/lib/ocf/resource.d/heartbeat/exportfs updated by hand from: https://github.com/ClusterLabs/resource-agents/commits/master/heartbeat/exportfs Thank you very much Marji Cermak _______________________________________________ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems