I just ran into this same issue last week. There is a JIRA ticket on it at Intel but in a nutshell mkfs.lustre on zfs will only record the last mgsnode you specify in your command. To add an additional fail node you can use the zfs command to update the configuration:
zfs set lustre:failover.node=<mgsnode1@<network>:<mgsnode2>@<network> <zpool name>/<zpool volume> Hope this helps. ==== Joe Mervini Sandia National Laboratories High Performance Computing 505.844.6770 jame...@sandia.gov On Dec 1, 2014, at 10:41 AM, Ron Croonenberg <r...@lanl.gov> wrote: > Hello, > > We're running/building Lustre on ZFS and I noticed that when using > mkfs.lustre on a zpool, when creating the MDT, with two --mgsnid parameters, > one for the MGS and one for the MGS fail over, causes a problem resulting in > not being able to mount the MDT. (I think it tries to connect to the fail > over instead of the actual MGS) > > In ldiskfs it just works and I can mount the MDT. > > For MDT/MDS fail over, is it enough to just specify the --failnode parameter > or does the --mgsnid parameter need to be specified too? > > thanks, > > Ron > _______________________________________________ > Lustre-discuss mailing list > Lustre-discuss@lists.lustre.org > http://lists.lustre.org/mailman/listinfo/lustre-discuss _______________________________________________ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss