Oh - BTW. You will need to do the same thing with your OSTs for setting both the mgsnodes.
Also, you can use zfs show <zpool name>/<zpool volume> to get the same info as you would with tunefs.lustre ==== Joe Mervini Sandia National Laboratories High Performance Computing 505.844.6770 jame...@sandia.gov On Dec 1, 2014, at 11:38 AM, Joe Mervini <jame...@sandia.gov> wrote: > I just ran into this same issue last week. There is a JIRA ticket on it at > Intel but in a nutshell mkfs.lustre on zfs will only record the last mgsnode > you specify in your command. To add an additional fail node you can use the > zfs command to update the configuration: > > zfs set lustre:failover.node=<mgsnode1@<network>:<mgsnode2>@<network> <zpool > name>/<zpool volume> > > Hope this helps. > ==== > > Joe Mervini > Sandia National Laboratories > High Performance Computing > 505.844.6770 > jame...@sandia.gov > > > > On Dec 1, 2014, at 10:41 AM, Ron Croonenberg <r...@lanl.gov> wrote: > >> Hello, >> >> We're running/building Lustre on ZFS and I noticed that when using >> mkfs.lustre on a zpool, when creating the MDT, with two --mgsnid parameters, >> one for the MGS and one for the MGS fail over, causes a problem resulting in >> not being able to mount the MDT. (I think it tries to connect to the fail >> over instead of the actual MGS) >> >> In ldiskfs it just works and I can mount the MDT. >> >> For MDT/MDS fail over, is it enough to just specify the --failnode parameter >> or does the --mgsnid parameter need to be specified too? >> >> thanks, >> >> Ron >> _______________________________________________ >> Lustre-discuss mailing list >> Lustre-discuss@lists.lustre.org >> http://lists.lustre.org/mailman/listinfo/lustre-discuss > _______________________________________________ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss