On Tue, Jul 8, 2014, at 06:06, Andrew Beekhof wrote: > > On 5 Jul 2014, at 1:00 am, Giuseppe Ragusa <giuseppe.rag...@hotmail.com> > wrote: > > > From: and...@beekhof.net > > Date: Fri, 4 Jul 2014 22:50:28 +1000 > > To: pacemaker@oss.clusterlabs.org > > Subject: Re: [Pacemaker] Pacemaker 1.1: cloned stonith resources require > > --force to be added to levels > > > > > > On 4 Jul 2014, at 1:29 pm, Giuseppe Ragusa <giuseppe.rag...@hotmail.com> > > wrote: > > > > > >> Hi all, > > > >> while creating a cloned stonith resource > > > > > > > > Any particular reason you feel the need to clone it? > > > > > > In the end, I suppose it's only a "purist mindset" :) because it is a PDU > > > whose power outlets control both nodes, so > > > its resource "should be" active (and monitored) on both nodes > > > "independently". > > > I understand that it would work anyway, leaving it not cloned and not > > > location-constrained > > > just as regular, "dedicated" stonith devices would not need to be > > > location-constrained, right? > > > > > > >> for multi-level STONITH on a fully-up-to-date CentOS 6.5 > > > >> (pacemaker-1.1.10-14.el6_5.3.x86_64): > > > >> > > > >> pcs cluster cib stonith_cfg > > > >> pcs -f stonith_cfg stonith create pdu1 fence_apc action="off" \ > > > >> ipaddr="pdu1.verolengo.privatelan" login="cluster" passwd="test" \ > > > >> > > > >> pcmk_host_map="cluster1.verolengo.privatelan:3,cluster1.verolengo.privatelan:4,cluster2.verolengo.privatelan:6,cluster2.verolengo.privatelan:7" > > > >> \ > > > >> pcmk_host_check="static-list" > > > >> pcmk_host_list="cluster1.verolengo.privatelan,cluster2.verolengo.privatelan" > > > >> op monitor interval="240s" > > > >> pcs -f stonith_cfg resource clone pdu1 pdu1Clone > > > >> pcs -f stonith_cfg stonith level add 2 cluster1.verolengo.privatelan > > > >> pdu1Clone > > > >> pcs -f stonith_cfg stonith level add 2 cluster2.verolengo.privatelan > > > >> pdu1Clone > > > >> > > > >> > > > >> the last 2 lines do not succeed unless I add the option "--force" and > > > >> even so I still get errors when issuing verify: > > > >> > > > >> [root@cluster1 ~]# pcs stonith level verify > > > >> Error: pdu1Clone is not a stonith id > > > > > > > > If you check, I think you'll find there is no such resource as > > > > 'pdu1Clone'. > > > > I don't believe pcs lets you decide what the clone name is. > > > > > > You're right! (obviously ;> ) > > > It's been automatically named pdu1-clone > > > > > > I suppose that there's still too much crmsh in my memory :) > > > > > > Anyway, removing the stonith level (to start from scratch) and using the > > > correct clone name does not change the result: > > > > > > [root@cluster1 etc]# pcs -f stonith_cfg stonith level add 2 > > > cluster1.verolengo.privatelan pdu1-clone > > > Error: pdu1-clone is not a stonith id (use --force to override) > > > > I bet we didn't think of that. > > What if you just do: > > > > pcs -f stonith_cfg stonith level add 2 cluster1.verolengo.privatelan pdu1 > > > > Does that work? > > > > ------------------------------------------------------------------------ > > > > Yes, no errors at all and verify successful.
This initially passed by as a simple check for general sanity, while now, on second read, I think you were suggesting that I could clone as usual then configure with the primitive resource (which I usually avoid when working with regular clones) and it should automatically use instead the clone "at runtime", correct? > > Remember that a full real test (to verify actual second level functionality > > in presence of first level failure) > > is still pending for both the plain and cloned setup. > > > > Apropos: I read through the list archives that stonith resources (being > > resources, after all) > > could themselves cause fencing (!) if failing (start, monitor, stop) > > stop just unsets a flag in stonithd. > start does perform a monitor op though, which could fail. > > but by default only stop failure would result in fencing. I though that start-failure-is-fatal was true by default, but maybe not for stonith resources. > > and that an ad-hoc > > on-fail setting could be used to prevent that. > > Maybe my aforementioned naive testing procedure (pull the iLO cable) could > > provoke that? > > _shouldnt_ do so > > > Would you suggest to configure such an on-fail option? > > again, shouldn't be necessary Thanks again. Regards, Giuseppe > > Many thanks again for your help (and all your valuable work, of course!). > > > > Regards, > > Giuseppe > > _______________________________________________ > > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > > > Project Home: http://www.clusterlabs.org > > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > > Bugs: http://bugs.clusterlabs.org > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > Email had 1 attachment: > + signature.asc > 1k (application/pgp-signature) -- Giuseppe Ragusa giuseppe.rag...@fastmail.fm _______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org