----- Original Message -----
> From: "Andrew Beekhof" <and...@beekhof.net>
> To: "The Pacemaker cluster resource manager" <pacemaker@oss.clusterlabs.org>
> Sent: Monday, July 1, 2013 7:14:35 AM
> Subject: Re: [Pacemaker] Fixed! - Re: Problem with dual-PDU fencing node      
> with redundant PSUs
> 
> 
> On 01/07/2013, at 5:32 PM, Vladislav Bogdanov <bub...@hoster-ok.com>
> wrote:
> 
> > 29.06.2013 02:22, Andrew Beekhof wrote:
> >> 
> >> On 29/06/2013, at 12:22 AM, Digimer <li...@alteeve.ca> wrote:
> >> 
> >>> On 06/28/2013 06:21 AM, Andrew Beekhof wrote:
> >>>> 
> >>>> On 28/06/2013, at 5:22 PM, Lars Marowsky-Bree <l...@suse.com>
> >>>> wrote:
> >>>> 
> >>>>> On 2013-06-27T12:53:01, Digimer <li...@alteeve.ca> wrote:
> >>>>> 
> >>>>>> primitive fence_n01_psu1_off stonith:fence_apc_snmp \
> >>>>>>      params ipaddr="an-p01" pcmk_reboot_action="off" port="1"
> >>>>>> pcmk_host_list="an-c03n01.alteeve.ca"
> >>>>>> primitive fence_n01_psu1_on stonith:fence_apc_snmp \
> >>>>>>      params ipaddr="an-p01" pcmk_reboot_action="on" port="1"
> >>>>>> pcmk_host_list="an-c03n01.alteeve.ca"
> >>>>> 
> >>>>> So every device twice, including location constraints? I see
> >>>>> potential
> >>>>> for optimization by improving how the fence code handles this
> >>>>> ... That's
> >>>>> abhorrently complex. (And I'm not sure the 'action' parameter
> >>>>> ought to
> >>>>> be overwritten.)
> >>>> 
> >>>> I'm not crazy about it either because it means the device is
> >>>> tied to a specific command.
> >>>> But it seems to be something all the RHCS people try to do...
> >>> 
> >>> Maybe something in the rhcs water cooler made us all mad... ;)
> >>> 
> >>>>> Glad you got it working, though.
> >>>>> 
> >>>>>> location loc_fence_n01_ipmi fence_n01_ipmi -inf:
> >>>>>> an-c03n01.alteeve.ca
> >>>>> [...]
> >>>>> 
> >>>>> I'm not sure you need any of these location constraints, by the
> >>>>> way. Did
> >>>>> you test if it works without them?
> >>>>> 
> >>>>>> Again, this is after just one test. I will want to test it
> >>>>>> several more
> >>>>>> times before I consider it reliable. Ideally, I would love to
> >>>>>> hear
> >>>>>> Andrew or others confirm this looks sane/correct.
> >>>>> 
> >>>>> It looks correct, but not quite sane. ;-) That seems not to be
> >>>>> something you can address, though. I'm thinking that fencing
> >>>>> topology
> >>>>> should be smart enough to, if multiple fencing devices are
> >>>>> specified, to
> >>>>> know how to expand them to "first all off (if off fails
> >>>>> anywhere, it's a
> >>>>> failure), then all on (if on fails, it is not a failure)".
> >>>>> That'd
> >>>>> greatly simplify the syntax.
> >>>> 
> >>>> The RH agents have apparently already been updated to support
> >>>> multiple ports.
> >>>> I'm really not keen on having the stonith-ng doing this.
> >>> 
> >>> This doesn't help people who have dual power rails/PDUs for power
> >>> redundancy.
> >> 
> >> I'm yet to be convinced that having two PDUs is helping those
> >> people in the first place.
> >> If it were actually useful, I suspect more than two/three people
> >> would have asked for it in the last decade.
> > 
> > I'm just silently waiting for this to happen.
> 
> Rarely a good plan.
> Better to make my life so miserable that implementing it seems like a
> vacation in comparison :)
> 
> > Although I use different fencing scheme (and plan to use even more
> > different one), that is very nice fall-back path for me. And I
> > strongly
> > prefer all complexities like reboot -> off-off-on-on to be hidden
> > from
> > the configuration. Naturally, that is task for the entity which has
> > whole picture of what to do - stonithd. Just my 'IMHO'.
> 
> If the tides of public opinion change, then yes, stonithd is the
> place.
> But I can't justify the effort for only a handful of deployments.

+1 here

> 
> > 
> > And, to PSU/PDU. I, like Digimer, always separate power circuits as
> > much
> > as possible. Of course I always use redundant PSUs.

and +1 here

Jake

> > 
> > Vladislav
> > 
> > 
> > 
> > _______________________________________________
> > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> > 
> > Project Home: http://www.clusterlabs.org
> > Getting started:
> > http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
> 
> 
> _______________________________________________
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started:
> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
> 
> 

_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to