Re: [Pacemaker] [Linux-ha-dev] CFP: Linux Plumbers Mini-Conf on High-Availability/Clustering

2010-08-10 Thread Guillaume Chanaud
Envoyé depuis un mobile Samsung Lars Marowsky-Bree a écrit : >On 2010-08-04T15:59:27, Lars Marowsky-Bree wrote: > >> Hi all, >> >> there will (hopefully!) be a mini-conference on HA/Clustering at this >> year's LPC in Cambridge, MA, Nov 3-5th. > >Just a quick reminder, there've not been many

Re: [Pacemaker] [Linux-ha-dev] CFP: Linux Plumbers Mini-Conf on High-Availability/Clustering

2010-08-10 Thread Lars Marowsky-Bree
On 2010-08-04T15:59:27, Lars Marowsky-Bree wrote: > Hi all, > > there will (hopefully!) be a mini-conference on HA/Clustering at this > year's LPC in Cambridge, MA, Nov 3-5th. Just a quick reminder, there've not been many proposals submitted yet. If the trend continues, the mini-conf slot might

[Pacemaker] Question on fix for Bug lf#2433

2010-08-10 Thread Stepan, Troy
Hi, I applied the changeset for Bug lf#2433 (No services should be stopped until probes finish) to pacemaker 1.0.7-4.1. Either I misinterpreted the bugfix or it's not working that way I thought it would. While both of my dummy rscs are running, issuing a clean to dummy0 stops dummy1 (dummy1 i

Re: [Pacemaker] An issue about failcount of resources when start action failed

2010-08-10 Thread Dejan Muhamedagic
Hi, On Tue, Aug 10, 2010 at 06:29:45PM +0800, Jingcheng zhang wrote: > Dear Beekhof, > I configured a two node cluster with clone resource A. When the > resource A start failed, I saw the failed actions (start operation) in > crm_mon but the failcount displayed by "crm resource" is 0. Is

Re: [Pacemaker] stonith sbd problem

2010-08-10 Thread Dejan Muhamedagic
Hi, On Tue, Aug 10, 2010 at 10:16:05AM +0200, philipp.achmuel...@arz.at wrote: > hi, > > following configuration: > > node lnx0047a > node lnx0047b > primitive lnx0101a ocf:heartbeat:KVM \ > params name="lnx0101a" \ > meta allow-migrate="1" target-role="Started" \ > op mi

Re: [Pacemaker] Preventing resource from becoming inactive

2010-08-10 Thread Dejan Muhamedagic
Hi, On Tue, Aug 10, 2010 at 07:56:47AM +0200, Torsten Bronger wrote: > Hallöchen! > > Sometimes Pacemaker just switches off my Lighttpd. How do you mean "just switches off"? > It becomes > inactive and is never reanimated. Only restarting the Heartbeat > service helps. How can I tell Pacemaker

Re: [Pacemaker] rsc_order plus resource_set problem

2010-08-10 Thread Dejan Muhamedagic
Hi, On Tue, Aug 10, 2010 at 11:00:29AM +0800, Michael Fung wrote: > Hello Dejan, > > > > > That cannot be the case, because tabs are treated as space. The > > problem comes from specifying 'sequential="true"' which, because > > default, is not generated hence the original XML and the > > generat

[Pacemaker] An issue about failcount of resources when start action failed

2010-08-10 Thread Jingcheng zhang
Dear Beekhof, I configured a two node cluster with clone resource A. When the resource A start failed, I saw the failed actions (start operation) in crm_mon but the failcount displayed by "crm resource" is 0. Is the result expected? Why isn't the fail count of resource greater than 1?

Re: [Pacemaker] Preventing resource from becoming inactive

2010-08-10 Thread Torsten Bronger
Hallöchen! David Mohr writes: > On Tue, 10 Aug 2010 07:56:47 +0200, Torsten Bronger > wrote: > >> Sometimes Pacemaker just switches off my Lighttpd. It becomes >> inactive and is never reanimated. Only restarting the Heartbeat >> service helps. How can I tell Pacemaker to retry to restart >> l

Re: [Pacemaker] Preventing resource from becoming inactive

2010-08-10 Thread David Mohr
Hi! On Tue, 10 Aug 2010 07:56:47 +0200, Torsten Bronger wrote: > Hallöchen! > > Sometimes Pacemaker just switches off my Lighttpd. It becomes > inactive and is never reanimated. Only restarting the Heartbeat > service helps. How can I tell Pacemaker to retry to restart > lighttpd without ever

[Pacemaker] stonith sbd problem

2010-08-10 Thread philipp . achmueller
hi, following configuration: node lnx0047a node lnx0047b primitive lnx0101a ocf:heartbeat:KVM \ params name="lnx0101a" \ meta allow-migrate="1" target-role="Started" \ op migrate_from interval="0" timeout="3600s" \ op migrate_to interval="0" timeout="3600s" \