[Pacemaker] The state of the sg_persist RA (SCSI3 reservations)

2011-09-01 Thread Max Williams
Is it likely to make it in to the next version of the resource-agents package? I've tested it about as much as I can and it seems solid but would be keen to hear other peoples opinions and experiences. Cheers, Max Williams I

[Pacemaker] Reloading a resource after a failover

2011-09-28 Thread Max Williams
Hi, I have a pair of clustered DNS servers with a virtual IP (VIP) configured. The problem is that when the VIP fails over, named on the new host of the VIP will not listen on port 53/UDP of the VIP until it is reloaded (I think this is because this daemon uses UDP, not TCP). So I'd like to be

Re: [Pacemaker] Reloading a resource after a failover

2011-09-29 Thread Max Williams
:17 To: The Pacemaker cluster resource manager Subject: Re: [Pacemaker] Reloading a resource after a failover Put bind itself under pacemaker control. You can use LSB RA or OCF RA that I recently created. On Sep 28, 2011 10:46 AM, "Max Williams" mailto:max.willi...@betfair.com>>

Re: [Pacemaker] Reloading a resource after a failover

2011-09-29 Thread Max Williams
you use? LSB one doesn't support reload, OCF - does. You need to get OCF RA from github,. On Thu, Sep 29, 2011 at 3:04 AM, Max Williams mailto:max.willi...@betfair.com>> wrote: Yes this is what I would like to do. Ideally have named as a clone and then have an order like this: crm(liv

Re: [Pacemaker] Reloading a resource after a failover

2011-10-04 Thread Max Williams
a failover Here: https://github.com/ClusterLabs/resource-agents/blob/master/heartbeat/named Let me know how it works for you. On Sep 29, 2011 8:25 AM, "Max Williams" mailto:max.willi...@betfair.com>> wrote: > Yes I was using the LSB RA. Can you give me a link to the OCF RA on git

Re: [Pacemaker] Reloading a resource after a failover

2011-10-05 Thread Max Williams
an Haas mailto:flor...@hastexo.com>> wrote: On 2011-10-04 12:26, Max Williams wrote: > Thanks Serge. > > How exactly do I tell pacemaker to use this RA? Is there an 'import' > command I need to run? No, you just drop the resource agent in the appropriate provider directory (/usr/li

Re: [Pacemaker] pacemaker kills corosync

2011-10-05 Thread Max Williams
Hi Oliver, Normally pacemaker is loaded by Corosync, ie you just start corosync and leave pacemaker stopped and disabled via chkconfig. This is set in corosync.conf by this: service { # Load the Pacemaker Cluster Resource Manager name: pacemaker ver: 0 } http://www.clus

Re: [Pacemaker] pacemaker kills corosync

2011-10-05 Thread Max Williams
Actually I see now the documentation says use version 1 where I had previously used the version 0 method. Is version 1 the recommended way with the latest packages now? -Original Message- From: Max Williams [mailto:max.willi...@betfair.com] Sent: 05 October 2011 12:17 To: Pacemaker

Re: [Pacemaker] Stonith Device on a VM

2011-10-05 Thread Max Williams
I have, using fence_virsh. Here is my config for one fence device mapped to one node: primitive fence_cluster2.test stonith:fence_virsh params ipaddr="192.168.100.1" action="reboot" login="root" passwd="password" port="cluster-test-2" pcmk_host_list="cluster2.test" pcmk_host_check="static-list"

Re: [Pacemaker] Reloading a resource after a failover

2011-11-10 Thread Max Williams
moves or fails over named will get reloaded so it will listen on the VIP interface. Result. There is one minor issue and that is when pacemaker tries to stop rndc_reload resource it fails as the process does not stay running. It logs the failure but this doesn't seem to make any differenc

[Pacemaker] What versions should we be using and where to get packages?

2011-12-01 Thread Max Williams
Hi All, Are the very latest versions of pacemaker and corosync the most stable? I can find versions pacemaker-1.1.6-4 for corosync-1.4.2-1 for FC16 but packages for EL6 are hard to find. The clusterlabs website doesn't even have any EL6 packages. Is everyone else using older version or building th

[Pacemaker] Using SCSI reservations with Pacemaker

2011-05-05 Thread Max Williams
ent about this method when using a shared LUN. Do others feel the same? How have other people approached the problem of ensuring data consistency when using shared storage with Pacemaker/Corosync? SFEX? Thanks in advance, Max Williams

Re: [Pacemaker] Using SCSI reservations with Pacemaker

2011-05-06 Thread Max Williams
OK will do. Cheers, Max -Original Message- From: Florian Haas [mailto:florian.h...@linbit.com] Sent: 06 May 2011 08:25 To: The Pacemaker cluster resource manager; Evgeny Nifontov Subject: Re: [Pacemaker] Using SCSI reservations with Pacemaker Hi Max, On 2011-05-05 18:39, Max Williams

Re: [Pacemaker] Using SCSI reservations with Pacemaker

2011-05-10 Thread Max Williams
heers, Max -Original Message- From: Florian Haas [mailto:florian.h...@linbit.com] Sent: 06 May 2011 08:25 To: The Pacemaker cluster resource manager; Evgeny Nifontov Subject: Re: [Pacemaker] Using SCSI reservations with Pacemaker Hi Max, On 2011-05-05 18:39, Max Williams wrote: > I d

[Pacemaker] Failover when storage fails

2011-05-11 Thread Max Williams
Hi, I want to configure pacemaker to failover a group of resources and sg_persist (master/slave) when there is a problem with the storage but when I cause the iSCSI LUN to disappear simulating a failure, the cluster always gets stuck in this state: Last updated: Wed May 11 10:52:43 2011 Stack:

Re: [Pacemaker] Using SCSI reservations with Pacemaker

2011-05-11 Thread Max Williams
Max, On 2011-05-10 16:57, Max Williams wrote: > Hi Florian, > I've installed the sg_persist RA, configured it and it works as expected with > a Netapp iSCSI LUN, excellent! > > Here is my configuration: > primitive scsi_reservation ocf:heartbeat:sg_persist params

Re: [Pacemaker] Failover when storage fails

2011-05-11 Thread Max Williams
e" colocation MyApp_group_on_scsi_reservation inf: MyApp_group ms_MyApp_scsi_reservation:Master order MyApp_group_after_scsi_reservation inf: ms_MyApp_scsi_reservation:promote MyApp_group:start property $id="cib-bootstrap-options" \ dc-version="1.1.2-f059ec7ced7a86f18e5490b

Re: [Pacemaker] Using SCSI reservations with Pacemaker

2011-05-12 Thread Max Williams
Good point, thanks! Max -Original Message- From: Lars Marowsky-Bree [mailto:l...@suse.de] Sent: 12 May 2011 12:50 To: The Pacemaker cluster resource manager; Evgeny Nifontov; Florian Haas Subject: Re: [Pacemaker] Using SCSI reservations with Pacemaker On 2011-05-10T14:57:58, Max

Re: [Pacemaker] Failover when storage fails

2011-05-13 Thread Max Williams
al Message- From: Tim Serong [mailto:tser...@novell.com] Sent: 13 May 2011 04:22 To: The Pacemaker cluster resource manager (pacemaker@oss.clusterlabs.org) Subject: Re: [Pacemaker] Failover when storage fails On 5/12/2011 at 02:28 AM, Max Williams wrote: > After further testing even with st

Re: [Pacemaker] Preventing auto-fail-back

2011-05-18 Thread Max Williams
Hi Daniel, You might want to set "on-fail=standby" for the resource group or individual resources. This will put the host in to standby when a failure occurs thus preventing failback: http://www.clusterlabs.org/doc/en-US/Pacemaker/1.1/html/Pacemaker_Explained/s-resource-operations.html#s-resource

Re: [Pacemaker] Failover when storage fails

2011-06-02 Thread Max Williams
Just to update the list with the outcome of this issue, it's resolved in Pacemaker 1.1.5. Cheers, Max -Original Message- From: Max Williams [mailto:max.willi...@betfair.com] Sent: 13 May 2011 09:55 To: The Pacemaker cluster resource manager (pacemaker@oss.clusterlabs.org) Subjec

Re: [Pacemaker] (Linux-HA)resource-stickiness automatically move all resources in certain time

2011-06-08 Thread Max Williams
I'm not quite sure about your requirements because it sounds like you want to be able to perform a manual check of a host post failure AND have automatic fail back but perhaps you could remove the resource stickiness and look at setting migration-threshold to 1 and failure-timeout to 3600: http:

Re: [Pacemaker] RHEL 6.0 STONITH configuration

2011-06-09 Thread Max Williams
I don't have any ILOs but I do use RHEL6 with Dell DRACs. Perhaps something like this: primitive stonith-node1 stonith:fence_ilo \ params ipaddr=" ilo1" action="reboot" login="some_login " passwd="some_pass " pcmk_host_list=" node1" pcmk_host_check="static-list" \ meta target-ro