- Original Message -
> From: "Andrew Beekhof"
> To: "The Pacemaker cluster resource manager"
> Sent: Thursday, June 5, 2014 6:36:51 PM
> Subject: Re: [Pacemaker] Upstart support in cluster-glue 1.0.11 and
> pacemaker 1.1.11
>
>
> On 6
Hello,
I'm working on backporting the latest cluster-glue (1.0.11), pacemaker
(1.1.11), corosync, and related packages to Ubuntu 12.04:
https://launchpad.net/~xespackages/+archive/clustertesting
I've installed these packages and started testing, however I seem to be unable
to add an upstart res
- Original Message -
> From: "David Vossel"
> To: "The Pacemaker cluster resource manager"
> Sent: Monday, April 21, 2014 10:39:22 AM
> Subject: Re: [Pacemaker] Best practice for quorum nodes
> > Is there a better way to designate a node as a quorum node, so that
> > resources
> > do not
Hello,
I've read several guides about how to configure a 3-node cluster with one node
that can't actually run the resources, but just serves as a quorum node. One
practice for configuring this node is to put it in "standby", which prevents it
from running resources. In my experience, this seems
Hello,
Reading through the "Highly Available NFS storage with DRBD and Pacemaker"
guide
available here, http://www.linbit.com/en/downloads/tech-guides, I see that it
recommends that the lsb:nfs-kernel-server resource is configured as a clone:
clone cl_lsb_nfsserver p_lsb_nfsserver
I also found
David/Andrew,
Once 1.1.11 final is released, is it considered the new stable series of
Pacemaker, or should 1.1.10 still be used in very stable/critical production
environments?
Thanks,
Andrew
- Original Message -
> From: "David Vossel"
> To: "The Pacemaker cluster resource manager"
- Original Message -
> From: "Andrew Beekhof"
> To: "The Pacemaker cluster resource manager"
> Sent: Thursday, August 8, 2013 7:42:22 PM
> Subject: Re: [Pacemaker] Announce: Pacemaker 1.1.10 now available
>
>
> On 08/08/2013, at 11:48 PM, Andr
- Original Message -
> From: "Andrew Beekhof"
> To: "The Pacemaker cluster resource manager"
> Sent: Thursday, August 8, 2013 2:35:53 AM
> Subject: Re: [Pacemaker] Announce: Pacemaker 1.1.10 now available
>
>
> On 08/08/2013, at 5:13 PM, Vladislav Bogdanov
> wrote:
>
> > 26.07.2013 03
- Original Message -
> From: "Michael Furman"
> To: pacemaker@oss.clusterlabs.org
> Sent: Tuesday, June 11, 2013 3:19:52 AM
> Subject: Re: [Pacemaker] Can I use Pacemaker release 1.1.8 for production
> clusters?
>
>
>
> Thank you for the reply!
> What problem you encounter in a product
- Original Message -
> From: "Florian Crouzat"
> To: pacemaker@oss.clusterlabs.org
> Sent: Monday, June 10, 2013 10:01:35 AM
> Subject: Re: [Pacemaker] Can I use Pacemaker release 1.1.8 for production
> clusters?
>
> Le 10/06/2013 16:46, Michael Furman a écrit :
> > Hi all!
> >
> > Accor
Andrew,
I'd recommend adding more than one host to your p_ping resource and see if that
improves the situation. When I had this problem, I observed better behavior
after adding more than one IP to the list of hosts and changing the p_ping
location constraint to be as follows:
location loc_run_o
- Original Message -
> From: "Robert Parsons"
> To: "The Pacemaker cluster resource manager"
> Sent: Wednesday, May 1, 2013 4:03:46 PM
> Subject: [Pacemaker] Pacemaker/Corosync on Ubuntu 12.04
>
>
>
> We're wanting to build a web server farm with approx 18 nodes. We're
> aware of scala
- Original Message -
> From: "Andrew Beekhof"
> To: "The Pacemaker cluster resource manager"
> Sent: Thursday, April 25, 2013 6:52:08 PM
> Subject: Re: [Pacemaker] best setup for corosync + pacemaker in ubuntu 12.04
>
>
> On 25/04/2013, at 4:03 PM, Angel L. Mateo wrote:
>
> > Hello ev
- Original Message -
> From: "Ante Karamatić"
> To: pacemaker@oss.clusterlabs.org
> Sent: Thursday, February 14, 2013 3:57:38 AM
> Subject: Re: [Pacemaker] Reason for cluster resource migration
>
> On 13.02.2013 16:27, Andrew Martin wrote::
>
> > Unf
- Original Message -
> From: "Andrew Beekhof"
> To: "The Pacemaker cluster resource manager"
> Sent: Tuesday, February 12, 2013 10:52:23 PM
> Subject: Re: [Pacemaker] Reason for cluster resource migration
>
> On Wed, Feb 13, 2013 at 2:04 AM, Andr
- Original Message -
> From: "Ante Karamatic"
> To: pacemaker@oss.clusterlabs.org
> Sent: Wednesday, February 13, 2013 1:53:34 AM
> Subject: Re: [Pacemaker] Reason for cluster resource migration
>
> On 13.02.2013 05:57, Andrew Beekhof wrote::
>
> > This link has some useful info:
> >
>
w Beekhof
> wrote:
> > On Tue, Feb 12, 2013 at 3:01 PM, Andrew Beekhof
> > wrote:
> >> On Tue, Feb 12, 2013 at 1:40 PM, Andrew Martin
> >> wrote:
> >>> Hello,
> >>>
> >>> Unfortunately this same failure occurred again tonig
perhaps not
quickly enough to prevent the STONITH and resource migration?
Here is the pe-core dump file mentioned in the log:
http://sources.xes-inc.com/downloads/pe-core.bz2
Thanks,
Andrew
- Original Message -
> From: "Andrew Martin"
> To: "The Pacem
Hello,
I am running a 3-node Pacemaker (1.1.8) + Corosync (2.1.0) cluster on Ubuntu
12.04. Two of the nodes are "real" nodes, hosting a DRBD filesystem mount and
some daemons:
http://pastebin.com/n1sNMhuE
The third node cannot run resources and acts as a quorum node in standby.
Recently, the no
- Original Message -
> From: "Andrew Beekhof"
> To: "The Pacemaker cluster resource manager"
> Sent: Thursday, December 6, 2012 8:36:27 PM
> Subject: Re: [Pacemaker] Reason for cluster resource migration
>
> On Wed, Dec 5, 2012 at 8:29 AM, Andrew Ma
Andreas,
- Original Message -
> From: "Andreas Kurz"
> To: pacemaker@oss.clusterlabs.org
> Sent: Friday, December 21, 2012 6:22:57 PM
> Subject: Re: [Pacemaker] Best way to recover from failed STONITH?
>
> On 12/21/2012 07:47 PM, Andrew Martin wrote:
> >
> On 12/21/2012 04:18 PM, Andrew Martin wrote:
> > Hello,
> >
> > Yesterday a power failure took out one of the nodes and its STONITH
> > device (they share an upstream power source) in a 3-node
> > active/passive cluster (Corosync 2.1.0, Pacemaker 1.1.8).
Hello,
Yesterday a power failure took out one of the nodes and its STONITH device
(they share an upstream power source) in a 3-node active/passive cluster
(Corosync 2.1.0, Pacemaker 1.1.8). After logging into the cluster, I saw that
the STONITH operation had given up in failure and that none of
Brett,
The ocf:heartbeat:pingd resource agent is used to monitor network availability.
This resource agent
is actually deprecated - the recommended replacement is ocf:pacemaker:pingd.
You can use ocf:pacemaker:pingd with a location constraint to move resources
away from a node if it
loses netw
Hello,
I am running a 3-node Pacemaker cluster (2 "real" nodes and 1 quorum node in
standby) on Ubuntu 12.04 server (amd64) with Pacemaker 1.1.8 and Corosync
2.1.0. My cluster configuration is:
http://pastebin.com/6TPkWtbt
Recently, pengine died on storage0 (where the resources were runnin
Art,
The Linbit Tech Guides are nice for this purpose as they give step-by-step
instructions for configuring a simple example cluster. Since you mentioned
using DRBD in your setup, I would recommend the DRBD Users Guide, which also
has information specific to Pacemaker:
http://www.drbd.org/u
Art,
I would recommend building your cluster with the latest version of Pacemaker,
1.1.8, and a newer release of Corosync as well. I also recently built a cluster
on Ubuntu 12.04 and encountered bugs in the older versions of Corosync and
Pacemaker in the repositories.
There are two stable
the IPC possibility.
Thanks,
Andrew
- Original Message -
From: "Jan Friesse"
To: "Andrew Martin"
Cc: "Angus Salkeld" , disc...@corosync.org,
pacemaker@oss.clusterlabs.org
Sent: Thursday, November 8, 2012 7:39:45 AM
Subject: Re: [corosync] [Pacemaker] Cor
midnight.blogspot.com/2012/07/cluster-building-ubuntu-1204-revised.html
Thanks,
Andrew
- Original Message -
From: "Jan Friesse"
To: "Andrew Martin"
Cc: "Angus Salkeld" , disc...@corosync.org,
pacemaker@oss.clusterlabs.org
Sent: Wednesday, November 7,
rage0.
What else can I do to debug this problem? Or, should I just try to downgrade to
corosync 1.4.2 (the version available in the Ubuntu repositories)?
Thanks,
Andrew
- Original Message -
From: "Andrew Martin"
To: "Angus Salkeld"
Cc: disc...@corosync.org, pace
quot;
To: pacemaker@oss.clusterlabs.org, disc...@corosync.org
Sent: Friday, November 2, 2012 8:18:51 PM
Subject: Re: [corosync] [Pacemaker] Corosync 2.1.0 dies on both nodes in cluster
On 02/11/12 13:07 -0500, Andrew Martin wrote:
>Hi Angus,
>
>
>Corosync died again while using libqb 0.14.3.
jfrie...@redhat.com >
To: pacemaker@oss.clusterlabs.org, disc...@corosync.org
Sent: Monday, November 5, 2012 2:21:09 AM
Subject: Re: [Pacemaker] [corosync] Corosync 2.1.0 dies on both nodes in
cluster
Angus Salkeld napsal(a):
> On 02/11/12 13:07 -0500, Andrew Martin wrote:
>>
hanks,
Andrew
- Original Message -
From: "Angus Salkeld"
To: pacemaker@oss.clusterlabs.org, disc...@corosync.org
Sent: Thursday, November 1, 2012 5:47:16 PM
Subject: Re: [corosync] [Pacemaker] Corosync 2.1.0 dies on both nodes in cluster
On 01/11/12 17:27 -0500, Andre
corosync] [Pacemaker] Corosync 2.1.0 dies on both nodes in cluster
On 01/11/12 14:32 -0500, Andrew Martin wrote:
>Hi Honza,
>
>
>Thanks for the help. I enabled core dumps in /etc/security/limits.conf but
>didn't have a chance to reboot and apply the changes so I don't have a co
ge -
From: "Jan Friesse"
To: "Andrew Martin"
Cc: disc...@corosync.org, "The Pacemaker cluster resource manager"
Sent: Thursday, November 1, 2012 7:55:52 AM
Subject: Re: [corosync] Corosync 2.1.0 dies on both nodes in cluster
Ansdrew,
I was not able to find any
STONTIHed (must be the localhost interface on storage1). Is
there anyway to prevent this?
Does this help to determine why corosync is dying, and what I can do to fix it?
Thanks,
Andrew
- Original Message -
From: "Andrew Martin"
To: disc...@corosync.org
Sent: Thursday, Novem
Hi Andreas,
- Original Message -
> From: "Andreas Kurz"
> To: pacemaker@oss.clusterlabs.org
> Sent: Wednesday, October 24, 2012 4:13:03 AM
> Subject: Re: [Pacemaker] Behavior of Corosync+Pacemaker with DRBD primary
> power loss
>
> On 10/23/201
lure)? Are split-brains prevented because Pacemaker ensures
that only one node is promoted to Primary at any time?
Is it possible to recover from such a failure without allow-two-primaries?
Thanks,
Andrew
- Original Message -
From: "Andrew Martin"
To: "The Pacemaker
Hello,
I have a 3 node Pacemaker + Corosync cluster with 2 "real" nodes, node0 and
node1, running a DRBD resource (single-primary) and the 3rd node in standby
acting as a quorum node. If node0 were running the DRBD resource, and thus is
DRBD primary, and its power supply fails, will the DRBD r
Hi Mars,
Did you also remove the node from the CIB?
http://www.clusterlabs.org/doc/en-US/Pacemaker/1.0/html/Pacemaker_Explained/s-node-delete.html
Andrew
- Original Message -
From: "Mars gu"
To: "pacemaker"
Sent: Thursday, October 18, 2012 5:47:03 AM
Subject: [Pacemaker] crm_n
"The Pacemaker cluster resource manager"
Sent: Monday, October 15, 2012 5:31:51 AM
Subject: Re: [Pacemaker] STONITHed node cannot rejoin cluster for over 1000
elections
On Sat, Oct 13, 2012 at 1:53 AM, Andrew Martin wrote:
> Hi Andrew,
>
> Thanks, I'll compile Pa
uot;Andrew Beekhof"
To: "The Pacemaker cluster resource manager"
Sent: Thursday, October 11, 2012 8:08:13 PM
Subject: Re: [Pacemaker] STONITHed node cannot rejoin cluster for over 1000
elections
On Fri, Oct 12, 2012 at 7:12 AM, Andrew Martin wrote:
> Hello,
>
>
efore storage1 is brought back into the
cluster. What is the cause of this and how can I modify my cluster
configuration to have nodes rejoin right away?
Thanks,
Andrew Martin
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.cl
unfortunate side-effect.
Is there another way to achieve the same network connectivity check instead of
using ocf:pacemaker:ping? I know the other *ping* resource agents are
deprecated.
Thanks,
Andrew
- Original Message -
From: "Jake Smith"
To: "Andrew Martin"
Cc: &q
"The Pacemaker cluster resource manager"
Sent: Monday, August 27, 2012 9:39:30 AM
Subject: Re: [Pacemaker] Loss of ocf:pacemaker:ping target forces resources to
restart?
- Original Message -----
> From: "Andrew Martin"
> To: "The Pacemaker cluster resource manager&
bs.org
Sent: Thursday, August 23, 2012 3:57:02 AM
Subject: Re: [Pacemaker] Loss of ocf:pacemaker:ping target forces resources to
restart?
Le 22/08/2012 18:23, Andrew Martin a écrit :
> Hello,
>
>
> I have a 3 node Pacemaker + Heartbeat cluster (two real nodes and 1 quorum
> node
2
could ping 192.168.0.128 but node1 could not (move the resources to where
things are better-connected). Is this understanding incorrect? If so, is there
a way I can change my configuration so that it will only restart/migrate
resources if one node is found to be better connected?
Can you tell me why t
i,
>
> On Thu, Jun 28, 2012 at 6:13 PM, Andrew Martin wrote:
>> Hi Dan,
>>
>> Thanks for the help. If I configure the network as I described - ring 0 as
>> the network all 3 nodes are on, ring 1 as the network only 2 of the nodes
>> are on, and using &quo
"
Sent: Wednesday, June 27, 2012 3:42:42 AM
Subject: Re: [Pacemaker] Different Corosync Rings for Different Nodes in Same
Cluster?
Hi,
On Tue, Jun 26, 2012 at 9:53 PM, Andrew Martin wrote:
> Hello,
>
> I am setting up a 3 node cluster with Corosync + Pacemaker on Ubuntu 12.0
Hello,
I am setting up a 3 node cluster with Corosync + Pacemaker on Ubuntu 12.04
server. Two of the nodes are "real" nodes, while the 3rd is in standby mode as
a quorum node. The two "real" nodes each have two NICs, one that is connected
to a shared LAN and the other that is directly connect
l not trigger a failover, but longer will.
Thanks again,
Andrew
- Original Message -
From: "Lars Ellenberg"
To: pacemaker@oss.clusterlabs.org
Sent: Tuesday, June 19, 2012 5:33:46 PM
Subject: Re: [Pacemaker] Why Did Pacemaker Restart this VirtualDomain Resource?
On Tue, Jun 1
ain resource fail to see which operation
resource report the problem
2012/6/19 Andrew Martin < amar...@xes-inc.com >
Hi Emmanuel,
Thanks for the idea. I looked through the rest of the log and these "return
code 8" errors on the ocf:linbit:drbd resources are occurring at other
]: info: rsc:p_drbd_mount1:0 monitor[54]
(pid 12396)
=
it can be a drbd problem, but i tell you the true i'm not sure
==
http://www.clusterlabs.org/doc/en-US/Pacemaker/1.1/html/Pacemaker_Explained/s-ocf-return-codes.html
===
Hello,
I have a 3 node Pacemaker+Heartbeat cluster (two real nodes and one "standby"
quorum node) with Ubuntu 10.04 LTS on the nodes and using the
Pacemaker+Heartbeat packages from the Ubuntu HA Team PPA (
https://launchpad.net/~ubuntu-ha-maintainers/+archive/ppa ). I have configured
3 DRBD
Hi Andreas,
- Original Message -
> From: "Andreas Kurz"
> To: pacemaker@oss.clusterlabs.org
> Sent: Tuesday, April 10, 2012 5:28:15 AM
> Subject: Re: [Pacemaker] Nodes will not promote DRBD resources to
> master on failover
> On 04/10/2012 06:17 AM, Andrew
8:21 AM
Subject: Re: [Pacemaker] VirtualDomain Shutdown Timeout
On Sun, Mar 25, 2012 at 6:27 AM, Andrew Martin wrote:
> Hello,
>
> I have configured a KVM virtual machine primitive using Pacemaker 1.1.6 and
> Heartbeat 3.0.5 on Ubuntu 10.04 Server using DRBD as the storage device (
drbd daemon? It will be started/stopped/managed
automatically by my ocf:linbit:drbd resources (and I can remove the /etc/rc*
symlinks)?
Thanks,
Andrew
- Original Message -
From: "Andreas Kurz" < andr...@hastexo.com >
To: pacemaker@oss.clusterlabs.org
Sent: Wednesday, Ma
Original Message -
From: "Andreas Kurz"
To: pacemaker@oss.clusterlabs.org
Sent: Monday, March 26, 2012 5:56:22 PM
Subject: Re: [Pacemaker] Nodes will not promote DRBD resources to master on
failover
On 03/24/2012 08:15 PM, Andrew Martin wrote:
> Hi Andreas,
>
> My com
Hello,
I have configured a KVM virtual machine primitive using Pacemaker 1.1.6 and
Heartbeat 3.0.5 on Ubuntu 10.04 Server using DRBD as the storage device (so
there is no shared storage, no live-migration):
primitive p_vm ocf:heartbeat:VirtualDomain \
params config="/vmstore/config/vm.xml"
t in /var/log/daemon.log, but I can't find an error message printed
about why it will not promote node2. At this point the DRBD devices are as
follows:
node2: cstate = WFConnection dstate=UpToDate
node1: cstate = StandAlone dstate=Outdated
I don't see any reason why node2 can't becom
21, 2012 9:25:55 PM
Subject: Re: [Pacemaker] Always Run Clone Resource
Shouldnt it have sent you a message that it was being stopped?
On Thu, Mar 22, 2012 at 4:51 AM, Andrew Martin wrote:
> Hello,
>
> I have a pacemaker/heartbeat cluster that uses several DRBD primitives. The
> clu
Hello,
I have a pacemaker/heartbeat cluster that uses several DRBD primitives. The
cluster resources are all colocated and ordered to start after the DRBD
primitives . I have configured an ocf:heartbeat:MailTo primitive and clone for
notifying me of any changes in the cluster state:
primiti
3rd Node as Quorum Node in 2 Node Cluster
It looks like we're waiting for the other node to respond, which it wont do.
Is running pacemaker on the other node but with standby=true an option for you?
On Sat, Feb 25, 2012 at 6:25 AM, Andrew Martin < amar...@xes-inc.com > wrote:
> Hi
day, January 23, 2012 1:53:27 PM
Subject: Re: [Pacemaker] Configuring 3rd Node as Quorum Node in 2 Node Cluster
On 01/23/2012 03:36 PM, Andrew Martin wrote:
> I think I will configure the 3rd (quorum) node in standby mode. In the
> near future I am looking into setting up 2 additional clusters
ntu.com >
To: pacemaker@oss.clusterlabs.org
Sent: Thursday, February 23, 2012 1:48:59 AM
Subject: Re: [Pacemaker] Clone Set does not start its lsb service on all nodes
On 13.02.2012 23:29, Andrew Martin wrote:
> I am creating a cluster that will manage virtual machines using libvirt
> on
rew
- Original Message -
From: "Andreas Ntaflos"
To: "The Pacemaker cluster resource manager" ,
amar...@xes-inc.com
Sent: Wednesday, February 15, 2012 4:09:53 PM
Subject: Re: [Pacemaker] Clone Set does not start its lsb service on all nodes
On 2012-02-13 23:29,
Hello,
I am creating a cluster that will manage virtual machines using libvirt on
Ubuntu 10.04 Server. I am following this guide (except using DRBD instead of
iSCSI):
http://www.linbit.com/fileadmin/tech-guides/ha-kvm.pdf
When configuring the clone set for the libvirt-bin service, I create
anks,
Andrew
- Original Message -
From: "Jake Smith"
To: "The Pacemaker cluster resource manager"
Sent: Friday, February 10, 2012 12:06:15 PM
Subject: Re: [Pacemaker] Help with colocation and order of primitives
- Original Message -
> From: "An
Hello,
I am working on configuring a 2-node cluster with 3 DRBD devices, their
corresponding filesystem mounts, and a VirtualDomain resource agent. I want to
add the appropriate constraints to pacemaker so that these resources only start
together on the same node, and that they start in the p
I think I will configure the 3rd (quorum) node in standby mode. In the near
future I am looking into setting up 2 additional clusters (each of these are
also 2-node clusters) and would like to use this same server as the quorum node
for those clusters as well. Is this possible? If so, how do I h
Here is a guide for setting up a highly-available NFS server using DRBD as the
storage solution (with Pacemaker):
http://www.linbit.com/fileadmin/tech-guides/ha-nfs.pdf
Thanks,
Andrew
- Original Message -
From: "emmanuel segura"
To: "The Pacemaker cluster resource manager"
Se
Hi Dennis,
Have you also added /var/lib/nfs to the shared DRBD resource? This is an
important step to ensure that data about currently-open files and mount
information is transferred to the other node during failover. See the end of
Step 4:
http://www.howtoforge.com/highly-available-nfs-serv
node1
and uname ne node2
Is there anything else that needs to be done to add this 3rd quorum node to the
cluster?
Thanks,
Andrew Martin
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemake
t; > - Original Message -
> > > From: "Rasto Levrinc"
> > > To: "The Pacemaker cluster resource manager"
> > >
> > > Sent: Monday, January 9, 2012 2:12:54 PM
> > > Subject: Re: [Pacemaker] Cannot Create Primitive in CRM S
ter resource manager"
> >
> > Sent: Monday, January 9, 2012 2:12:54 PM
> > Subject: Re: [Pacemaker] Cannot Create Primitive in CRM Shell
> >
> > On Mon, Jan 9, 2012 at 3:34 PM, Andrew Martin
> > wrote:
> > > Hi Florian,
> > >
>
at - 3.0.5
pacemaker - 1.1.6
Thanks,
Andrew
- Original Message -
From: "Andrew Martin"
To: "The Pacemaker cluster resource manager"
Sent: Monday, January 9, 2012 8:34:48 AM
Subject: Re: [Pacemaker] Cannot Create Primitive in CRM Shell
Hi Florian,
Thanks for
ger" < pacemaker@oss.clusterlabs.org >
Sent: Monday, January 9, 2012 5:44:55 AM
Subject: Re: [Pacemaker] Cannot Create Primitive in CRM Shell
On Mon, Jan 9, 2012 at 11:42 AM, Dan Frincu < df.clus...@gmail.com > wrote:
> Hi,
>
> On Fri, Jan 6, 2012 at 11:24 PM, A
Hello,
I am working with DRBD + Heartbeat + Pacemaker to create a 2-node
highly-available cluster. I have been following this official guide on DRBD's
website for configuring all of the components:
http://www.linbit.com/fileadmin/tech-guides/ha-nfs.pdf
However, once I go to configure the p
78 matches
Mail list logo