I was successfully running 1.1.8 on a pair of CentOS 6.4 servers and
after updating to CentOS 6.5 and 1.1.10, pacemaker miss-behaves.
The first symptoms appeared with the 1.1.10-14.el6 packages. About 20
hours after the upgrade, the first drbd_monitor issues came out.
Dec 09 18:50:12 Updated:
On 05/13/2013 10:03 AM, Diego Remolina wrote:
Hi,
I was wondering if anybody can tell me what is the best way to replicate
the following crm commands in pcs. It seems pcs cannot do any
collocation rules using groups (or I just do not know how to properly do
it). I was also not able to figure
Hi,
I was wondering if anybody can tell me what is the best way to replicate
the following crm commands in pcs. It seems pcs cannot do any
collocation rules using groups (or I just do not know how to properly do
it). I was also not able to figure out how to colocate a drbd resource
on the con
Why are you trying to compile from source, use the clusterlabs repo...
http://www.clusterlabs.org/wiki/Install#Installing_on_EPEL_Compatible_Distributions:_RHEL.2C_CentOS.2C_etc
;)
Diego
shejimshad M wrote:
Dear sir
I am using redhat 5. We are trying to implement HIGH
AVAILABIL
Amazing what impact 5 missing characters can have.
http://hg.clusterlabs.org/pacemaker/stable-1.0/rev/89bd754939df
1.0.9.1 is already up on clusterlabs.org
Humble apologies for the screw up.
Much better!!!
[r...@phys-ha02 ~]# rpm -qa | grep coro
corosynclib-1.2.5-1.3.el5
corosync-1.2.5-1.3.
Hi,
I have just updated my corosync and pacemaker packages and now whenever
I try to start corosync, I get segfault messages in the logs.
Here is what I did.
crm node offline phys-ha02.physics.gatech.edu
/etc/init.d/corosync stop
yum -y update (updated corosync and pacemaker to 1.2.5-1.3 an
unknown. I have tried RHEL 5.5 as well as CentOS 5.5 with clusterrepo
rpms and been unable to reproduce. I'll keep looking.
Regards
-steve
On 05/27/2010 06:07 AM, Diego Remolina wrote:
Hi,
I was running the old rpms from the opensuse repo and wanted to change
over to the latest packages
unknown. I have tried RHEL 5.5 as well as CentOS 5.5 with clusterrepo
rpms and been unable to reproduce. I'll keep looking.
Regards
-steve
On 05/27/2010 06:07 AM, Diego Remolina wrote:
Hi,
I was running the old rpms from the opensuse repo and wanted to change
over to the latest pac
Hi,
I was running the old rpms from the opensuse repo and wanted to change
over to the latest packages from the clusterlabs repo in my RHEL 5.5
machines.
Steps I took
1. Disabled the old repo
2. Set the nodes to standby (two node drbd cluster) and turned of openais
3. Enabled the new repo.
4.
One thing that I believe would really complement the howto is if you
could explain the configuration for a more complex setup, say two
separate drbd Primary devices on the same cluster.
Going from one to two drbd devices is not always obvious and there are
some location, collocation and order
ngine(hacluster to daemon) in 1.0.5, but not exactly sure.
hj
On Fri, Aug 21, 2009 at 9:11 AM, Diego Remolina
<mailto:diego.remol...@physics.gatech.edu>> wrote:
Here is what I am seeing now right after stopping openais, updating
heartbeat and pacemaker and trying to start openais a
samba (lsb:smb): Started phys-file01.physics.gatech.edu
Clone Set: pingd-clone
Started: [ phys-file01.physics.gatech.edu
phys-file02.physics.gatech.edu ]
[r...@phys-file02 ~]#
Diego
Andrew Beekhof wrote:
On Wed, Aug 12, 2009 at 3:35 PM, Diego
Remolina wrote:
could you instead at
I agree with Gerry, why would you disallow automatic failover, that is
the complete opposite idea to having a cluster.
In any case, I think you could put in some rule constraint tying the
resources to run in only one node with a value of infinity. If you
wanted to move the resource, you would
I have found a bug in the script .ocf-shellfunctions part of
resource-agents-1.0.27-3 which replaces heartbeat-resources.
There are 5 locations where the command "echo" is capitalized, this
produces errors such as:
lrmd: [29090]: info: RA output: (drbd_export:0:monitor:stderr)
/usr/lib/ocf/r
Do you have an entry for:
no-quorum-policy="ignore"
Here is how mine looks for a 2 node master/slave drbd fileserver cluster:
property $id="cib-bootstrap-options" \
dc-version="1.0.4-6dede86d6105786af3a5321ccf66b44b6914f0aa" \
cluster-infrastructure="openais" \
expected-
I noticed that you are using a non-cluster file system, ext3, so you should be using a master slave resource, not a simple resource for drbd. (unless you seem to be starting drbd with the system init scripts that may not be the best thing to do).Please look at my previous post to the list "Master/S
could you instead attach to it with gdb and see what it was doing?
I will try, but cannot promise it will be soon, beginning of the
semester is very busy and I am not familiar with gdb...
that looks suspicious... are you invoking the shell or crm_shadow?
This is probably when I type crm st
Aug 12 07:57:17 phys-file02 openais[9380]: [crm ] info: process_ais_conf:
Reading configure
Aug 12 07:57:17 phys-file02 openais[9380]: [MAIN ] info: config_find_next:
Processing additional logging options...
Aug 12 07:57:17 phys-file02 openais[9380]: [MAIN ] info: get_config_opt:
Found 'on' for o
Can you define "not correctly" please?
I'd rather not ignore such behavior.
The machine would come up and not join the cluster. Checking the status
of openais would show as "Running". crm status would show:
Connection to cluster failed: connection failed
A look at the log file shows:
Aug 12
Solution:
1) clone the pingd
2) Delete you colocation constraint. It is useless.
3) Make a location constatint the allows the ip address only run on a node
that gets points from the pingd.
I want to thank Michael for pointing out my mistake. I have also
migrated away from using heartbeat to op
What is the failconter of the resource on node phys-file02? Please do a
crm_mon -1f
What does it say?
[r...@phys-file02 ~]# crm_mon -1f
Last updated: Fri Aug 7 08:29:20 2009
Stack: Heartbeat
Current DC: phys-file02.physics.gatech.edu
(db786ace-4c9b-4ba1-b272-95b4d81b40a9) - par
Hi,
I am fairly new to pacemaker, and while I had things working correctly
for a while, in testing failovers and playing with my machines I got
them to a state where one resource cannot start (ms-drbd_export:1).
Last updated: Fri Aug 7 07:27:52 2009
Stack: Heartbeat
Current DC:
22 matches
Mail list logo