Re: [Pacemaker] Announce: Pacemaker 1.1.10 now available

2013-07-29 Thread Trevor Hemsley
On 30/07/13 02:51, Andrew Beekhof wrote:
>> [k.proskurin@video-build libqb]$ ./autogen.sh
>> > autoreconf: Entering directory `.'
>> > autoreconf: configure.ac: not using Gettext
>> > autoreconf: running: aclocal -I m4
>> > configure.ac:4: error: Autoconf version 2.61 or higher is required
>> > 
>> > autoconf-2.59-12
>> > 
>> > Dead end.
> Create the srpm for libqb on rhel6, then pass it to rpmbuild on rhel5

Also the IUS Community repo has autoconf26x-2.63-4.ius.centos5.noarch
package that might be useful...

T

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Announce: Pacemaker 1.1.10 now available

2013-07-29 Thread Andrew Beekhof

On 29/07/2013, at 10:37 PM, Kirill Proskurin  wrote:

> On 7/29/13 03:50 , Andrew Beekhof wrote:
>> 
>> On 26/07/2013, at 8:20 PM, Kirill Proskurin  wrote:
>> 
>>> Hi!
>>> 
>>> Is it RHEL5.x friendly?
>> 
>> I'm assuming yes, as no-one complained during the 3 month release candidate 
>> phase ;-)
> 
> Look like it's not.

A good reason to get involved earlier next time ;-)

> 
> [k.proskurin@video-build pacemaker]$ make release
> sort: invalid option -- V
> Try `sort --help' for more information.

try bypassing sort:

   make LAST_RELEASE=Pacemaker-1.1.10 rpm

> 
> And lots of error below.
> Old sort...
> coreutils-5.97-34.el5_8.1
> 
> Ok, let's try my old spec... boom!
> configure: error: Version of libqb is too old: v0.13 or greater requried
> 
> Oops, forgot about this. Ok,
> 
> [k.proskurin@video-build libqb]$ ./autogen.sh
> autoreconf: Entering directory `.'
> autoreconf: configure.ac: not using Gettext
> autoreconf: running: aclocal -I m4
> configure.ac:4: error: Autoconf version 2.61 or higher is required
> 
> autoconf-2.59-12
> 
> Dead end.

Create the srpm for libqb on rhel6, then pass it to rpmbuild on rhel5

> 
>> 
>>> 
>>> On 7/26/13 04:43 , Andrew Beekhof wrote:
 Announcing the release of Pacemaker 1.1.10
 
https://github.com/ClusterLabs/pacemaker/releases/Pacemaker-1.1.10
>>> 
>>> --
>>> Best regards,
>>> Proskurin Kirill
>>> 
>>> ___
>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>> 
>>> Project Home: http://www.clusterlabs.org
>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>> Bugs: http://bugs.clusterlabs.org
>> 
>> 
>> ___
>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>> 
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>> 
>> 
> 
> 
> 
> -- 
> Best regards,
> Proskurin Kirill
> 
> ___
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org


___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Problems after updating from debian squeeze to wheezy

2013-07-29 Thread Oriol Mula-Valls

Hi,

There is a bug already open:
http://bugs.debian.org/cgi-bin/bugreport.cgi?bug=697357

Kind regards,
Oriol

On 29/07/13 21:41, Jelle de Jong wrote:

-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

On 29/07/13 20:58, Arnold Krille wrote:

So further testing show that the bonded interfaces below seem to
eat all broadcasts and all forwards (for the virtual machines).
Don't know why yet. But as soon as I switched my network-config to
use the bridges on eth0 instead of bond0, corosync came back into
the cluster (with the second ring faulty) and virtual machines
where reachable...


Is this with 3.x try using the 2.6 kernel from oldstable, I cant get
the 3.2 kernel to work with bonding bridge and vlans and kvm and
corosync, been having this problem for a while now with all 3.x kernel
and cant seem to fix it, contacted the bonding and bridging mailing
list and no useful help yet.

Please contact me if you find a way to have the same networking
working with both 3.x and 2.6...

Kind regards,

Jelle
-BEGIN PGP SIGNATURE-
Version: GnuPG v1.4.12 (GNU/Linux)

iJwEAQECAAYFAlH2xWIACgkQ1WclBW9j5HlqcAP/QE6JiJfiXkkW7z7bb7sJFslF
kEbIDYlYpqtrMyx/sL7YFkxXpc/D7bXSMqToM6ATE3v+16fBd7xA+OYprCYcp1hJ
95+lapR3jL0Gs5WvT6j8DcgqMCBr6UA4f5dLETdgENnV0EjyLv6lREFK6Qsvp8BE
UeFZwN559bQMPZck4t8=
=oTxj
-END PGP SIGNATURE-

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org




--
Oriol Mula Valls
Institut Català de Ciències del Clima (IC3)
Doctor Trueta 203 - 08005 Barcelona
Tel:+34 93 567 99 77

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Problems after updating from debian squeeze to wheezy

2013-07-29 Thread Jelle de Jong
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

On 29/07/13 20:58, Arnold Krille wrote:
> So further testing show that the bonded interfaces below seem to
> eat all broadcasts and all forwards (for the virtual machines).
> Don't know why yet. But as soon as I switched my network-config to
> use the bridges on eth0 instead of bond0, corosync came back into
> the cluster (with the second ring faulty) and virtual machines
> where reachable...

Is this with 3.x try using the 2.6 kernel from oldstable, I cant get
the 3.2 kernel to work with bonding bridge and vlans and kvm and
corosync, been having this problem for a while now with all 3.x kernel
and cant seem to fix it, contacted the bonding and bridging mailing
list and no useful help yet.

Please contact me if you find a way to have the same networking
working with both 3.x and 2.6...

Kind regards,

Jelle
-BEGIN PGP SIGNATURE-
Version: GnuPG v1.4.12 (GNU/Linux)

iJwEAQECAAYFAlH2xWIACgkQ1WclBW9j5HlqcAP/QE6JiJfiXkkW7z7bb7sJFslF
kEbIDYlYpqtrMyx/sL7YFkxXpc/D7bXSMqToM6ATE3v+16fBd7xA+OYprCYcp1hJ
95+lapR3jL0Gs5WvT6j8DcgqMCBr6UA4f5dLETdgENnV0EjyLv6lREFK6Qsvp8BE
UeFZwN559bQMPZck4t8=
=oTxj
-END PGP SIGNATURE-

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Problems after updating from debian squeeze to wheezy

2013-07-29 Thread Arnold Krille

Am 2013-07-29 18:46, schrieb Arnold Krille:

Hi all,

I have a little problem here and would like to get some help:

I have (had?) a working three-node cluster of two active nodes
(nebel1 and nebel2) and one standby-node (nebel3) running debian
squeeze + backports. That is pacemaker 1.1.7-1~bpo60+1 and corosync
1.4.2-1~bpo60+1.

Now I updated the standby-node nebel3 to debian wheezy which went
without problems itself. But as neither the versions of pacemaker and
corosync changed, I expected the updated nebel3 to join the original
cluster again. Little did I know... So while nebel3 has pacemaker
1.1.7-1 and corosync 1.4.2-3, it seems something in the update broke
it. /etc/corosync/corosync.conf is still the same on all nodes.



So further testing show that the bonded interfaces below seem to eat 
all broadcasts and all forwards (for the virtual machines). Don't know 
why yet. But as soon as I switched my network-config to use the bridges 
on eth0 instead of bond0, corosync came back into the cluster (with the 
second ring faulty) and virtual machines where reachable...


Have fun,

Arnold

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] libqb installed in non-standard dircausesconfigurefailures

2013-07-29 Thread Matthew O'Connor

On 07/28/2013 08:12 PM, Andrew Beekhof wrote:
> On 28/07/2013, at 2:12 PM, Matthew O'Connor  wrote:
>
>> On 07/23/2013 07:04 PM, Andrew Beekhof wrote:
>>> On 23/07/2013, at 12:32 AM, Matthew O'Connor  wrote:
>>>
 Hi Andrew,

 On 07/19/2013 12:22 AM, Andrew Beekhof wrote:
>> I've added the PKG_CONFIG_PATH and the two libqb_ lines in an attempt to
>>> make things work, as recommended by the configure help.  So far, no
>>> dice.  Is this something that needs to be fixed in the autoconf/autogen
>>> stuff?  Something I can submit a patch for?  (sadly, not versed at all
>>> in autoconf/autogen, but willing to learn!)
> we get libqb's details from pnk-config, so depending on what how it 
> processes PKG_CONFIG_PATH, maybe its returning the "main" installation 
> instead of the one you put in /opt/cluster
>
 The only libqb on my release of Ubuntu (12.04) is the one I built and
 installed myself - no released packages available last I checked. 
 Managed to get ./configure working by putting the library and include
 paths into LDFLAGS and CFLAGS (respectively), and putting the env vars
 AFTER ./configure instead of before.  Maybe PKG_CONFIG_PATH would work
 better there as well.  Unfortunately it's still not a totally clean
 install on Ubuntu as the cluster-glue stuff dictates where the ocf files
 go.  Will either have to build cluster-glue from scratch or take what I
 have as a compromise and go.
>>> What version were you trying to build btw?
>>> Because configure has changed a bit in that area. Maybe HEAD will behave 
>>> better for you...
>> 1.1.9.  I did eventually find that supplying PKG_CONFIG_PATH as an
>> _argument_ to configure instead of as an environment var made the
>> necessary difference.  I thereafter succeeded in building a full suite
>> to a non-standard folder.  Sadly I couldn't quite get around not using
>> sudo for most of the "make install"s...changing ownership as an
>> underprivileged user is obviously a no-no.
> Glad you made progress in the end :)
>
>> One issue I came across: when configuring the resource-agents package
>> for install, I had to use a peculiar configure line to get the ocf
>> scripts in the right place:
>>
>>   ./configure --prefix=/opt/cluster --with-ocf-root=lib/ocf
>>
>> I came to that after digging through how configure was using the
>> variables related to the OCF root.  This seems to be dissimilar to
>> cluster-glue's usage of --with-ocf-root (which was
>> --with-ocf-root=/opt/cluster/lib/ocf).
> Passing along to David who looks after resource-agents for RH.
> However, once they're no longer in /usr/lib/ocf, they're technically not 
> compliant with the OCF standard anymore ;-)

Hmm good point.

>> One other question: which corosync version do you recommend to use with
>> 1.1.9+?  I'm on 1.4.6 right now.
> Pacemaker can function with anything since 1.2 iirc.

Great!  Thanks again!!

>
>> Thanks!
>>
>> -- Matthew
>> P.S. I'm looking forward to trying 1.1.10!
>>
 Thanks!

 -- 
 Thank you!
 Matthew O'Connor
 (GPG Key ID: 55F981C4)


 CONFIDENTIAL NOTICE: The information contained in this electronic message 
 is legally privileged, confidential and exempt from disclosure under 
 applicable law. It is intended only for the use of the individual or 
 entity named above. If the reader of this message is not the intended 
 recipient, you are hereby notified that any dissemination, distribution or 
 copying of this message is strictly prohibited. If you have received this 
 communication in error, please notify the sender immediately by return 
 e-mail and delete the original message and any copies of it from your 
 computer system. Thank you.

 EXPORT CONTROL WARNING:  This document may contain technical data that is 
 subject to the International Traffic in Arms Regulations (ITAR) controls 
 and may not be exported or otherwise disclosed to any foreign person or 
 firm, whether in the US or abroad, without first complying with all 
 requirements of the ITAR, 22 CFR 120-130, including the requirement for 
 obtaining an export license if applicable. In addition, this document may 
 contain technology that is subject to the Export Administration 
 Regulations (EAR) and may not be exported or otherwise disclosed to any 
 non-U.S. person, whether in the US or abroad, without first complying with 
 all requirements of the EAR, 15 CFR 730-774, including the requirement for 
 obtaining an export license if applicable. Violation of these export laws 
 is subject to severe criminal penalties.


 ___
 Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
 http://oss.clusterlabs.org/mailman/listinfo/pacemaker

 Project Home: http://www.clusterlabs.org
 Getting started: http://www.clusterlabs.org/doc/Cluster_from

[Pacemaker] Problems after updating from debian squeeze to wheezy

2013-07-29 Thread Arnold Krille

Hi all,

I have a little problem here and would like to get some help:

I have (had?) a working three-node cluster of two active nodes (nebel1 
and nebel2) and one standby-node (nebel3) running debian squeeze + 
backports. That is pacemaker 1.1.7-1~bpo60+1 and corosync 
1.4.2-1~bpo60+1.


Now I updated the standby-node nebel3 to debian wheezy which went 
without problems itself. But as neither the versions of pacemaker and 
corosync changed, I expected the updated nebel3 to join the original 
cluster again. Little did I know... So while nebel3 has pacemaker 
1.1.7-1 and corosync 1.4.2-3, it seems something in the update broke it. 
/etc/corosync/corosync.conf is still the same on all nodes.


I suspect the problem is somewhere in corosync as nebel1 and nebel2 
only see each other:


$ ssh root@nebel2 --  corosync-objctl |grep member
runtime.totem.pg.mrp.srp.members.33648138.ip=r(0) ip(10.110.1.2) r(1) 
ip(10.112.0.2)

runtime.totem.pg.mrp.srp.members.33648138.join_count=1
runtime.totem.pg.mrp.srp.members.33648138.status=joined
runtime.totem.pg.mrp.srp.members.16870922.ip=r(0) ip(10.110.1.1) r(1) 
ip(10.112.0.1)

runtime.totem.pg.mrp.srp.members.16870922.join_count=1
runtime.totem.pg.mrp.srp.members.16870922.status=joined
runtime.totem.pg.mrp.srp.members.50425354.ip=r(0) ip(10.110.1.3) r(1) 
ip(10.112.0.3)

runtime.totem.pg.mrp.srp.members.50425354.join_count=39
runtime.totem.pg.mrp.srp.members.50425354.status=left

nebel3 on the other hand:

$ ssh root@nebel3 --  corosync-objctl |grep member
runtime.totem.pg.mrp.srp.members.50425354.ip=r(0) ip(10.110.1.3) r(1) 
ip(10.112.0.3)

runtime.totem.pg.mrp.srp.members.50425354.join_count=1
runtime.totem.pg.mrp.srp.members.50425354.status=joined

Both nebel2 and nebel3 think the communication-rings are free of 
faults:


$ ssh root@nebel2 --  corosync-cfgtool -s
Printing ring status.
Local node ID 33648138
RING ID 0
id  = 10.110.1.2
status  = ring 0 active with no faults
RING ID 1
id  = 10.112.0.2
status  = ring 1 active with no faults
$ ssh root@nebel3 --  corosync-cfgtool -s
Printing ring status.
Local node ID 50425354
RING ID 0
id  = 10.110.1.3
status  = ring 0 active with no faults
RING ID 1
id  = 10.112.0.3
status  = ring 1 active with no faults

I can ping all the participating nodes via all their connections and 
IPs from all nodes


The corosync.log on nebel2 doesn't mention nebel3 after it leaving the 
cluster for reboot after the update. Likewise the corosync.log on nebel3 
doesn't mention nebel2 and nebel1 anymore.


So, what did I miss during the update? How can I get nebel3 to join 
back into the original cluster instead of forming its own 1-out-of-3 
cluster (with the same resources defined)?


Any helps is highly appreciated!

- Arnold

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Announce: Pacemaker 1.1.10 now available

2013-07-29 Thread Kirill Proskurin

On 7/29/13 03:50 , Andrew Beekhof wrote:


On 26/07/2013, at 8:20 PM, Kirill Proskurin  wrote:


Hi!

Is it RHEL5.x friendly?


I'm assuming yes, as no-one complained during the 3 month release candidate 
phase ;-)


Look like it's not.

[k.proskurin@video-build pacemaker]$ make release
sort: invalid option -- V
Try `sort --help' for more information.

And lots of error below.
Old sort...
coreutils-5.97-34.el5_8.1

Ok, let's try my old spec... boom!
configure: error: Version of libqb is too old: v0.13 or greater requried

Oops, forgot about this. Ok,

[k.proskurin@video-build libqb]$ ./autogen.sh
autoreconf: Entering directory `.'
autoreconf: configure.ac: not using Gettext
autoreconf: running: aclocal -I m4
configure.ac:4: error: Autoconf version 2.61 or higher is required

autoconf-2.59-12

Dead end.





On 7/26/13 04:43 , Andrew Beekhof wrote:

Announcing the release of Pacemaker 1.1.10

https://github.com/ClusterLabs/pacemaker/releases/Pacemaker-1.1.10


--
Best regards,
Proskurin Kirill

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org



___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org






--
Best regards,
Proskurin Kirill

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster

2013-07-29 Thread emmanuel segura
Hello

That would say you have problem with your multicast :)




2013/7/29 Enric Muñoz 

>  Using unicast it is working well. Thank you very much.
>
> ** **
>
> *De:* Michael Schwartzkopff [mailto:mi...@clusterbau.com]
> *Enviado el:* lunes, 29 de julio de 2013 13:17
> *Para:* The Pacemaker cluster resource manager
> *Asunto:* Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node
> cluster
>
> ** **
>
> Am Montag, 29. Juli 2013, 10:49:46 schrieb Enric Muñoz:
>
> > I only see the multicast packets in the node that is online but not in
> the
>
> > other one. How can I configure corosync to use unicast??
>
>  
>
> Googlesays:
>
>
> http://www.google.de/url?sa=t&rct=j&q=corosync%20udpu&source=web&cd=1&cad=rja&ved=0CDQQFjAA&url=https%3A%2F%2Fgithub.com%2Ffghaas%2Fcorosync%2Fblob%2Fmaster%2Fconf%2Fcorosync.conf.example.udpu&ei=nE72Uca2BYfVswbKpYGwDw&usg=AFQjCNFxgQSKNt6qDqZG47dWbnEMAa9HgQ
> 
>
>  
>
>  
>
> -- 
>
> Dr. Michael Schwartzkopff
>
> Guardinistr. 63
>
> 81375 München
>
>  
>
> Tel: (0163) 172 50 98
>
> ___
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
>


-- 
esta es mi vida e me la vivo hasta que dios quiera
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster

2013-07-29 Thread Enric Muñoz
Using unicast it is working well. Thank you very much.

De: Michael Schwartzkopff [mailto:mi...@clusterbau.com]
Enviado el: lunes, 29 de julio de 2013 13:17
Para: The Pacemaker cluster resource manager
Asunto: Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster


Am Montag, 29. Juli 2013, 10:49:46 schrieb Enric Muñoz:

> I only see the multicast packets in the node that is online but not in the

> other one. How can I configure corosync to use unicast??



Googlesays:

http://www.google.de/url?sa=t&rct=j&q=corosync%20udpu&source=web&cd=1&cad=rja&ved=0CDQQFjAA&url=https%3A%2F%2Fgithub.com%2Ffghaas%2Fcorosync%2Fblob%2Fmaster%2Fconf%2Fcorosync.conf.example.udpu&ei=nE72Uca2BYfVswbKpYGwDw&usg=AFQjCNFxgQSKNt6qDqZG47dWbnEMAa9HgQ





--

Dr. Michael Schwartzkopff

Guardinistr. 63

81375 München



Tel: (0163) 172 50 98
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster

2013-07-29 Thread Michael Schwartzkopff
Am Montag, 29. Juli 2013, 10:49:46 schrieb Enric Muñoz:
> I only see the multicast packets in the node that is online but not in the
> other one. How can I configure corosync to use unicast??

Googlesays:
http://www.google.de/url?sa=t&rct=j&q=corosync%20udpu&source=web&cd=1&cad=rja&ved=0CDQQFjAA&url=https%3A%2F%2Fgithub.com%2Ffghaas%2Fcorosync%2Fblob%2Fmaster%2Fconf%2Fcorosync.conf.example.udpu&ei=nE72Uca2BYfVswbKpYGwDw&usg=AFQjCNFxgQSKNt6qDqZG47dWbnEMAa9HgQ


-- 
Dr. Michael Schwartzkopff
Guardinistr. 63
81375 München

Tel: (0163) 172 50 98___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster

2013-07-29 Thread Enric Muñoz
I only see the multicast packets in the node that is online but not in the 
other one. How can I configure corosync to use unicast??

After enabling IGMP in the switch it appears a new message in the LCMC:

Could not initialize corosync configuration API error 6
Crm verify: Could not establish cib_ro connection: connection refused (111)
Live CIB query failed: Transport endpoint is not Connected.
Connection to cluster failed: transport endpoint is not Connected.

Thanks

De: Michael Schwartzkopff [mailto:mi...@clusterbau.com]
Enviado el: lunes, 29 de julio de 2013 12:31
Para: The Pacemaker cluster resource manager
Asunto: Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster


Am Montag, 29. Juli 2013, 10:06:29 schrieb Enric Muñoz:

> Iptables is disabled and selinux set to permissive in both nodes.



Do a tcpdump on both nodes. You should see the multicast packets as well as 
some unicast packets to the udp port you defined. As fasr as I remember there 
are three (or four?) unicast packets between any multicast packet.

If you do not see these packets, you have several options to search:



- do you use virtual machines? I have seen kvm problems with multicast packets.

- separate subnets? Talk to your network guys to use IGMP.

- Problems with your switches? Talk to the network guys.



On the other hand, you also could use udpu (unicast) for the transport.



--

Dr. Michael Schwartzkopff

Guardinistr. 63

81375 München



Tel: (0163) 172 50 98
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster

2013-07-29 Thread Michael Schwartzkopff
Am Montag, 29. Juli 2013, 10:06:29 schrieb Enric Muñoz:
> Iptables is disabled and selinux set to permissive in both nodes.

Do a tcpdump on both nodes. You should see the multicast packets as well as 
some unicast packets to the udp port you defined. As fasr as I remember there 
are three (or four?) unicast packets between any multicast packet.
If you do not see these packets, you have several options to search:

- do you use virtual machines? I have seen kvm problems with multicast 
packets.
- separate subnets? Talk to your network guys to use IGMP.
- Problems with your switches? Talk to the network guys.

On the other hand, you also could use udpu (unicast) for the transport.

-- 
Dr. Michael Schwartzkopff
Guardinistr. 63
81375 München

Tel: (0163) 172 50 98___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster

2013-07-29 Thread emmanuel segura
Hello

If you are using multicast check your igmp switch support is enabled.

Thanks


2013/7/29 Enric Muñoz 

>  Iptables is disabled and selinux set to permissive in both nodes.
>
> ** **
>
> *De:* Michael Schwartzkopff [mailto:mi...@clusterbau.com]
> *Enviado el:* lunes, 29 de julio de 2013 11:55
>
> *Para:* The Pacemaker cluster resource manager
> *Asunto:* Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node
> cluster
>
>  ** **
>
> Am Montag, 29. Juli 2013, 09:32:53 schrieb Enric Muñoz:
>
> > Hi,
>
> > 
>
> > It seems that the nodes don't see each other:
>
> > 
>
> > # corosync-objctl | grep members
>
> > runtime.totem.pg.mrp.srp.members.1275439296.ip=r(0) ip(192.168.5.204)***
> *
>
> > runtime.totem.pg.mrp.srp.members.1275439296.join_count=1
>
> > runtime.totem.pg.mrp.srp.members.1275439296.status=joined
>
> > 
>
> > I can ping from one node to the other succesfully. How can I join them??
> 
>
>  
>
> Plese check your firewall rules.
>
>  
>
> -- 
>
> Dr. Michael Schwartzkopff
>
> Guardinistr. 63
>
> 81375 München
>
>  
>
> Tel: (0163) 172 50 98
>
> ___
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
>


-- 
esta es mi vida e me la vivo hasta que dios quiera
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster

2013-07-29 Thread Enric Muñoz
Iptables is disabled and selinux set to permissive in both nodes.

De: Michael Schwartzkopff [mailto:mi...@clusterbau.com]
Enviado el: lunes, 29 de julio de 2013 11:55
Para: The Pacemaker cluster resource manager
Asunto: Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster


Am Montag, 29. Juli 2013, 09:32:53 schrieb Enric Muñoz:

> Hi,

>

> It seems that the nodes don't see each other:

>

> # corosync-objctl | grep members

> runtime.totem.pg.mrp.srp.members.1275439296.ip=r(0) ip(192.168.5.204)

> runtime.totem.pg.mrp.srp.members.1275439296.join_count=1

> runtime.totem.pg.mrp.srp.members.1275439296.status=joined

>

> I can ping from one node to the other succesfully. How can I join them??



Plese check your firewall rules.



--

Dr. Michael Schwartzkopff

Guardinistr. 63

81375 München



Tel: (0163) 172 50 98
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster

2013-07-29 Thread Michael Schwartzkopff
Am Montag, 29. Juli 2013, 09:32:53 schrieb Enric Muñoz:
> Hi,
> 
> It seems that the nodes don't see each other:
> 
> # corosync-objctl | grep members
> runtime.totem.pg.mrp.srp.members.1275439296.ip=r(0) ip(192.168.5.204)
> runtime.totem.pg.mrp.srp.members.1275439296.join_count=1
> runtime.totem.pg.mrp.srp.members.1275439296.status=joined
> 
> I can ping from one node to the other succesfully. How can I join them??

Plese check your firewall rules.

-- 
Dr. Michael Schwartzkopff
Guardinistr. 63
81375 München

Tel: (0163) 172 50 98___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster

2013-07-29 Thread Enric Muñoz
Hi,

It seems that the nodes don't see each other:

# corosync-objctl | grep members
runtime.totem.pg.mrp.srp.members.1275439296.ip=r(0) ip(192.168.5.204)
runtime.totem.pg.mrp.srp.members.1275439296.join_count=1
runtime.totem.pg.mrp.srp.members.1275439296.status=joined

I can ping from one node to the other succesfully. How can I join them??

About the STONITH, I disabled it in the LCMC but it seems that it applied the 
changes only to the first node. I don't have any command crm available ( i read 
it's something about this version of pacemaker), so how can I disable it??

Sorry, I'm totally new on pacemaker/corosync.

Thanks

De: Michael Schwartzkopff [mailto:mi...@clusterbau.com]
Enviado el: lunes, 29 de julio de 2013 11:15
Para: The Pacemaker cluster resource manager
Asunto: Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster


Am Montag, 29. Juli 2013, 09:01:01 schrieb Enric Muñoz:

> Hi all,

>

> I am trying to build a 2 node iSCSI HA storage cluster with Pacemaker,

> Corosync and DRBD on CentOS 6.4. I have problems while Building the

> cluster. The problem is that pacemaker is not starting in the second node.

> When doing corosync-cfgtool -s both nodes are active with no faults



Do the nodes see each other? Please check with



# corosync-objctl



In the section "members" at the end of the output there should be the names of 
both nodes and status should be "joined"



the other error message is self-explaining: The cluster does STONITH by 
default. Since you have no STONITH resources defined, it refuses to start 
resources. Please configure STONITH resources or disable the STONITH feature 
for testing by:



# crm configure property stonith-enabled="false"



But DO configure STONITH/fencing for a productive cluster that uses shared 
storage!



--

Dr. Michael Schwartzkopff

Guardinistr. 63

81375 München



Tel: (0163) 172 50 98
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster

2013-07-29 Thread emmanuel segura
Hello

You need configure stonith for your cluster.

Thanks


2013/7/29 Enric Muñoz 

>  Hi all,
>
> ** **
>
> I am trying to build a 2 node iSCSI HA storage cluster with Pacemaker,
> Corosync and DRBD on CentOS 6.4. I have problems while Building the
> cluster. The problem is that pacemaker is not starting in the second node.
>  When doing corosync-cfgtool  -s both nodes are active with no faults and
> the only error I see is the following:
>
> ** **
>
> # crm_verify -VL
>
>error: unpack_resources: Resource start-up disabled since no
> STONITH resources have been defined
>
>error: unpack_resources: Either configure some or disable STONITH
> with the stonith-enabled option
>
>error: unpack_resources: NOTE: Clusters with shared data need
> STONITH to ensure data integrity
>
> Errors found during check: config not valid
>
>   -V may provide more details
>
> ** **
>
> This is only for the 2nd node. In the first node it says config OK.
>
> ** **
>
> Can anyone help me??
>
> ** **
>
> Thanks in advance
>
> ** **
>
> ** **
>
> ___
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
>


-- 
esta es mi vida e me la vivo hasta que dios quiera
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster

2013-07-29 Thread Michael Schwartzkopff
Am Montag, 29. Juli 2013, 09:01:01 schrieb Enric Muñoz:
> Hi all,
> 
> I am trying to build a 2 node iSCSI HA storage cluster with Pacemaker,
> Corosync and DRBD on CentOS 6.4. I have problems while Building the
> cluster. The problem is that pacemaker is not starting in the second node. 
> When doing corosync-cfgtool  -s both nodes are active with no faults

Do the nodes see each other? Please check with

# corosync-objctl

In the section "members" at the end of the output there should be the names of 
both nodes and status should be "joined"

the other error message is self-explaining: The cluster does STONITH by 
default. Since you have no STONITH resources defined, it refuses to start 
resources. Please configure STONITH resources or disable the STONITH feature 
for testing by:

# crm configure property stonith-enabled="false"

But DO configure STONITH/fencing for a productive cluster that uses shared 
storage!

-- 
Dr. Michael Schwartzkopff
Guardinistr. 63
81375 München

Tel: (0163) 172 50 98___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[Pacemaker] Pacemaker not starting in 2nd node of a 2 node cluster

2013-07-29 Thread Enric Muñoz
Hi all,

I am trying to build a 2 node iSCSI HA storage cluster with Pacemaker, Corosync 
and DRBD on CentOS 6.4. I have problems while Building the cluster. The problem 
is that pacemaker is not starting in the second node.  When doing 
corosync-cfgtool  -s both nodes are active with no faults and the only error I 
see is the following:

# crm_verify -VL
   error: unpack_resources: Resource start-up disabled since no STONITH 
resources have been defined
   error: unpack_resources: Either configure some or disable STONITH with 
the stonith-enabled option
   error: unpack_resources: NOTE: Clusters with shared data need STONITH to 
ensure data integrity
Errors found during check: config not valid
  -V may provide more details

This is only for the 2nd node. In the first node it says config OK.

Can anyone help me??

Thanks in advance


___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org