Re: [Pacemaker] Missing autoconf check for sensors.h?
On Wed, Oct 21, 2009 at 8:53 AM, Florian Haas florian.h...@linbit.com wrote: OK, so building on a system where net-snmp-devel is not installed would circumvent this problem? yep, or if you installed the sensors devel package ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Re: [Pacemaker] Missing autoconf check for sensors.h?
On 10/21/2009 09:01 AM, Andrew Beekhof wrote: On Wed, Oct 21, 2009 at 8:53 AM, Florian Haas florian.h...@linbit.com wrote: OK, so building on a system where net-snmp-devel is not installed would circumvent this problem? yep, or if you installed the sensors devel package yeah, that I realized by myself. :) Cheers, Florian signature.asc Description: OpenPGP digital signature ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker
[Pacemaker] Why are fatal warnings enabled by default?
Andrew, Dejan, For pacemaker and agents, configure defaults to --enable-fatal-warnings. AFAIR, neither of these have ever built successfully with fatal warnings enabled. Is there a specific reason to keep the default as it is? Is this perhaps a deliberate entry barrier for packagers, so as to make sure they know what they're doing? :) Or would it be possible to actually fix those warnings, so the packages would compile -- *gasp* -- even with the configure defaults? Cheers, Florian signature.asc Description: OpenPGP digital signature ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Re: [Pacemaker] Dual powersupply stonith
On Wed, 2009-10-21 at 01:16 +0200, Dejan Muhamedagic wrote: In that case, we would need to send two synchronised stonith actions to two different stonith devices to succesfully stonith a node. Would this be best solved at pacemaker or at RA level? (Assuming this isn't solved at hardware level offcourse). That's not possible at all right now. Unless you go with another fencing solution such as sbd, but that's a different matter. Hmm. If a single RA would (transparantly) control two stonith devices, it should work? Error handling would require some thinking though. Unless there are some things I am missing, it should be possible with two of those IP Power things. Not a generic solution you want to use in the project though. J. ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Re: [Pacemaker] Why are fatal warnings enabled by default?
On Wed, Oct 21, 2009 at 9:31 AM, Florian Haas florian.h...@linbit.com wrote: Andrew, Dejan, For pacemaker and agents, configure defaults to --enable-fatal-warnings. So that the warnings get fixed. AFAIR, neither of these have ever built successfully with fatal warnings enabled. Only on Debian. shrug. Everywhere else it builds just fine. Is there a specific reason to keep the default as it is? Is this perhaps a deliberate entry barrier for packagers, so as to make sure they know what they're doing? :) Or would it be possible to actually fix those warnings, so the packages would compile -- *gasp* -- even with the configure defaults? Cheers, Florian ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Re: [Pacemaker] Dual powersupply stonith
Hi, On Wed, Oct 21, 2009 at 09:34:06AM +0200, Johan Verrept wrote: On Wed, 2009-10-21 at 01:16 +0200, Dejan Muhamedagic wrote: In that case, we would need to send two synchronised stonith actions to two different stonith devices to succesfully stonith a node. Would this be best solved at pacemaker or at RA level? (Assuming this isn't solved at hardware level offcourse). That's not possible at all right now. Unless you go with another fencing solution such as sbd, but that's a different matter. Hmm. If a single RA would (transparantly) control two stonith devices, it should work? Yes. Error handling would require some thinking though. Unless there are some things I am missing, it should be possible with two of those IP Power things. Not a generic solution you want to use in the project though. No. In this case, since the power distribution is rather robust, it would be great to have am on-board management card such as HP iLO or IBM RSA and then use that for fencing. Thanks, Dejan J. ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Re: [Pacemaker] Why are fatal warnings enabled by default?
Hi, On Wed, Oct 21, 2009 at 01:36:59PM +0200, Florian Haas wrote: On 10/21/2009 11:34 AM, Andrew Beekhof wrote: On Wed, Oct 21, 2009 at 11:04 AM, Florian Haas florian.h...@linbit.com wrote: On 2009-10-21 10:49, Andrew Beekhof wrote: On Wed, Oct 21, 2009 at 9:31 AM, Florian Haas florian.h...@linbit.com wrote: Andrew, Dejan, For pacemaker and agents, configure defaults to --enable-fatal-warnings. So that the warnings get fixed. AFAIR, neither of these have ever built successfully with fatal warnings enabled. Only on Debian. shrug. Everywhere else it builds just fine. Unless everywhere else excludes CentOS 5, where it doesn't build either. At least for me it doesn't. Don't know about upstream RHEL. The warnings being? In agents, a simple ./configure make leads to: [...] gmake[1]: Entering directory `/home/rpmbuild/hg/cluster-agents/heartbeat' if gcc -DHAVE_CONFIG_H -I. -I. -I../include -I../include -I../include -I../linux-ha -I/usr/include/glib-2.0 -I/usr/lib/glib-2.0/include-g -O2 -ggdb3 -O0 -fgnu89-inline -fstack-protector-all -Wall -Waggregate-return -Wbad-function-cast -Wcast-qual -Wcast-align -Wdeclaration-after-statement -Wendif-labels -Wfloat-equal -Wformat=2 -Wformat-security -Wformat-nonliteral -Winline -Wmissing-prototypes -Wmissing-declarations -Wmissing-format-attribute -Wnested-externs -Wno-long-long -Wno-strict-aliasing -Wpointer-arith -Wstrict-prototypes -Wwrite-strings -ansi -D_GNU_SOURCE -DANSI_ONLY -Werror -MT IPv6addr.o -MD -MP -MF .deps/IPv6addr.Tpo -c -o IPv6addr.o IPv6addr.c; \ then mv -f .deps/IPv6addr.Tpo .deps/IPv6addr.Po; else rm -f .deps/IPv6addr.Tpo; exit 1; fi cc1: warnings being treated as errors IPv6addr.c: In function ‘send_ua’: IPv6addr.c:453: warning: passing argument 2 of ‘libnet_pblock_record_ip_offset’ makes pointer from integer without a cast This doesn't happen here with libnet-1.1.2.1-140.75.i586. Which libnet version do you have? IPv6addr.c:438: warning: function call has aggregate value That function (libnet_name2addr6) indeed returns a struct and this is going to happen with -Waggregate-return since not all compilers support such function calls. Don't know what to do about this one. Thanks, Dejan gmake[1]: *** [IPv6addr.o] Error 1 gmake[1]: Leaving directory `/home/rpmbuild/hg/cluster-agents/heartbeat' make: *** [all-recursive] Error 1 In pacemaker, I build without openais-devel installed (don't want to build against OpenAIS 0.80.3 found on CentOS); Heartbeat and Glue are built and installed from current tip. ./configure make leads to: [...] gcc -DHAVE_CONFIG_H -I. -I. -I../../include -I../../include -I../../include -I../../include -I../../libltdl -I../../libltdl -I/usr/include/glib-2.0 -I/usr/lib/glib-2.0/include -I/usr/include/libxml2 -g -O2 -I/usr/include/heartbeat -ggdb3 -O0 -fgnu89-inline -fstack-protector-all -Wall -Waggregate-return -Wbad-function-cast -Wcast-align -Wdeclaration-after-statement -Wendif-labels -Wfloat-equal -Wformat=2 -Wformat-security -Wformat-nonliteral -Winline -Wmissing-prototypes -Wmissing-declarations -Wmissing-format-attribute -Wnested-externs -Wno-long-long -Wno-strict-aliasing -Wpointer-arith -Wstrict-prototypes -Wwrite-strings -ansi -D_GNU_SOURCE -DANSI_ONLY -Werror -fPIC -MT xml.lo -MD -MP -MF .deps/xml.Tpo -c xml.c -fPIC -DPIC -o .libs/xml.o cc1: warnings being treated as errors xml.c: In function 'string2xml': xml.c:490: warning: argument 2 of 'xmlSetGenericErrorFunc' might be a candidate for a format attribute xml.c: In function 'filename2xml': xml.c:626: warning: argument 2 of 'xmlSetGenericErrorFunc' might be a candidate for a format attribute gmake[2]: *** [xml.lo] Error 1 gmake[2]: Leaving directory `/home/rpmbuild/hg/pacemaker-stable-1.0/lib/common' gmake[1]: *** [all-recursive] Error 1 gmake[1]: Leaving directory `/home/rpmbuild/hg/pacemaker-stable-1.0/lib' make: *** [all-recursive] Error 1 As said, all of this is on CentOS 5.3. Cheers, Florian ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Re: [Pacemaker] Why are fatal warnings enabled by default?
On 2009-10-21 14:36, Dejan Muhamedagic wrote: The warnings being? In agents, a simple ./configure make leads to: [...] gmake[1]: Entering directory `/home/rpmbuild/hg/cluster-agents/heartbeat' if gcc -DHAVE_CONFIG_H -I. -I. -I../include -I../include -I../include -I../linux-ha -I/usr/include/glib-2.0 -I/usr/lib/glib-2.0/include-g -O2 -ggdb3 -O0 -fgnu89-inline -fstack-protector-all -Wall -Waggregate-return -Wbad-function-cast -Wcast-qual -Wcast-align -Wdeclaration-after-statement -Wendif-labels -Wfloat-equal -Wformat=2 -Wformat-security -Wformat-nonliteral -Winline -Wmissing-prototypes -Wmissing-declarations -Wmissing-format-attribute -Wnested-externs -Wno-long-long -Wno-strict-aliasing -Wpointer-arith -Wstrict-prototypes -Wwrite-strings -ansi -D_GNU_SOURCE -DANSI_ONLY -Werror -MT IPv6addr.o -MD -MP -MF .deps/IPv6addr.Tpo -c -o IPv6addr.o IPv6addr.c; \ then mv -f .deps/IPv6addr.Tpo .deps/IPv6addr.Po; else rm -f .deps/IPv6addr.Tpo; exit 1; fi cc1: warnings being treated as errors IPv6addr.c: In function ‘send_ua’: IPv6addr.c:453: warning: passing argument 2 of ‘libnet_pblock_record_ip_offset’ makes pointer from integer without a cast This doesn't happen here with libnet-1.1.2.1-140.75.i586. Which libnet version do you have? libnet-1.1.4-3.el5 Cheers, Florian signature.asc Description: OpenPGP digital signature ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker
[Pacemaker] corosync doesn't stop all services
Hi, perhaps this is the wrong list but anyway: I have corosync-1.1.1 and pacemaker-1.0.5 on debian lenny. When I start corosync everything looks fine. But when I stop corosync I still see a lot of heartbeart processes. I thought this was fixed in corosync-1.1.1. so what might be the problem? # ps uax | grep heart root 2083 0.0 0.4 4884 1220 pts/1S 17:04 0:00 /usr/lib/heartbeat/ha_logd -d root 2084 0.0 0.3 4884 820 pts/1S 17:04 0:00 /usr/lib/heartbeat/ha_logd -d root 2099 0.0 4.1 10712 10712 ?SLs 17:04 0:00 /usr/lib/heartbeat/stonithd 104 2100 0.1 1.4 12768 3748 ?S 17:04 0:00 /usr/lib/heartbeat/cib root 2101 0.0 0.7 5352 1800 ?S 17:04 0:00 /usr/lib/heartbeat/lrmd 104 2102 0.0 1.0 12260 2596 ?S 17:04 0:00 /usr/lib/heartbeat/attrd 104 2103 0.0 1.1 8880 3024 ?S 17:04 0:00 /usr/lib/heartbeat/pengine 104 2104 0.0 1.2 12404 3176 ?S 17:04 0:00 /usr/lib/heartbeat/crmd root 2140 0.0 0.2 3116 720 pts/1R+ 17:08 0:00 grep heart -- Dr. Michael Schwartzkopff MultiNET Services GmbH Addresse: Bretonischer Ring 7; 85630 Grasbrunn; Germany Tel: +49 - 89 - 45 69 11 0 Fax: +49 - 89 - 45 69 11 21 mob: +49 - 174 - 343 28 75 mail: mi...@multinet.de web: www.multinet.de Sitz der Gesellschaft: 85630 Grasbrunn Registergericht: Amtsgericht München HRB 114375 Geschäftsführer: Günter Jurgeneit, Hubert Martens --- PGP Fingerprint: F919 3919 FF12 ED5A 2801 DEA6 AA77 57A4 EDD8 979B Skype: misch42 ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Re: [Pacemaker] Why are fatal warnings enabled by default?
Hi, On Wed, Oct 21, 2009 at 03:08:53PM +0200, Florian Haas wrote: On 2009-10-21 14:36, Dejan Muhamedagic wrote: The warnings being? In agents, a simple ./configure make leads to: [...] gmake[1]: Entering directory `/home/rpmbuild/hg/cluster-agents/heartbeat' if gcc -DHAVE_CONFIG_H -I. -I. -I../include -I../include -I../include -I../linux-ha -I/usr/include/glib-2.0 -I/usr/lib/glib-2.0/include-g -O2 -ggdb3 -O0 -fgnu89-inline -fstack-protector-all -Wall -Waggregate-return -Wbad-function-cast -Wcast-qual -Wcast-align -Wdeclaration-after-statement -Wendif-labels -Wfloat-equal -Wformat=2 -Wformat-security -Wformat-nonliteral -Winline -Wmissing-prototypes -Wmissing-declarations -Wmissing-format-attribute -Wnested-externs -Wno-long-long -Wno-strict-aliasing -Wpointer-arith -Wstrict-prototypes -Wwrite-strings -ansi -D_GNU_SOURCE -DANSI_ONLY -Werror -MT IPv6addr.o -MD -MP -MF .deps/IPv6addr.Tpo -c -o IPv6addr.o IPv6addr.c; \ then mv -f .deps/IPv6addr.Tpo .deps/IPv6addr.Po; else rm -f .deps/IPv6addr.Tpo; exit 1; fi cc1: warnings being treated as errors IPv6addr.c: In function ‘send_ua’: IPv6addr.c:453: warning: passing argument 2 of ‘libnet_pblock_record_ip_offset’ makes pointer from integer without a cast This doesn't happen here with libnet-1.1.2.1-140.75.i586. Which libnet version do you have? libnet-1.1.4-3.el5 API changed. I'll see how this can be detected at the configure stage. Cheers, Dejan Cheers, Florian ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Re: [Pacemaker] corosync doesn't stop all services
On Wed, Oct 21, 2009 at 3:11 PM, Michael Schwartzkopff mi...@multinet.de wrote: Hi, perhaps this is the wrong list but anyway: I have corosync-1.1.1 and pacemaker-1.0.5 on debian lenny. Unless you have the latest Pacemaker from Hg, it still wont work. When I start corosync everything looks fine. But when I stop corosync I still see a lot of heartbeart processes. I thought this was fixed in corosync-1.1.1. so what might be the problem? # ps uax | grep heart root 2083 0.0 0.4 4884 1220 pts/1 S 17:04 0:00 /usr/lib/heartbeat/ha_logd -d root 2084 0.0 0.3 4884 820 pts/1 S 17:04 0:00 /usr/lib/heartbeat/ha_logd -d root 2099 0.0 4.1 10712 10712 ? SLs 17:04 0:00 /usr/lib/heartbeat/stonithd 104 2100 0.1 1.4 12768 3748 ? S 17:04 0:00 /usr/lib/heartbeat/cib root 2101 0.0 0.7 5352 1800 ? S 17:04 0:00 /usr/lib/heartbeat/lrmd 104 2102 0.0 1.0 12260 2596 ? S 17:04 0:00 /usr/lib/heartbeat/attrd 104 2103 0.0 1.1 8880 3024 ? S 17:04 0:00 /usr/lib/heartbeat/pengine 104 2104 0.0 1.2 12404 3176 ? S 17:04 0:00 /usr/lib/heartbeat/crmd root 2140 0.0 0.2 3116 720 pts/1 R+ 17:08 0:00 grep heart -- Dr. Michael Schwartzkopff MultiNET Services GmbH Addresse: Bretonischer Ring 7; 85630 Grasbrunn; Germany Tel: +49 - 89 - 45 69 11 0 Fax: +49 - 89 - 45 69 11 21 mob: +49 - 174 - 343 28 75 mail: mi...@multinet.de web: www.multinet.de Sitz der Gesellschaft: 85630 Grasbrunn Registergericht: Amtsgericht München HRB 114375 Geschäftsführer: Günter Jurgeneit, Hubert Martens --- PGP Fingerprint: F919 3919 FF12 ED5A 2801 DEA6 AA77 57A4 EDD8 979B Skype: misch42 ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Re: [Pacemaker] corosync doesn't stop all services
Am Mittwoch, 21. Oktober 2009 15:28:47 schrieb Andrew Beekhof: On Wed, Oct 21, 2009 at 3:11 PM, Michael Schwartzkopff mi...@multinet.de wrote: Hi, perhaps this is the wrong list but anyway: I have corosync-1.1.1 and pacemaker-1.0.5 on debian lenny. Unless you have the latest Pacemaker from Hg, it still wont work. OK, will patch. Thanks. -- Dr. Michael Schwartzkopff MultiNET Services GmbH Addresse: Bretonischer Ring 7; 85630 Grasbrunn; Germany Tel: +49 - 89 - 45 69 11 0 Fax: +49 - 89 - 45 69 11 21 mob: +49 - 174 - 343 28 75 mail: mi...@multinet.de web: www.multinet.de Sitz der Gesellschaft: 85630 Grasbrunn Registergericht: Amtsgericht München HRB 114375 Geschäftsführer: Günter Jurgeneit, Hubert Martens --- PGP Fingerprint: F919 3919 FF12 ED5A 2801 DEA6 AA77 57A4 EDD8 979B Skype: misch42 ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Re: [Pacemaker] corosync doesn't stop all services
We had to change both pacemaker and corosync for this problem. I suspect you don't have the updated pacemaker. Regards -steve On Wed, 2009-10-21 at 15:11 +0200, Michael Schwartzkopff wrote: Hi, perhaps this is the wrong list but anyway: I have corosync-1.1.1 and pacemaker-1.0.5 on debian lenny. When I start corosync everything looks fine. But when I stop corosync I still see a lot of heartbeart processes. I thought this was fixed in corosync-1.1.1. so what might be the problem? # ps uax | grep heart root 2083 0.0 0.4 4884 1220 pts/1S 17:04 0:00 /usr/lib/heartbeat/ha_logd -d root 2084 0.0 0.3 4884 820 pts/1S 17:04 0:00 /usr/lib/heartbeat/ha_logd -d root 2099 0.0 4.1 10712 10712 ?SLs 17:04 0:00 /usr/lib/heartbeat/stonithd 104 2100 0.1 1.4 12768 3748 ?S 17:04 0:00 /usr/lib/heartbeat/cib root 2101 0.0 0.7 5352 1800 ?S 17:04 0:00 /usr/lib/heartbeat/lrmd 104 2102 0.0 1.0 12260 2596 ?S 17:04 0:00 /usr/lib/heartbeat/attrd 104 2103 0.0 1.1 8880 3024 ?S 17:04 0:00 /usr/lib/heartbeat/pengine 104 2104 0.0 1.2 12404 3176 ?S 17:04 0:00 /usr/lib/heartbeat/crmd root 2140 0.0 0.2 3116 720 pts/1R+ 17:08 0:00 grep heart ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker
[Pacemaker] pacemaker unable to start
Hello guys, We are running corosync-1.0.0 heartbeat-2.99.1 pacemaker-1.0.4 the corosync.conf under /etc/corosync/ is # Please read the corosync.conf.5 manual page compatibility: whitetank aisexec { user: root group: root } totem { version: 2 secauth: off threads: 0 interface { ringnumber: 0 bindnetaddr: 172.30.0.0 mcastaddr:226.94.1.1 mcastport: 5406 } } logging { fileline: off to_stderr: yes to_logfile: yes to_syslog: yes logfile: /tmp/corosync.log debug: on timestamp: on logger_subsys { subsys: pacemaker debug: on tags: enter|leave|trace1|trace2| trace3|trace4|trace6 } } service { name: pacemaker ver: 0 # use_mgmtd: yes # use_logd:yes } corosync { user: root group: root } amf { mode: disabled } #service corosync start starts the messaging but fails to load pacemaker, /tmp/corosync.log --- == Oct 21 11:05:43 corosync [MAIN ] Corosync Cluster Engine ('trunk'): started and ready to provide service. Oct 21 11:05:43 corosync [MAIN ] Successfully read main configuration file '/etc/corosync/corosync.conf'. Oct 21 11:05:43 corosync [TOTEM ] Token Timeout (1000 ms) retransmit timeout (238 ms) Oct 21 11:05:43 corosync [TOTEM ] token hold (180 ms) retransmits before loss (4 retrans) Oct 21 11:05:43 corosync [TOTEM ] join (50 ms) send_join (0 ms) consensus (800 ms) merge (200 ms) Oct 21 11:05:43 corosync [TOTEM ] downcheck (1000 ms) fail to recv const (50 msgs) Oct 21 11:05:43 corosync [TOTEM ] seqno unchanged const (30 rotations) Maximum network MTU 1500 Oct 21 11:05:43 corosync [TOTEM ] window size per rotation (50 messages) maximum messages per rotation (17 messages) Oct 21 11:05:43 corosync [TOTEM ] send threads (0 threads) Oct 21 11:05:43 corosync [TOTEM ] RRP token expired timeout (238 ms) Oct 21 11:05:43 corosync [TOTEM ] RRP token problem counter (2000 ms) Oct 21 11:05:43 corosync [TOTEM ] RRP threshold (10 problem count) Oct 21 11:05:43 corosync [TOTEM ] RRP mode set to none. Oct 21 11:05:43 corosync [TOTEM ] heartbeat_failures_allowed (0) Oct 21 11:05:43 corosync [TOTEM ] max_network_delay (50 ms) Oct 21 11:05:43 corosync [TOTEM ] HeartBeat is Disabled. To enable set heartbeat_failures_allowed 0 Oct 21 11:05:43 corosync [TOTEM ] Initializing transmit/receive security: libtomcrypt SOBER128/SHA1HMAC (mode 0). Oct 21 11:05:43 corosync [TOTEM ] Receive multicast socket recv buffer size (262142 bytes). Oct 21 11:05:43 corosync [TOTEM ] Transmit multicast socket send buffer size (262142 bytes). Oct 21 11:05:43 corosync [TOTEM ] The network interface [172.30.0.145] is now up. Oct 21 11:05:43 corosync [TOTEM ] Created or loaded sequence id 184.172.30.0.145 for this ring. Oct 21 11:05:43 corosync [TOTEM ] entering GATHER state from 15. Oct 21 11:05:43 corosync [SERV ] *Service failed to load 'pacemaker'.* Oct 21 11:05:43 corosync [SERV ] Service initialized 'corosync extended virtual synchrony service' Oct 21 11:05:43 corosync [SERV ] Service initialized 'corosync configuration service' Oct 21 11:05:43 corosync [SERV ] Service initialized 'corosync cluster closed process group service v1.01' Oct 21 11:05:43 corosync [SERV ] Service initialized 'corosync cluster config database access v1.01' Oct 21 11:05:43 corosync [SERV ] Service initialized 'corosync profile loading service' Oct 21 11:05:43 corosync [MAIN ] Compatibility mode set to whitetank. Using V1 and V2 of the synchronization engine. Oct 21 11:05:43 corosync [TOTEM ] Creating commit token because I am the rep. Oct 21 11:05:43 corosync [TOTEM ] Saving state aru 0 high seq received 0 Oct 21 11:05:43 corosync [TOTEM ] Storing new sequence id for ring bc Oct 21 11:05:43 corosync [TOTEM ] entering COMMIT state. Oct 21 11:05:43 corosync [TOTEM ] got commit token Oct 21 11:05:43 corosync [TOTEM ] entering RECOVERY state. Oct 21 11:05:43 corosync [TOTEM ] position [0] member 172.30.0.145: Oct 21 11:05:43 corosync [TOTEM ] previous ring seq 184 rep 172.30.0.145 Oct 21 11:05:43 corosync [TOTEM ] aru 0 high delivered 0 received flag 1 Oct 21 11:05:43 corosync [TOTEM ] Did not need to originate any messages in recovery. Oct 21 11:05:43 corosync [TOTEM ] got commit token Oct 21 11:05:43 corosync [TOTEM ] Sending initial ORF token Oct 21 11:05:43 corosync [TOTEM ] token retrans flag is 0 my set retrans flag0 retrans queue empty 1 count 0, aru 0 Oct 21 11:05:43 corosync [TOTEM ] install seq 0 aru 0 high seq received 0 Oct 21 11:05:43 corosync [TOTEM ] token retrans flag is 0 my set retrans flag0 retrans queue empty 1 count 1, aru 0 Oct 21 11:05:43 corosync [TOTEM ] install seq 0 aru 0 high seq received 0 Oct 21 11:05:43 corosync [TOTEM ] token retrans flag is 0 my set retrans flag0 retrans queue empty 1 count 2, aru 0 Oct 21 11:05:43
Re: [Pacemaker] pacemaker unable to start
I recommend using corosync 1.1.1 - several bug fixes one critical for proper pacemaker operation. It won't fix this particular problem however. Corosync loads pacemaker by searching for a pacemaker lcrso file. These files are default installed in /usr/libexec/lcrso but may be in a different location depending on your distribution. Regards -steve On Wed, 2009-10-21 at 11:13 -0400, Shravan Mishra wrote: Hello guys, We are running corosync-1.0.0 heartbeat-2.99.1 pacemaker-1.0.4 the corosync.conf under /etc/corosync/ is # Please read the corosync.conf.5 manual page compatibility: whitetank aisexec { user: root group: root } totem { version: 2 secauth: off threads: 0 interface { ringnumber: 0 bindnetaddr: 172.30.0.0 mcastaddr:226.94.1.1 mcastport: 5406 } } logging { fileline: off to_stderr: yes to_logfile: yes to_syslog: yes logfile: /tmp/corosync.log debug: on timestamp: on logger_subsys { subsys: pacemaker debug: on tags: enter|leave|trace1|trace2| trace3|trace4|trace6 } } service { name: pacemaker ver: 0 # use_mgmtd: yes # use_logd:yes } corosync { user: root group: root } amf { mode: disabled } #service corosync start starts the messaging but fails to load pacemaker, /tmp/corosync.log --- == Oct 21 11:05:43 corosync [MAIN ] Corosync Cluster Engine ('trunk'): started and ready to provide service. Oct 21 11:05:43 corosync [MAIN ] Successfully read main configuration file '/etc/corosync/corosync.conf'. Oct 21 11:05:43 corosync [TOTEM ] Token Timeout (1000 ms) retransmit timeout (238 ms) Oct 21 11:05:43 corosync [TOTEM ] token hold (180 ms) retransmits before loss (4 retrans) Oct 21 11:05:43 corosync [TOTEM ] join (50 ms) send_join (0 ms) consensus (800 ms) merge (200 ms) Oct 21 11:05:43 corosync [TOTEM ] downcheck (1000 ms) fail to recv const (50 msgs) Oct 21 11:05:43 corosync [TOTEM ] seqno unchanged const (30 rotations) Maximum network MTU 1500 Oct 21 11:05:43 corosync [TOTEM ] window size per rotation (50 messages) maximum messages per rotation (17 messages) Oct 21 11:05:43 corosync [TOTEM ] send threads (0 threads) Oct 21 11:05:43 corosync [TOTEM ] RRP token expired timeout (238 ms) Oct 21 11:05:43 corosync [TOTEM ] RRP token problem counter (2000 ms) Oct 21 11:05:43 corosync [TOTEM ] RRP threshold (10 problem count) Oct 21 11:05:43 corosync [TOTEM ] RRP mode set to none. Oct 21 11:05:43 corosync [TOTEM ] heartbeat_failures_allowed (0) Oct 21 11:05:43 corosync [TOTEM ] max_network_delay (50 ms) Oct 21 11:05:43 corosync [TOTEM ] HeartBeat is Disabled. To enable set heartbeat_failures_allowed 0 Oct 21 11:05:43 corosync [TOTEM ] Initializing transmit/receive security: libtomcrypt SOBER128/SHA1HMAC (mode 0). Oct 21 11:05:43 corosync [TOTEM ] Receive multicast socket recv buffer size (262142 bytes). Oct 21 11:05:43 corosync [TOTEM ] Transmit multicast socket send buffer size (262142 bytes). Oct 21 11:05:43 corosync [TOTEM ] The network interface [172.30.0.145] is now up. Oct 21 11:05:43 corosync [TOTEM ] Created or loaded sequence id 184.172.30.0.145 for this ring. Oct 21 11:05:43 corosync [TOTEM ] entering GATHER state from 15. Oct 21 11:05:43 corosync [SERV ] Service failed to load 'pacemaker'. Oct 21 11:05:43 corosync [SERV ] Service initialized 'corosync extended virtual synchrony service' Oct 21 11:05:43 corosync [SERV ] Service initialized 'corosync configuration service' Oct 21 11:05:43 corosync [SERV ] Service initialized 'corosync cluster closed process group service v1.01' Oct 21 11:05:43 corosync [SERV ] Service initialized 'corosync cluster config database access v1.01' Oct 21 11:05:43 corosync [SERV ] Service initialized 'corosync profile loading service' Oct 21 11:05:43 corosync [MAIN ] Compatibility mode set to whitetank. Using V1 and V2 of the synchronization engine. Oct 21 11:05:43 corosync [TOTEM ] Creating commit token because I am the rep. Oct 21 11:05:43 corosync [TOTEM ] Saving state aru 0 high seq received 0 Oct 21 11:05:43 corosync [TOTEM ] Storing new sequence id for ring bc Oct 21 11:05:43 corosync [TOTEM ] entering COMMIT state. Oct 21 11:05:43 corosync [TOTEM ] got commit token Oct 21 11:05:43 corosync [TOTEM ] entering RECOVERY state. Oct 21 11:05:43 corosync [TOTEM ] position [0] member 172.30.0.145: Oct 21 11:05:43 corosync [TOTEM ] previous ring seq 184 rep 172.30.0.145 Oct 21 11:05:43 corosync [TOTEM ] aru 0 high delivered 0 received flag 1 Oct 21 11:05:43 corosync [TOTEM ] Did not need to originate any messages in recovery. Oct 21 11:05:43 corosync [TOTEM ] got commit token Oct 21
Re: [Pacemaker] Why are fatal warnings enabled by default?
On Wed, 21 Oct 2009, 17:56:25 +0200, Florian Haas wrote: On 10/21/2009 03:17 PM, Dejan Muhamedagic wrote: This doesn't happen here with libnet-1.1.2.1-140.75.i586. Which libnet version do you have? libnet-1.1.4-3.el5 API changed. I'll see how this can be detected at the configure stage. Now since this seems to be a libnet related problem, why does it also occur when I configure with --disable-libnet? By contrast it does go away when libnet is uninstalled. Is some autoconf magic broken there too? I'm not sure if there are any real rules as to which packages (and which versions) are required for successfull building. I remember similar issues with GCC over 10 years ago, after which we decided to document the stuff required for building it successfully. Do we have such rules documented for Pacemaker? Cheers. l8er manfred ___ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Re: [Pacemaker] pacemaker unable to start
Ya your missing the pacemaker lcrso file. Either you didn't build pacemaker with corosync support or pacemaker didn't install that binary in the proper place. try: updatedb locate lcrso Regards -steve On Wed, 2009-10-21 at 12:28 -0400, Shravan Mishra wrote: Steve, this is what my installation shows-- ls -l /usr/libexec/lcrso -rwxr-xr-x 1 root root 101243 Jul 29 11:21 coroparse.lcrso -rwxr-xr-x 1 root root 117688 Jul 29 11:21 objdb.lcrso -rwxr-xr-x 1 root root 92702 Jul 29 11:54 openaisserviceenable.lcrso -rwxr-xr-x 1 root root 110808 Jul 29 11:21 quorum_testquorum.lcrso -rwxr-xr-x 1 root root 159057 Jul 29 11:21 quorum_votequorum.lcrso -rwxr-xr-x 1 root root 1175430 Jul 29 11:54 service_amf.lcrso -rwxr-xr-x 1 root root 133976 Jul 29 11:21 service_cfg.lcrso -rwxr-xr-x 1 root root 218374 Jul 29 11:54 service_ckpt.lcrso -rwxr-xr-x 1 root root 139029 Jul 29 11:54 service_clm.lcrso -rwxr-xr-x 1 root root 122668 Jul 29 11:21 service_confdb.lcrso -rwxr-xr-x 1 root root 138412 Jul 29 11:21 service_cpg.lcrso -rwxr-xr-x 1 root root 125638 Jul 29 11:21 service_evs.lcrso -rwxr-xr-x 1 root root 196443 Jul 29 11:54 service_evt.lcrso -rwxr-xr-x 1 root root 194885 Jul 29 11:54 service_lck.lcrso -rwxr-xr-x 1 root root 235168 Jul 29 11:54 service_msg.lcrso -rwxr-xr-x 1 root root 120445 Jul 29 11:21 service_pload.lcrso -rwxr-xr-x 1 root root 135340 Jul 29 11:54 service_tmr.lcrso -rwxr-xr-x 1 root root 124092 Jul 29 11:21 vsf_quorum.lcrso -rwxr-xr-x 1 root root 121298 Jul 29 11:21 vsf_ykd.lcrso I also did export COROSYNC_DEFAULT_CONFIG_IFACE=openaisserviceenable:openaisparser In place of openaisparser I also tried corosyncparse and corosyncparser but to no avail. -sincerely Shravan On Wed, Oct 21, 2009 at 11:49 AM, Steven Dake sd...@redhat.com wrote: I recommend using corosync 1.1.1 - several bug fixes one critical for proper pacemaker operation. It won't fix this particular problem however. Corosync loads pacemaker by searching for a pacemaker lcrso file. These files are default installed in /usr/libexec/lcrso but may be in a different location depending on your distribution. Regards -steve On Wed, 2009-10-21 at 11:13 -0400, Shravan Mishra wrote: Hello guys, We are running corosync-1.0.0 heartbeat-2.99.1 pacemaker-1.0.4 the corosync.conf under /etc/corosync/ is # Please read the corosync.conf.5 manual page compatibility: whitetank aisexec { user: root group: root } totem { version: 2 secauth: off threads: 0 interface { ringnumber: 0 bindnetaddr: 172.30.0.0 mcastaddr:226.94.1.1 mcastport: 5406 } } logging { fileline: off to_stderr: yes to_logfile: yes to_syslog: yes logfile: /tmp/corosync.log debug: on timestamp: on logger_subsys { subsys: pacemaker debug: on tags: enter|leave|trace1|trace2| trace3|trace4|trace6 } } service { name: pacemaker ver: 0 # use_mgmtd: yes # use_logd:yes } corosync { user: root group: root } amf { mode: disabled } #service corosync start starts the messaging but fails to load pacemaker, /tmp/corosync.log --- == Oct 21 11:05:43 corosync [MAIN ] Corosync Cluster Engine ('trunk'): started and ready to provide service. Oct 21 11:05:43 corosync [MAIN ] Successfully read main configuration file '/etc/corosync/corosync.conf'. Oct 21 11:05:43 corosync [TOTEM ] Token Timeout (1000 ms) retransmit timeout (238 ms) Oct 21 11:05:43 corosync [TOTEM ] token hold (180 ms) retransmits before loss (4 retrans) Oct 21 11:05:43 corosync [TOTEM ] join (50 ms) send_join (0 ms) consensus (800 ms) merge (200 ms) Oct 21 11:05:43 corosync [TOTEM ] downcheck (1000 ms) fail to recv const (50 msgs) Oct 21 11:05:43 corosync [TOTEM ] seqno unchanged const (30 rotations) Maximum network MTU 1500 Oct 21 11:05:43 corosync [TOTEM ] window size per rotation (50 messages) maximum messages per rotation (17 messages) Oct 21 11:05:43 corosync [TOTEM ] send threads (0 threads) Oct 21 11:05:43 corosync [TOTEM ] RRP token expired timeout (238 ms) Oct 21 11:05:43 corosync [TOTEM ] RRP token problem counter (2000 ms) Oct 21 11:05:43 corosync [TOTEM ] RRP threshold (10 problem count) Oct 21 11:05:43 corosync [TOTEM ] RRP mode set to none. Oct 21 11:05:43 corosync [TOTEM ] heartbeat_failures_allowed (0) Oct 21 11:05:43 corosync [TOTEM ] max_network_delay (50 ms) Oct 21 11:05:43 corosync [TOTEM ] HeartBeat is Disabled. To enable set heartbeat_failures_allowed 0 Oct 21 11:05:43 corosync [TOTEM ] Initializing transmit/receive security: