Re: [Pacemaker] Missing autoconf check for sensors.h?

2009-10-21 Thread Andrew Beekhof
On Wed, Oct 21, 2009 at 8:53 AM, Florian Haas florian.h...@linbit.com wrote:
 OK, so building on a system where net-snmp-devel is not installed would
 circumvent this problem?


yep, or if you installed the sensors devel package

___
Pacemaker mailing list
Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker


Re: [Pacemaker] Missing autoconf check for sensors.h?

2009-10-21 Thread Florian Haas
On 10/21/2009 09:01 AM, Andrew Beekhof wrote:
 On Wed, Oct 21, 2009 at 8:53 AM, Florian Haas florian.h...@linbit.com wrote:
 OK, so building on a system where net-snmp-devel is not installed would
 circumvent this problem?

 
 yep, or if you installed the sensors devel package

yeah, that I realized by myself. :)

Cheers,
Florian



signature.asc
Description: OpenPGP digital signature
___
Pacemaker mailing list
Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker


[Pacemaker] Why are fatal warnings enabled by default?

2009-10-21 Thread Florian Haas
Andrew, Dejan,

For pacemaker and agents, configure defaults to --enable-fatal-warnings.
AFAIR, neither of these have ever built successfully with fatal warnings
enabled. Is there a specific reason to keep the default as it is? Is
this perhaps a deliberate entry barrier for packagers, so as to make
sure they know what they're doing? :) Or would it be possible to
actually fix those warnings, so the packages would compile -- *gasp* --
even with the configure defaults?

Cheers,
Florian





signature.asc
Description: OpenPGP digital signature
___
Pacemaker mailing list
Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker


Re: [Pacemaker] Dual powersupply stonith

2009-10-21 Thread Johan Verrept
On Wed, 2009-10-21 at 01:16 +0200, Dejan Muhamedagic wrote:
  In that case, we would need to send two synchronised stonith actions to
  two different stonith devices to succesfully stonith a node. Would this
  be best solved at pacemaker or at RA level? (Assuming this isn't solved
  at hardware level offcourse).
 
 That's not possible at all right now. Unless you go with another
 fencing solution such as sbd, but that's a different matter.

Hmm. If a single RA would (transparantly) control two stonith devices,
it should work?  Error handling would require some thinking though.
Unless there are some things I am missing, it should be possible with
two of those IP Power things. Not a generic solution you want to use in
the project though.

J.


___
Pacemaker mailing list
Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker


Re: [Pacemaker] Why are fatal warnings enabled by default?

2009-10-21 Thread Andrew Beekhof
On Wed, Oct 21, 2009 at 9:31 AM, Florian Haas florian.h...@linbit.com wrote:
 Andrew, Dejan,

 For pacemaker and agents, configure defaults to --enable-fatal-warnings.

So that the warnings get fixed.

 AFAIR, neither of these have ever built successfully with fatal warnings
 enabled.

Only on Debian.  shrug.
Everywhere else it builds just fine.

 Is there a specific reason to keep the default as it is? Is
 this perhaps a deliberate entry barrier for packagers, so as to make
 sure they know what they're doing? :) Or would it be possible to
 actually fix those warnings, so the packages would compile -- *gasp* --
 even with the configure defaults?

 Cheers,
 Florian




 ___
 Pacemaker mailing list
 Pacemaker@oss.clusterlabs.org
 http://oss.clusterlabs.org/mailman/listinfo/pacemaker



___
Pacemaker mailing list
Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker


Re: [Pacemaker] Dual powersupply stonith

2009-10-21 Thread Dejan Muhamedagic
Hi,

On Wed, Oct 21, 2009 at 09:34:06AM +0200, Johan Verrept wrote:
 On Wed, 2009-10-21 at 01:16 +0200, Dejan Muhamedagic wrote:
   In that case, we would need to send two synchronised stonith actions to
   two different stonith devices to succesfully stonith a node. Would this
   be best solved at pacemaker or at RA level? (Assuming this isn't solved
   at hardware level offcourse).
  
  That's not possible at all right now. Unless you go with another
  fencing solution such as sbd, but that's a different matter.
 
 Hmm. If a single RA would (transparantly) control two stonith devices,
 it should work?

Yes.

 Error handling would require some thinking though.
 Unless there are some things I am missing, it should be possible with
 two of those IP Power things. Not a generic solution you want to use in
 the project though.

No. In this case, since the power distribution is rather robust,
it would be great to have am on-board management card such as HP
iLO or IBM RSA and then use that for fencing.

Thanks,

Dejan

   J.
 
 
 ___
 Pacemaker mailing list
 Pacemaker@oss.clusterlabs.org
 http://oss.clusterlabs.org/mailman/listinfo/pacemaker

___
Pacemaker mailing list
Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker


Re: [Pacemaker] Why are fatal warnings enabled by default?

2009-10-21 Thread Dejan Muhamedagic
Hi,

On Wed, Oct 21, 2009 at 01:36:59PM +0200, Florian Haas wrote:
 On 10/21/2009 11:34 AM, Andrew Beekhof wrote:
  On Wed, Oct 21, 2009 at 11:04 AM, Florian Haas florian.h...@linbit.com 
  wrote:
  On 2009-10-21 10:49, Andrew Beekhof wrote:
  On Wed, Oct 21, 2009 at 9:31 AM, Florian Haas florian.h...@linbit.com 
  wrote:
  Andrew, Dejan,
 
  For pacemaker and agents, configure defaults to --enable-fatal-warnings.
  So that the warnings get fixed.
 
  AFAIR, neither of these have ever built successfully with fatal warnings
  enabled.
  Only on Debian.  shrug.
  Everywhere else it builds just fine.
  Unless everywhere else excludes CentOS 5, where it doesn't build
  either. At least for me it doesn't. Don't know about upstream RHEL.
  
  
  The warnings being?
 
 In agents, a simple ./configure  make leads to:
 
 [...]
 gmake[1]: Entering directory `/home/rpmbuild/hg/cluster-agents/heartbeat'
 if gcc -DHAVE_CONFIG_H -I. -I. -I../include -I../include -I../include
 -I../linux-ha  -I/usr/include/glib-2.0 -I/usr/lib/glib-2.0/include-g
 -O2 -ggdb3 -O0  -fgnu89-inline -fstack-protector-all -Wall
 -Waggregate-return -Wbad-function-cast -Wcast-qual -Wcast-align
 -Wdeclaration-after-statement -Wendif-labels -Wfloat-equal -Wformat=2
 -Wformat-security -Wformat-nonliteral -Winline -Wmissing-prototypes
 -Wmissing-declarations -Wmissing-format-attribute -Wnested-externs
 -Wno-long-long -Wno-strict-aliasing -Wpointer-arith -Wstrict-prototypes
 -Wwrite-strings -ansi -D_GNU_SOURCE -DANSI_ONLY -Werror -MT IPv6addr.o
 -MD -MP -MF .deps/IPv6addr.Tpo -c -o IPv6addr.o IPv6addr.c; \
   then mv -f .deps/IPv6addr.Tpo .deps/IPv6addr.Po; else rm -f
 .deps/IPv6addr.Tpo; exit 1; fi
 cc1: warnings being treated as errors
 IPv6addr.c: In function ‘send_ua’:
 IPv6addr.c:453: warning: passing argument 2 of
 ‘libnet_pblock_record_ip_offset’ makes pointer from integer without a cast

This doesn't happen here with libnet-1.1.2.1-140.75.i586. Which
libnet version do you have?

 IPv6addr.c:438: warning: function call has aggregate value

That function (libnet_name2addr6) indeed returns a struct and
this is going to happen with -Waggregate-return since not all
compilers support such function calls. Don't know what to do
about this one.

Thanks,

Dejan

 gmake[1]: *** [IPv6addr.o] Error 1
 gmake[1]: Leaving directory `/home/rpmbuild/hg/cluster-agents/heartbeat'
 make: *** [all-recursive] Error 1
 
 In pacemaker, I build without openais-devel installed (don't want to
 build against OpenAIS 0.80.3 found on CentOS); Heartbeat and Glue are
 built and installed from current tip. ./configure  make leads to:
 
 [...]
 gcc -DHAVE_CONFIG_H -I. -I. -I../../include -I../../include
 -I../../include -I../../include -I../../libltdl -I../../libltdl
 -I/usr/include/glib-2.0 -I/usr/lib/glib-2.0/include
 -I/usr/include/libxml2 -g -O2 -I/usr/include/heartbeat -ggdb3 -O0
 -fgnu89-inline -fstack-protector-all -Wall -Waggregate-return
 -Wbad-function-cast -Wcast-align -Wdeclaration-after-statement
 -Wendif-labels -Wfloat-equal -Wformat=2 -Wformat-security
 -Wformat-nonliteral -Winline -Wmissing-prototypes -Wmissing-declarations
 -Wmissing-format-attribute -Wnested-externs -Wno-long-long
 -Wno-strict-aliasing -Wpointer-arith -Wstrict-prototypes -Wwrite-strings
 -ansi -D_GNU_SOURCE -DANSI_ONLY -Werror -fPIC -MT xml.lo -MD -MP -MF
 .deps/xml.Tpo -c xml.c  -fPIC -DPIC -o .libs/xml.o
 cc1: warnings being treated as errors
 xml.c: In function 'string2xml':
 xml.c:490: warning: argument 2 of 'xmlSetGenericErrorFunc' might be a
 candidate for a format attribute
 xml.c: In function 'filename2xml':
 xml.c:626: warning: argument 2 of 'xmlSetGenericErrorFunc' might be a
 candidate for a format attribute
 gmake[2]: *** [xml.lo] Error 1
 gmake[2]: Leaving directory
 `/home/rpmbuild/hg/pacemaker-stable-1.0/lib/common'
 gmake[1]: *** [all-recursive] Error 1
 gmake[1]: Leaving directory `/home/rpmbuild/hg/pacemaker-stable-1.0/lib'
 make: *** [all-recursive] Error 1
 
 As said, all of this is on CentOS 5.3.
 
 Cheers,
 Florian
 



 ___
 Pacemaker mailing list
 Pacemaker@oss.clusterlabs.org
 http://oss.clusterlabs.org/mailman/listinfo/pacemaker


___
Pacemaker mailing list
Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker


Re: [Pacemaker] Why are fatal warnings enabled by default?

2009-10-21 Thread Florian Haas
On 2009-10-21 14:36, Dejan Muhamedagic wrote:
 The warnings being?
 In agents, a simple ./configure  make leads to:

 [...]
 gmake[1]: Entering directory `/home/rpmbuild/hg/cluster-agents/heartbeat'
 if gcc -DHAVE_CONFIG_H -I. -I. -I../include -I../include -I../include
 -I../linux-ha  -I/usr/include/glib-2.0 -I/usr/lib/glib-2.0/include-g
 -O2 -ggdb3 -O0  -fgnu89-inline -fstack-protector-all -Wall
 -Waggregate-return -Wbad-function-cast -Wcast-qual -Wcast-align
 -Wdeclaration-after-statement -Wendif-labels -Wfloat-equal -Wformat=2
 -Wformat-security -Wformat-nonliteral -Winline -Wmissing-prototypes
 -Wmissing-declarations -Wmissing-format-attribute -Wnested-externs
 -Wno-long-long -Wno-strict-aliasing -Wpointer-arith -Wstrict-prototypes
 -Wwrite-strings -ansi -D_GNU_SOURCE -DANSI_ONLY -Werror -MT IPv6addr.o
 -MD -MP -MF .deps/IPv6addr.Tpo -c -o IPv6addr.o IPv6addr.c; \
  then mv -f .deps/IPv6addr.Tpo .deps/IPv6addr.Po; else rm -f
 .deps/IPv6addr.Tpo; exit 1; fi
 cc1: warnings being treated as errors
 IPv6addr.c: In function ‘send_ua’:
 IPv6addr.c:453: warning: passing argument 2 of
 ‘libnet_pblock_record_ip_offset’ makes pointer from integer without a cast
 
 This doesn't happen here with libnet-1.1.2.1-140.75.i586. Which
 libnet version do you have?

libnet-1.1.4-3.el5

Cheers,
Florian



signature.asc
Description: OpenPGP digital signature
___
Pacemaker mailing list
Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker


[Pacemaker] corosync doesn't stop all services

2009-10-21 Thread Michael Schwartzkopff
Hi,

perhaps this is the wrong list but anyway:

I have corosync-1.1.1 and pacemaker-1.0.5 on debian lenny.

When I start corosync everything looks fine. But when I stop corosync I still 
see a lot of heartbeart processes. I thought this was fixed in corosync-1.1.1. 
so what might be the problem?

# ps uax | grep heart
root  2083  0.0  0.4   4884  1220 pts/1S   17:04   0:00 
/usr/lib/heartbeat/ha_logd -d
root  2084  0.0  0.3   4884   820 pts/1S   17:04   0:00 
/usr/lib/heartbeat/ha_logd -d
root  2099  0.0  4.1  10712 10712 ?SLs 17:04   0:00 
/usr/lib/heartbeat/stonithd
104   2100  0.1  1.4  12768  3748 ?S   17:04   0:00 
/usr/lib/heartbeat/cib
root  2101  0.0  0.7   5352  1800 ?S   17:04   0:00 
/usr/lib/heartbeat/lrmd
104   2102  0.0  1.0  12260  2596 ?S   17:04   0:00 
/usr/lib/heartbeat/attrd
104   2103  0.0  1.1   8880  3024 ?S   17:04   0:00 
/usr/lib/heartbeat/pengine
104   2104  0.0  1.2  12404  3176 ?S   17:04   0:00 
/usr/lib/heartbeat/crmd
root  2140  0.0  0.2   3116   720 pts/1R+  17:08   0:00 grep heart

-- 
Dr. Michael Schwartzkopff
MultiNET Services GmbH
Addresse: Bretonischer Ring 7; 85630 Grasbrunn; Germany
Tel: +49 - 89 - 45 69 11 0
Fax: +49 - 89 - 45 69 11 21
mob: +49 - 174 - 343 28 75

mail: mi...@multinet.de
web: www.multinet.de

Sitz der Gesellschaft: 85630 Grasbrunn
Registergericht: Amtsgericht München HRB 114375
Geschäftsführer: Günter Jurgeneit, Hubert Martens

---

PGP Fingerprint: F919 3919 FF12 ED5A 2801 DEA6 AA77 57A4 EDD8 979B
Skype: misch42

___
Pacemaker mailing list
Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker


Re: [Pacemaker] Why are fatal warnings enabled by default?

2009-10-21 Thread Dejan Muhamedagic
Hi,

On Wed, Oct 21, 2009 at 03:08:53PM +0200, Florian Haas wrote:
 On 2009-10-21 14:36, Dejan Muhamedagic wrote:
  The warnings being?
  In agents, a simple ./configure  make leads to:
 
  [...]
  gmake[1]: Entering directory `/home/rpmbuild/hg/cluster-agents/heartbeat'
  if gcc -DHAVE_CONFIG_H -I. -I. -I../include -I../include -I../include
  -I../linux-ha  -I/usr/include/glib-2.0 -I/usr/lib/glib-2.0/include-g
  -O2 -ggdb3 -O0  -fgnu89-inline -fstack-protector-all -Wall
  -Waggregate-return -Wbad-function-cast -Wcast-qual -Wcast-align
  -Wdeclaration-after-statement -Wendif-labels -Wfloat-equal -Wformat=2
  -Wformat-security -Wformat-nonliteral -Winline -Wmissing-prototypes
  -Wmissing-declarations -Wmissing-format-attribute -Wnested-externs
  -Wno-long-long -Wno-strict-aliasing -Wpointer-arith -Wstrict-prototypes
  -Wwrite-strings -ansi -D_GNU_SOURCE -DANSI_ONLY -Werror -MT IPv6addr.o
  -MD -MP -MF .deps/IPv6addr.Tpo -c -o IPv6addr.o IPv6addr.c; \
 then mv -f .deps/IPv6addr.Tpo .deps/IPv6addr.Po; else rm -f
  .deps/IPv6addr.Tpo; exit 1; fi
  cc1: warnings being treated as errors
  IPv6addr.c: In function ‘send_ua’:
  IPv6addr.c:453: warning: passing argument 2 of
  ‘libnet_pblock_record_ip_offset’ makes pointer from integer without a cast
  
  This doesn't happen here with libnet-1.1.2.1-140.75.i586. Which
  libnet version do you have?
 
 libnet-1.1.4-3.el5

API changed. I'll see how this can be detected at the configure
stage.

Cheers,

Dejan


 Cheers,
 Florian
 



 ___
 Pacemaker mailing list
 Pacemaker@oss.clusterlabs.org
 http://oss.clusterlabs.org/mailman/listinfo/pacemaker


___
Pacemaker mailing list
Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker


Re: [Pacemaker] corosync doesn't stop all services

2009-10-21 Thread Andrew Beekhof
On Wed, Oct 21, 2009 at 3:11 PM, Michael Schwartzkopff
mi...@multinet.de wrote:
 Hi,

 perhaps this is the wrong list but anyway:

 I have corosync-1.1.1 and pacemaker-1.0.5 on debian lenny.

Unless you have the latest Pacemaker from Hg, it still wont work.


 When I start corosync everything looks fine. But when I stop corosync I still
 see a lot of heartbeart processes. I thought this was fixed in corosync-1.1.1.
 so what might be the problem?

 # ps uax | grep heart
 root      2083  0.0  0.4   4884  1220 pts/1    S   17:04   0:00
 /usr/lib/heartbeat/ha_logd -d
 root      2084  0.0  0.3   4884   820 pts/1    S   17:04   0:00
 /usr/lib/heartbeat/ha_logd -d
 root      2099  0.0  4.1  10712 10712 ?        SLs 17:04   0:00
 /usr/lib/heartbeat/stonithd
 104       2100  0.1  1.4  12768  3748 ?        S   17:04   0:00
 /usr/lib/heartbeat/cib
 root      2101  0.0  0.7   5352  1800 ?        S   17:04   0:00
 /usr/lib/heartbeat/lrmd
 104       2102  0.0  1.0  12260  2596 ?        S   17:04   0:00
 /usr/lib/heartbeat/attrd
 104       2103  0.0  1.1   8880  3024 ?        S   17:04   0:00
 /usr/lib/heartbeat/pengine
 104       2104  0.0  1.2  12404  3176 ?        S   17:04   0:00
 /usr/lib/heartbeat/crmd
 root      2140  0.0  0.2   3116   720 pts/1    R+  17:08   0:00 grep heart

 --
 Dr. Michael Schwartzkopff
 MultiNET Services GmbH
 Addresse: Bretonischer Ring 7; 85630 Grasbrunn; Germany
 Tel: +49 - 89 - 45 69 11 0
 Fax: +49 - 89 - 45 69 11 21
 mob: +49 - 174 - 343 28 75

 mail: mi...@multinet.de
 web: www.multinet.de

 Sitz der Gesellschaft: 85630 Grasbrunn
 Registergericht: Amtsgericht München HRB 114375
 Geschäftsführer: Günter Jurgeneit, Hubert Martens

 ---

 PGP Fingerprint: F919 3919 FF12 ED5A 2801 DEA6 AA77 57A4 EDD8 979B
 Skype: misch42

 ___
 Pacemaker mailing list
 Pacemaker@oss.clusterlabs.org
 http://oss.clusterlabs.org/mailman/listinfo/pacemaker


___
Pacemaker mailing list
Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker


Re: [Pacemaker] corosync doesn't stop all services

2009-10-21 Thread Michael Schwartzkopff
Am Mittwoch, 21. Oktober 2009 15:28:47 schrieb Andrew Beekhof:
 On Wed, Oct 21, 2009 at 3:11 PM, Michael Schwartzkopff

 mi...@multinet.de wrote:
  Hi,
 
  perhaps this is the wrong list but anyway:
 
  I have corosync-1.1.1 and pacemaker-1.0.5 on debian lenny.

 Unless you have the latest Pacemaker from Hg, it still wont work.

OK, will patch. Thanks.

-- 
Dr. Michael Schwartzkopff
MultiNET Services GmbH
Addresse: Bretonischer Ring 7; 85630 Grasbrunn; Germany
Tel: +49 - 89 - 45 69 11 0
Fax: +49 - 89 - 45 69 11 21
mob: +49 - 174 - 343 28 75

mail: mi...@multinet.de
web: www.multinet.de

Sitz der Gesellschaft: 85630 Grasbrunn
Registergericht: Amtsgericht München HRB 114375
Geschäftsführer: Günter Jurgeneit, Hubert Martens

---

PGP Fingerprint: F919 3919 FF12 ED5A 2801 DEA6 AA77 57A4 EDD8 979B
Skype: misch42

___
Pacemaker mailing list
Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker


Re: [Pacemaker] corosync doesn't stop all services

2009-10-21 Thread Steven Dake
We had to change both pacemaker and corosync for this problem.  I
suspect you don't have the updated pacemaker.

Regards
-steve

On Wed, 2009-10-21 at 15:11 +0200, Michael Schwartzkopff wrote:
 Hi,
 
 perhaps this is the wrong list but anyway:
 
 I have corosync-1.1.1 and pacemaker-1.0.5 on debian lenny.
 
 When I start corosync everything looks fine. But when I stop corosync I still 
 see a lot of heartbeart processes. I thought this was fixed in 
 corosync-1.1.1. 
 so what might be the problem?
 
 # ps uax | grep heart
 root  2083  0.0  0.4   4884  1220 pts/1S   17:04   0:00 
 /usr/lib/heartbeat/ha_logd -d
 root  2084  0.0  0.3   4884   820 pts/1S   17:04   0:00 
 /usr/lib/heartbeat/ha_logd -d
 root  2099  0.0  4.1  10712 10712 ?SLs 17:04   0:00 
 /usr/lib/heartbeat/stonithd
 104   2100  0.1  1.4  12768  3748 ?S   17:04   0:00 
 /usr/lib/heartbeat/cib
 root  2101  0.0  0.7   5352  1800 ?S   17:04   0:00 
 /usr/lib/heartbeat/lrmd
 104   2102  0.0  1.0  12260  2596 ?S   17:04   0:00 
 /usr/lib/heartbeat/attrd
 104   2103  0.0  1.1   8880  3024 ?S   17:04   0:00 
 /usr/lib/heartbeat/pengine
 104   2104  0.0  1.2  12404  3176 ?S   17:04   0:00 
 /usr/lib/heartbeat/crmd
 root  2140  0.0  0.2   3116   720 pts/1R+  17:08   0:00 grep heart
 


___
Pacemaker mailing list
Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker


[Pacemaker] pacemaker unable to start

2009-10-21 Thread Shravan Mishra
Hello guys,

We are running

corosync-1.0.0
heartbeat-2.99.1
pacemaker-1.0.4

the corosync.conf  under /etc/corosync/ is


# Please read the corosync.conf.5 manual page
compatibility: whitetank

aisexec {
   user: root
   group: root
}
totem {
   version: 2
   secauth: off
   threads: 0
   interface {
   ringnumber: 0
   bindnetaddr: 172.30.0.0
   mcastaddr:226.94.1.1
   mcastport: 5406
   }
}

logging {
   fileline: off
   to_stderr: yes
   to_logfile: yes
   to_syslog: yes
   logfile: /tmp/corosync.log
   debug: on
   timestamp: on
   logger_subsys {
   subsys: pacemaker
   debug: on
   tags: enter|leave|trace1|trace2| trace3|trace4|trace6
   }
}


service {
   name: pacemaker
   ver: 0
#   use_mgmtd: yes
 #  use_logd:yes
}


corosync {
   user: root
   group: root
}


amf {
   mode: disabled
}



#service corosync start

starts the messaging but fails to load pacemaker,

/tmp/corosync.log  ---

==

Oct 21 11:05:43 corosync [MAIN  ] Corosync Cluster Engine ('trunk'): started
and ready to provide service.
Oct 21 11:05:43 corosync [MAIN  ] Successfully read main configuration file
'/etc/corosync/corosync.conf'.
Oct 21 11:05:43 corosync [TOTEM ] Token Timeout (1000 ms) retransmit timeout
(238 ms)
Oct 21 11:05:43 corosync [TOTEM ] token hold (180 ms) retransmits before
loss (4 retrans)
Oct 21 11:05:43 corosync [TOTEM ] join (50 ms) send_join (0 ms) consensus
(800 ms) merge (200 ms)
Oct 21 11:05:43 corosync [TOTEM ] downcheck (1000 ms) fail to recv const (50
msgs)
Oct 21 11:05:43 corosync [TOTEM ] seqno unchanged const (30 rotations)
Maximum network MTU 1500
Oct 21 11:05:43 corosync [TOTEM ] window size per rotation (50 messages)
maximum messages per rotation (17 messages)
Oct 21 11:05:43 corosync [TOTEM ] send threads (0 threads)
Oct 21 11:05:43 corosync [TOTEM ] RRP token expired timeout (238 ms)
Oct 21 11:05:43 corosync [TOTEM ] RRP token problem counter (2000 ms)
Oct 21 11:05:43 corosync [TOTEM ] RRP threshold (10 problem count)
Oct 21 11:05:43 corosync [TOTEM ] RRP mode set to none.
Oct 21 11:05:43 corosync [TOTEM ] heartbeat_failures_allowed (0)
Oct 21 11:05:43 corosync [TOTEM ] max_network_delay (50 ms)
Oct 21 11:05:43 corosync [TOTEM ] HeartBeat is Disabled. To enable set
heartbeat_failures_allowed  0
Oct 21 11:05:43 corosync [TOTEM ] Initializing transmit/receive security:
libtomcrypt SOBER128/SHA1HMAC (mode 0).
Oct 21 11:05:43 corosync [TOTEM ] Receive multicast socket recv buffer size
(262142 bytes).
Oct 21 11:05:43 corosync [TOTEM ] Transmit multicast socket send buffer size
(262142 bytes).
Oct 21 11:05:43 corosync [TOTEM ] The network interface [172.30.0.145] is
now up.
Oct 21 11:05:43 corosync [TOTEM ] Created or loaded sequence id
184.172.30.0.145 for this ring.
Oct 21 11:05:43 corosync [TOTEM ] entering GATHER state from 15.
Oct 21 11:05:43 corosync [SERV  ] *Service failed to load 'pacemaker'.*
Oct 21 11:05:43 corosync [SERV  ] Service initialized 'corosync extended
virtual synchrony service'
Oct 21 11:05:43 corosync [SERV  ] Service initialized 'corosync
configuration service'
Oct 21 11:05:43 corosync [SERV  ] Service initialized 'corosync cluster
closed process group service v1.01'
Oct 21 11:05:43 corosync [SERV  ] Service initialized 'corosync cluster
config database access v1.01'
Oct 21 11:05:43 corosync [SERV  ] Service initialized 'corosync profile
loading service'
Oct 21 11:05:43 corosync [MAIN  ] Compatibility mode set to whitetank.
 Using V1 and V2 of the synchronization engine.
Oct 21 11:05:43 corosync [TOTEM ] Creating commit token because I am the
rep.
Oct 21 11:05:43 corosync [TOTEM ] Saving state aru 0 high seq received 0
Oct 21 11:05:43 corosync [TOTEM ] Storing new sequence id for ring bc
Oct 21 11:05:43 corosync [TOTEM ] entering COMMIT state.
Oct 21 11:05:43 corosync [TOTEM ] got commit token
Oct 21 11:05:43 corosync [TOTEM ] entering RECOVERY state.
Oct 21 11:05:43 corosync [TOTEM ] position [0] member 172.30.0.145:
Oct 21 11:05:43 corosync [TOTEM ] previous ring seq 184 rep 172.30.0.145
Oct 21 11:05:43 corosync [TOTEM ] aru 0 high delivered 0 received flag 1
Oct 21 11:05:43 corosync [TOTEM ] Did not need to originate any messages in
recovery.
Oct 21 11:05:43 corosync [TOTEM ] got commit token
Oct 21 11:05:43 corosync [TOTEM ] Sending initial ORF token
Oct 21 11:05:43 corosync [TOTEM ] token retrans flag is 0 my set retrans
flag0 retrans queue empty 1 count 0, aru 0
Oct 21 11:05:43 corosync [TOTEM ] install seq 0 aru 0 high seq received 0
Oct 21 11:05:43 corosync [TOTEM ] token retrans flag is 0 my set retrans
flag0 retrans queue empty 1 count 1, aru 0
Oct 21 11:05:43 corosync [TOTEM ] install seq 0 aru 0 high seq received 0
Oct 21 11:05:43 corosync [TOTEM ] token retrans flag is 0 my set retrans
flag0 retrans queue empty 1 count 2, aru 0
Oct 21 11:05:43 

Re: [Pacemaker] pacemaker unable to start

2009-10-21 Thread Steven Dake
I recommend using corosync 1.1.1 - several bug fixes one critical for
proper pacemaker operation.  It won't fix this particular problem
however.

Corosync loads pacemaker by searching for a pacemaker lcrso file.  These
files are default installed in /usr/libexec/lcrso but may be in a
different location depending on your distribution.

Regards
-steve

On Wed, 2009-10-21 at 11:13 -0400, Shravan Mishra wrote:
 Hello guys,
 
 We are running 
 
 corosync-1.0.0
 heartbeat-2.99.1
 pacemaker-1.0.4
 
 the corosync.conf  under /etc/corosync/ is 
 
 
 # Please read the corosync.conf.5 manual page
 compatibility: whitetank
 
 aisexec {
user: root
group: root
 }
 totem {
version: 2
secauth: off
threads: 0
interface {
ringnumber: 0
bindnetaddr: 172.30.0.0
mcastaddr:226.94.1.1
mcastport: 5406
}
 }
 
 logging {
fileline: off
to_stderr: yes
to_logfile: yes
to_syslog: yes
logfile: /tmp/corosync.log
debug: on
timestamp: on
logger_subsys {
subsys: pacemaker
debug: on
tags: enter|leave|trace1|trace2| trace3|trace4|trace6
}
 }
 
 
 service {
name: pacemaker
ver: 0
 #   use_mgmtd: yes
  #  use_logd:yes
 }
 
 
 corosync {
user: root
group: root
 }
 
 
 amf {
mode: disabled
 }
 
 
 
 #service corosync start   
 
 starts the messaging but fails to load pacemaker,
 
 /tmp/corosync.log  ---   
 
 ==
 
 Oct 21 11:05:43 corosync [MAIN  ] Corosync Cluster Engine ('trunk'):
 started and ready to provide service.
 Oct 21 11:05:43 corosync [MAIN  ] Successfully read main configuration
 file '/etc/corosync/corosync.conf'.
 Oct 21 11:05:43 corosync [TOTEM ] Token Timeout (1000 ms) retransmit
 timeout (238 ms)
 Oct 21 11:05:43 corosync [TOTEM ] token hold (180 ms) retransmits
 before loss (4 retrans)
 Oct 21 11:05:43 corosync [TOTEM ] join (50 ms) send_join (0 ms)
 consensus (800 ms) merge (200 ms)
 Oct 21 11:05:43 corosync [TOTEM ] downcheck (1000 ms) fail to recv
 const (50 msgs)
 Oct 21 11:05:43 corosync [TOTEM ] seqno unchanged const (30 rotations)
 Maximum network MTU 1500
 Oct 21 11:05:43 corosync [TOTEM ] window size per rotation (50
 messages) maximum messages per rotation (17 messages)
 Oct 21 11:05:43 corosync [TOTEM ] send threads (0 threads)
 Oct 21 11:05:43 corosync [TOTEM ] RRP token expired timeout (238 ms)
 Oct 21 11:05:43 corosync [TOTEM ] RRP token problem counter (2000 ms)
 Oct 21 11:05:43 corosync [TOTEM ] RRP threshold (10 problem count)
 Oct 21 11:05:43 corosync [TOTEM ] RRP mode set to none.
 Oct 21 11:05:43 corosync [TOTEM ] heartbeat_failures_allowed (0)
 Oct 21 11:05:43 corosync [TOTEM ] max_network_delay (50 ms)
 Oct 21 11:05:43 corosync [TOTEM ] HeartBeat is Disabled. To enable set
 heartbeat_failures_allowed  0
 Oct 21 11:05:43 corosync [TOTEM ] Initializing transmit/receive
 security: libtomcrypt SOBER128/SHA1HMAC (mode 0).
 Oct 21 11:05:43 corosync [TOTEM ] Receive multicast socket recv buffer
 size (262142 bytes).
 Oct 21 11:05:43 corosync [TOTEM ] Transmit multicast socket send
 buffer size (262142 bytes).
 Oct 21 11:05:43 corosync [TOTEM ] The network interface [172.30.0.145]
 is now up.
 Oct 21 11:05:43 corosync [TOTEM ] Created or loaded sequence id
 184.172.30.0.145 for this ring.
 Oct 21 11:05:43 corosync [TOTEM ] entering GATHER state from 15.
 Oct 21 11:05:43 corosync [SERV  ] Service failed to load 'pacemaker'.
 Oct 21 11:05:43 corosync [SERV  ] Service initialized 'corosync
 extended virtual synchrony service'
 Oct 21 11:05:43 corosync [SERV  ] Service initialized 'corosync
 configuration service'
 Oct 21 11:05:43 corosync [SERV  ] Service initialized 'corosync
 cluster closed process group service v1.01'
 Oct 21 11:05:43 corosync [SERV  ] Service initialized 'corosync
 cluster config database access v1.01'
 Oct 21 11:05:43 corosync [SERV  ] Service initialized 'corosync
 profile loading service'
 Oct 21 11:05:43 corosync [MAIN  ] Compatibility mode set to
 whitetank.  Using V1 and V2 of the synchronization engine.
 Oct 21 11:05:43 corosync [TOTEM ] Creating commit token because I am
 the rep.
 Oct 21 11:05:43 corosync [TOTEM ] Saving state aru 0 high seq received
 0
 Oct 21 11:05:43 corosync [TOTEM ] Storing new sequence id for ring bc
 Oct 21 11:05:43 corosync [TOTEM ] entering COMMIT state.
 Oct 21 11:05:43 corosync [TOTEM ] got commit token
 Oct 21 11:05:43 corosync [TOTEM ] entering RECOVERY state.
 Oct 21 11:05:43 corosync [TOTEM ] position [0] member 172.30.0.145:
 Oct 21 11:05:43 corosync [TOTEM ] previous ring seq 184 rep
 172.30.0.145
 Oct 21 11:05:43 corosync [TOTEM ] aru 0 high delivered 0 received flag
 1
 Oct 21 11:05:43 corosync [TOTEM ] Did not need to originate any
 messages in recovery.
 Oct 21 11:05:43 corosync [TOTEM ] got commit token
 Oct 21 

Re: [Pacemaker] Why are fatal warnings enabled by default?

2009-10-21 Thread Manfred Hollstein
On Wed, 21 Oct 2009, 17:56:25 +0200, Florian Haas wrote:
 On 10/21/2009 03:17 PM, Dejan Muhamedagic wrote:
  This doesn't happen here with libnet-1.1.2.1-140.75.i586. Which
  libnet version do you have?
  libnet-1.1.4-3.el5
  
  API changed. I'll see how this can be detected at the configure
  stage.
 
 Now since this seems to be a libnet related problem, why does it also
 occur when I configure with --disable-libnet? By contrast it does go
 away when libnet is uninstalled. Is some autoconf magic broken there too?

I'm not sure if there are any real rules as to which packages (and which
versions) are required for successfull building. I remember similar issues
with GCC over 10 years ago, after which we decided to document the stuff
required for building it successfully. Do we have such rules documented
for Pacemaker?

Cheers.

l8er
manfred

___
Pacemaker mailing list
Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker


Re: [Pacemaker] pacemaker unable to start

2009-10-21 Thread Steven Dake
Ya your missing the pacemaker lcrso file.  Either you didn't build
pacemaker with corosync support or pacemaker didn't install that binary
in the proper place.

try:

updatedb
locate lcrso

Regards
-steve

On Wed, 2009-10-21 at 12:28 -0400, Shravan Mishra wrote:
 Steve, this is what my installation shows--
 
 ls -l /usr/libexec/lcrso
 
 -rwxr-xr-x  1 root root  101243 Jul 29 11:21 coroparse.lcrso
 -rwxr-xr-x  1 root root  117688 Jul 29 11:21 objdb.lcrso
 -rwxr-xr-x  1 root root   92702 Jul 29 11:54 openaisserviceenable.lcrso
 -rwxr-xr-x  1 root root  110808 Jul 29 11:21 quorum_testquorum.lcrso
 -rwxr-xr-x  1 root root  159057 Jul 29 11:21 quorum_votequorum.lcrso
 -rwxr-xr-x  1 root root 1175430 Jul 29 11:54 service_amf.lcrso
 -rwxr-xr-x  1 root root  133976 Jul 29 11:21 service_cfg.lcrso
 -rwxr-xr-x  1 root root  218374 Jul 29 11:54 service_ckpt.lcrso
 -rwxr-xr-x  1 root root  139029 Jul 29 11:54 service_clm.lcrso
 -rwxr-xr-x  1 root root  122668 Jul 29 11:21 service_confdb.lcrso
 -rwxr-xr-x  1 root root  138412 Jul 29 11:21 service_cpg.lcrso
 -rwxr-xr-x  1 root root  125638 Jul 29 11:21 service_evs.lcrso
 -rwxr-xr-x  1 root root  196443 Jul 29 11:54 service_evt.lcrso
 -rwxr-xr-x  1 root root  194885 Jul 29 11:54 service_lck.lcrso
 -rwxr-xr-x  1 root root  235168 Jul 29 11:54 service_msg.lcrso
 -rwxr-xr-x  1 root root  120445 Jul 29 11:21 service_pload.lcrso
 -rwxr-xr-x  1 root root  135340 Jul 29 11:54 service_tmr.lcrso
 -rwxr-xr-x  1 root root  124092 Jul 29 11:21 vsf_quorum.lcrso
 -rwxr-xr-x  1 root root  121298 Jul 29 11:21 vsf_ykd.lcrso
 
 I also did
 
 export COROSYNC_DEFAULT_CONFIG_IFACE=openaisserviceenable:openaisparser
 
 In place of openaisparser I also tried corosyncparse and
 corosyncparser but to no avail.
 
 -sincerely
 Shravan
 
 On Wed, Oct 21, 2009 at 11:49 AM, Steven Dake sd...@redhat.com wrote:
  I recommend using corosync 1.1.1 - several bug fixes one critical for
  proper pacemaker operation.  It won't fix this particular problem
  however.
 
  Corosync loads pacemaker by searching for a pacemaker lcrso file.  These
  files are default installed in /usr/libexec/lcrso but may be in a
  different location depending on your distribution.
 
  Regards
  -steve
 
  On Wed, 2009-10-21 at 11:13 -0400, Shravan Mishra wrote:
  Hello guys,
 
  We are running
 
  corosync-1.0.0
  heartbeat-2.99.1
  pacemaker-1.0.4
 
  the corosync.conf  under /etc/corosync/ is
 
  
  # Please read the corosync.conf.5 manual page
  compatibility: whitetank
 
  aisexec {
 user: root
 group: root
  }
  totem {
 version: 2
 secauth: off
 threads: 0
 interface {
 ringnumber: 0
 bindnetaddr: 172.30.0.0
 mcastaddr:226.94.1.1
 mcastport: 5406
 }
  }
 
  logging {
 fileline: off
 to_stderr: yes
 to_logfile: yes
 to_syslog: yes
 logfile: /tmp/corosync.log
 debug: on
 timestamp: on
 logger_subsys {
 subsys: pacemaker
 debug: on
 tags: enter|leave|trace1|trace2| trace3|trace4|trace6
 }
  }
 
 
  service {
 name: pacemaker
 ver: 0
  #   use_mgmtd: yes
   #  use_logd:yes
  }
 
 
  corosync {
 user: root
 group: root
  }
 
 
  amf {
 mode: disabled
  }
  
 
 
  #service corosync start
 
  starts the messaging but fails to load pacemaker,
 
  /tmp/corosync.log  ---
 
  ==
 
  Oct 21 11:05:43 corosync [MAIN  ] Corosync Cluster Engine ('trunk'):
  started and ready to provide service.
  Oct 21 11:05:43 corosync [MAIN  ] Successfully read main configuration
  file '/etc/corosync/corosync.conf'.
  Oct 21 11:05:43 corosync [TOTEM ] Token Timeout (1000 ms) retransmit
  timeout (238 ms)
  Oct 21 11:05:43 corosync [TOTEM ] token hold (180 ms) retransmits
  before loss (4 retrans)
  Oct 21 11:05:43 corosync [TOTEM ] join (50 ms) send_join (0 ms)
  consensus (800 ms) merge (200 ms)
  Oct 21 11:05:43 corosync [TOTEM ] downcheck (1000 ms) fail to recv
  const (50 msgs)
  Oct 21 11:05:43 corosync [TOTEM ] seqno unchanged const (30 rotations)
  Maximum network MTU 1500
  Oct 21 11:05:43 corosync [TOTEM ] window size per rotation (50
  messages) maximum messages per rotation (17 messages)
  Oct 21 11:05:43 corosync [TOTEM ] send threads (0 threads)
  Oct 21 11:05:43 corosync [TOTEM ] RRP token expired timeout (238 ms)
  Oct 21 11:05:43 corosync [TOTEM ] RRP token problem counter (2000 ms)
  Oct 21 11:05:43 corosync [TOTEM ] RRP threshold (10 problem count)
  Oct 21 11:05:43 corosync [TOTEM ] RRP mode set to none.
  Oct 21 11:05:43 corosync [TOTEM ] heartbeat_failures_allowed (0)
  Oct 21 11:05:43 corosync [TOTEM ] max_network_delay (50 ms)
  Oct 21 11:05:43 corosync [TOTEM ] HeartBeat is Disabled. To enable set
  heartbeat_failures_allowed  0
  Oct 21 11:05:43 corosync [TOTEM ] Initializing transmit/receive
  security: