Hi, You have a problem with the cluster stonithd:"error: crm_abort: crm_glib_handler: Forked child 6186 to record non-fatal assert at logging.c:73 "
Try to post your cluster version(packages), maybe someone can tell you if this is a known bug or new. 2014-12-29 10:29 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>: > ok, sorry for that.. please use this instead. > > http://pastebin.centos.org/14771/ > > thanks. > > On Mon, Dec 29, 2014 at 5:25 PM, emmanuel segura <emi2f...@gmail.com> wrote: > >> Sorry, >> >> But your paste is empty. >> >> 2014-12-29 10:19 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>: >> > hi, >> > >> > uploaded it here. >> > >> > http://susepaste.org/45413433 >> > >> > thanks. >> > >> > On Mon, Dec 29, 2014 at 5:09 PM, Marlon Guao <marlon.g...@gmail.com> >> wrote: >> > >> >> Ok, i attached the log file of one of the nodes. >> >> >> >> On Mon, Dec 29, 2014 at 4:42 PM, emmanuel segura <emi2f...@gmail.com> >> >> wrote: >> >> >> >>> please use pastebin and show your whole logs >> >>> >> >>> 2014-12-29 9:06 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>: >> >>> > by the way.. just to note that.. for a normal testing (manual >> failover, >> >>> > rebooting the active node)... the cluster is working fine. I only >> >>> encounter >> >>> > this error if I try to poweroff/shutoff the active node. >> >>> > >> >>> > On Mon, Dec 29, 2014 at 4:05 PM, Marlon Guao <marlon.g...@gmail.com> >> >>> wrote: >> >>> > >> >>> >> Hi. >> >>> >> >> >>> >> >> >>> >> Dec 29 13:47:16 s1 LVM(vg1)[1601]: WARNING: LVM Volume cluvg1 is not >> >>> >> available (stopped) >> >>> >> Dec 29 13:47:16 s1 crmd[1515]: notice: process_lrm_event: >> Operation >> >>> >> vg1_monitor_0: not running (node= >> >>> >> s1, call=23, rc=7, cib-update=40, confirmed=true) >> >>> >> Dec 29 13:47:16 s1 crmd[1515]: notice: te_rsc_command: Initiating >> >>> action >> >>> >> 9: monitor fs1_monitor_0 on >> >>> >> s1 (local) >> >>> >> Dec 29 13:47:16 s1 crmd[1515]: notice: te_rsc_command: Initiating >> >>> action >> >>> >> 16: monitor vg1_monitor_0 on >> >>> >> s2 >> >>> >> Dec 29 13:47:16 s1 Filesystem(fs1)[1618]: WARNING: Couldn't find >> device >> >>> >> [/dev/mapper/cluvg1-clulv1]. Ex >> >>> >> pected /dev/??? to exist >> >>> >> >> >>> >> >> >>> >> from the LVM agent, it checked if the volume is already available.. >> and >> >>> >> will raise the above error if not. But, I don't see that it tries to >> >>> >> activate it before raising the VG. Perhaps, it assumes that the VG >> is >> >>> >> already activated... so, I'm not sure who should be activating it >> >>> (should >> >>> >> it be LVM?). >> >>> >> >> >>> >> >> >>> >> if [ $rc -ne 0 ]; then >> >>> >> ocf_log $loglevel "LVM Volume $1 is not available >> >>> >> (stopped)" >> >>> >> rc=$OCF_NOT_RUNNING >> >>> >> else >> >>> >> case $(get_vg_mode) in >> >>> >> 1) # exclusive with tagging. >> >>> >> # If vg is running, make sure the correct >> tag >> >>> is >> >>> >> present. Otherwise we >> >>> >> # can not guarantee exclusive activation. >> >>> >> if ! check_tags; then >> >>> >> ocf_exit_reason "WARNING: >> >>> >> $OCF_RESKEY_volgrpname is active without the cluster tag, >> \"$OUR_TAG\"" >> >>> >> >> >>> >> On Mon, Dec 29, 2014 at 3:36 PM, emmanuel segura < >> emi2f...@gmail.com> >> >>> >> wrote: >> >>> >> >> >>> >>> logs? >> >>> >>> >> >>> >>> 2014-12-29 6:54 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>: >> >>> >>> > Hi, >> >>> >>> > >> >>> >>> > just want to ask regarding the LVM resource agent on >> >>> pacemaker/corosync. >> >>> >>> > >> >>> >>> > I setup 2 nodes cluster (opensuse13.2 -- my config below). The >> >>> cluster >> >>> >>> > works as expected, like doing a manual failover (via crm resource >> >>> move), >> >>> >>> > and automatic failover (by rebooting the active node for >> instance). >> >>> >>> But, if >> >>> >>> > i try to just "shutoff" the active node (it's a VM, so I can do a >> >>> >>> > poweroff). The resources won't be able to failover to the passive >> >>> node. >> >>> >>> > when I did an investigation, it's due to an LVM resource not >> >>> starting >> >>> >>> > (specifically, the VG). I found out that the LVM resource won't >> try >> >>> to >> >>> >>> > activate the volume group in the passive node. Is this an >> expected >> >>> >>> > behaviour? >> >>> >>> > >> >>> >>> > what I really expect is that, in the event that the active node >> be >> >>> >>> shutoff >> >>> >>> > (by a power outage for instance), all resources should be >> failover >> >>> >>> > automatically to the passive. LVM should re-activate the VG. >> >>> >>> > >> >>> >>> > >> >>> >>> > here's my config. >> >>> >>> > >> >>> >>> > node 1: s1 >> >>> >>> > node 2: s2 >> >>> >>> > primitive cluIP IPaddr2 \ >> >>> >>> > params ip=192.168.13.200 cidr_netmask=32 \ >> >>> >>> > op monitor interval=30s >> >>> >>> > primitive clvm ocf:lvm2:clvmd \ >> >>> >>> > params daemon_timeout=30 \ >> >>> >>> > op monitor timeout=90 interval=30 >> >>> >>> > primitive dlm ocf:pacemaker:controld \ >> >>> >>> > op monitor interval=60s timeout=90s on-fail=ignore \ >> >>> >>> > op start interval=0 timeout=90 >> >>> >>> > primitive fs1 Filesystem \ >> >>> >>> > params device="/dev/mapper/cluvg1-clulv1" directory="/data" >> >>> fstype=btrfs >> >>> >>> > primitive mariadb mysql \ >> >>> >>> > params config="/etc/my.cnf" >> >>> >>> > primitive sbd stonith:external/sbd \ >> >>> >>> > op monitor interval=15s timeout=60s >> >>> >>> > primitive vg1 LVM \ >> >>> >>> > params volgrpname=cluvg1 exclusive=yes \ >> >>> >>> > op start timeout=10s interval=0 \ >> >>> >>> > op stop interval=0 timeout=10 \ >> >>> >>> > op monitor interval=10 timeout=30 on-fail=restart depth=0 >> >>> >>> > group base-group dlm clvm >> >>> >>> > group rgroup cluIP vg1 fs1 mariadb \ >> >>> >>> > meta target-role=Started >> >>> >>> > clone base-clone base-group \ >> >>> >>> > meta interleave=true target-role=Started >> >>> >>> > property cib-bootstrap-options: \ >> >>> >>> > dc-version=1.1.12-1.1.12.git20140904.266d5c2 \ >> >>> >>> > cluster-infrastructure=corosync \ >> >>> >>> > no-quorum-policy=ignore \ >> >>> >>> > last-lrm-refresh=1419514875 \ >> >>> >>> > cluster-name=xxx \ >> >>> >>> > stonith-enabled=true >> >>> >>> > rsc_defaults rsc-options: \ >> >>> >>> > resource-stickiness=100 >> >>> >>> > >> >>> >>> > -- >> >>> >>> >>>> import this >> >>> >>> > _______________________________________________ >> >>> >>> > Linux-HA mailing list >> >>> >>> > Linux-HA@lists.linux-ha.org >> >>> >>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha >> >>> >>> > See also: http://linux-ha.org/ReportingProblems >> >>> >>> >> >>> >>> >> >>> >>> >> >>> >>> -- >> >>> >>> esta es mi vida e me la vivo hasta que dios quiera >> >>> >>> _______________________________________________ >> >>> >>> Linux-HA mailing list >> >>> >>> Linux-HA@lists.linux-ha.org >> >>> >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha >> >>> >>> See also: http://linux-ha.org/ReportingProblems >> >>> >>> >> >>> >> >> >>> >> >> >>> >> >> >>> >> -- >> >>> >> >>> import this >> >>> >> >> >>> > >> >>> > >> >>> > >> >>> > -- >> >>> >>>> import this >> >>> > _______________________________________________ >> >>> > Linux-HA mailing list >> >>> > Linux-HA@lists.linux-ha.org >> >>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha >> >>> > See also: http://linux-ha.org/ReportingProblems >> >>> >> >>> >> >>> >> >>> -- >> >>> esta es mi vida e me la vivo hasta que dios quiera >> >>> _______________________________________________ >> >>> Linux-HA mailing list >> >>> Linux-HA@lists.linux-ha.org >> >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha >> >>> See also: http://linux-ha.org/ReportingProblems >> >>> >> >> >> >> >> >> >> >> -- >> >> >>> import this >> >> >> > >> > >> > >> > -- >> >>>> import this >> > _______________________________________________ >> > Linux-HA mailing list >> > Linux-HA@lists.linux-ha.org >> > http://lists.linux-ha.org/mailman/listinfo/linux-ha >> > See also: http://linux-ha.org/ReportingProblems >> >> >> >> -- >> esta es mi vida e me la vivo hasta que dios quiera >> _______________________________________________ >> Linux-HA mailing list >> Linux-HA@lists.linux-ha.org >> http://lists.linux-ha.org/mailman/listinfo/linux-ha >> See also: http://linux-ha.org/ReportingProblems >> > > > > -- >>>> import this > _______________________________________________ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems -- esta es mi vida e me la vivo hasta que dios quiera _______________________________________________ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems