hmm.. but as far as I can see, looks like that messages can still be
ignored. My original problem is that.. LVM resource agent doesn't try to
activate the VG on the passive node if the active node goes power off.

On Mon, Dec 29, 2014 at 5:33 PM, emmanuel segura <emi2f...@gmail.com> wrote:

> Hi,
>
> You have  a problem with the cluster stonithd:"error: crm_abort:
> crm_glib_handler: Forked child 6186 to record non-fatal assert at
> logging.c:73 "
>
> Try to post your cluster version(packages), maybe someone can tell you
> if this is a known bug or new.
>
>
>
> 2014-12-29 10:29 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
> > ok, sorry for that.. please use this instead.
> >
> > http://pastebin.centos.org/14771/
> >
> > thanks.
> >
> > On Mon, Dec 29, 2014 at 5:25 PM, emmanuel segura <emi2f...@gmail.com>
> wrote:
> >
> >> Sorry,
> >>
> >> But your paste is empty.
> >>
> >> 2014-12-29 10:19 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
> >> > hi,
> >> >
> >> > uploaded it here.
> >> >
> >> > http://susepaste.org/45413433
> >> >
> >> > thanks.
> >> >
> >> > On Mon, Dec 29, 2014 at 5:09 PM, Marlon Guao <marlon.g...@gmail.com>
> >> wrote:
> >> >
> >> >> Ok, i attached the log file of one of the nodes.
> >> >>
> >> >> On Mon, Dec 29, 2014 at 4:42 PM, emmanuel segura <emi2f...@gmail.com
> >
> >> >> wrote:
> >> >>
> >> >>> please use pastebin and show your whole logs
> >> >>>
> >> >>> 2014-12-29 9:06 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
> >> >>> > by the way.. just to note that.. for a normal testing (manual
> >> failover,
> >> >>> > rebooting the active node)... the cluster is working fine. I only
> >> >>> encounter
> >> >>> > this error if I try to poweroff/shutoff the active node.
> >> >>> >
> >> >>> > On Mon, Dec 29, 2014 at 4:05 PM, Marlon Guao <
> marlon.g...@gmail.com>
> >> >>> wrote:
> >> >>> >
> >> >>> >> Hi.
> >> >>> >>
> >> >>> >>
> >> >>> >> Dec 29 13:47:16 s1 LVM(vg1)[1601]: WARNING: LVM Volume cluvg1 is
> not
> >> >>> >> available (stopped)
> >> >>> >> Dec 29 13:47:16 s1 crmd[1515]:   notice: process_lrm_event:
> >> Operation
> >> >>> >> vg1_monitor_0: not running (node=
> >> >>> >> s1, call=23, rc=7, cib-update=40, confirmed=true)
> >> >>> >> Dec 29 13:47:16 s1 crmd[1515]:   notice: te_rsc_command:
> Initiating
> >> >>> action
> >> >>> >> 9: monitor fs1_monitor_0 on
> >> >>> >> s1 (local)
> >> >>> >> Dec 29 13:47:16 s1 crmd[1515]:   notice: te_rsc_command:
> Initiating
> >> >>> action
> >> >>> >> 16: monitor vg1_monitor_0 on
> >> >>> >>  s2
> >> >>> >> Dec 29 13:47:16 s1 Filesystem(fs1)[1618]: WARNING: Couldn't find
> >> device
> >> >>> >> [/dev/mapper/cluvg1-clulv1]. Ex
> >> >>> >> pected /dev/??? to exist
> >> >>> >>
> >> >>> >>
> >> >>> >> from the LVM agent, it checked if the volume is already
> available..
> >> and
> >> >>> >> will raise the above error if not. But, I don't see that it
> tries to
> >> >>> >> activate it before raising the VG. Perhaps, it assumes that the
> VG
> >> is
> >> >>> >> already activated... so, I'm not sure who should be activating it
> >> >>> (should
> >> >>> >> it be LVM?).
> >> >>> >>
> >> >>> >>
> >> >>> >>  if [ $rc -ne 0 ]; then
> >> >>> >>                 ocf_log $loglevel "LVM Volume $1 is not available
> >> >>> >> (stopped)"
> >> >>> >>                 rc=$OCF_NOT_RUNNING
> >> >>> >>         else
> >> >>> >>                 case $(get_vg_mode) in
> >> >>> >>                 1) # exclusive with tagging.
> >> >>> >>                         # If vg is running, make sure the correct
> >> tag
> >> >>> is
> >> >>> >> present. Otherwise we
> >> >>> >>                         # can not guarantee exclusive activation.
> >> >>> >>                         if ! check_tags; then
> >> >>> >>                                 ocf_exit_reason "WARNING:
> >> >>> >> $OCF_RESKEY_volgrpname is active without the cluster tag,
> >> \"$OUR_TAG\""
> >> >>> >>
> >> >>> >> On Mon, Dec 29, 2014 at 3:36 PM, emmanuel segura <
> >> emi2f...@gmail.com>
> >> >>> >> wrote:
> >> >>> >>
> >> >>> >>> logs?
> >> >>> >>>
> >> >>> >>> 2014-12-29 6:54 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
> >> >>> >>> > Hi,
> >> >>> >>> >
> >> >>> >>> > just want to ask regarding the LVM resource agent on
> >> >>> pacemaker/corosync.
> >> >>> >>> >
> >> >>> >>> > I setup 2 nodes cluster (opensuse13.2 -- my config below). The
> >> >>> cluster
> >> >>> >>> > works as expected, like doing a manual failover (via crm
> resource
> >> >>> move),
> >> >>> >>> > and automatic failover (by rebooting the active node for
> >> instance).
> >> >>> >>> But, if
> >> >>> >>> > i try to just "shutoff" the active node (it's a VM, so I can
> do a
> >> >>> >>> > poweroff). The resources won't be able to failover to the
> passive
> >> >>> node.
> >> >>> >>> > when I did an investigation, it's due to an LVM resource not
> >> >>> starting
> >> >>> >>> > (specifically, the VG). I found out that the LVM resource
> won't
> >> try
> >> >>> to
> >> >>> >>> > activate the volume group in the passive node. Is this an
> >> expected
> >> >>> >>> > behaviour?
> >> >>> >>> >
> >> >>> >>> > what I really expect is that, in the event that the active
> node
> >> be
> >> >>> >>> shutoff
> >> >>> >>> > (by a power outage for instance), all resources should be
> >> failover
> >> >>> >>> > automatically to the passive. LVM should re-activate the VG.
> >> >>> >>> >
> >> >>> >>> >
> >> >>> >>> > here's my config.
> >> >>> >>> >
> >> >>> >>> > node 1: s1
> >> >>> >>> > node 2: s2
> >> >>> >>> > primitive cluIP IPaddr2 \
> >> >>> >>> > params ip=192.168.13.200 cidr_netmask=32 \
> >> >>> >>> > op monitor interval=30s
> >> >>> >>> > primitive clvm ocf:lvm2:clvmd \
> >> >>> >>> > params daemon_timeout=30 \
> >> >>> >>> > op monitor timeout=90 interval=30
> >> >>> >>> > primitive dlm ocf:pacemaker:controld \
> >> >>> >>> > op monitor interval=60s timeout=90s on-fail=ignore \
> >> >>> >>> > op start interval=0 timeout=90
> >> >>> >>> > primitive fs1 Filesystem \
> >> >>> >>> > params device="/dev/mapper/cluvg1-clulv1" directory="/data"
> >> >>> fstype=btrfs
> >> >>> >>> > primitive mariadb mysql \
> >> >>> >>> > params config="/etc/my.cnf"
> >> >>> >>> > primitive sbd stonith:external/sbd \
> >> >>> >>> > op monitor interval=15s timeout=60s
> >> >>> >>> > primitive vg1 LVM \
> >> >>> >>> > params volgrpname=cluvg1 exclusive=yes \
> >> >>> >>> > op start timeout=10s interval=0 \
> >> >>> >>> > op stop interval=0 timeout=10 \
> >> >>> >>> > op monitor interval=10 timeout=30 on-fail=restart depth=0
> >> >>> >>> > group base-group dlm clvm
> >> >>> >>> > group rgroup cluIP vg1 fs1 mariadb \
> >> >>> >>> > meta target-role=Started
> >> >>> >>> > clone base-clone base-group \
> >> >>> >>> > meta interleave=true target-role=Started
> >> >>> >>> > property cib-bootstrap-options: \
> >> >>> >>> > dc-version=1.1.12-1.1.12.git20140904.266d5c2 \
> >> >>> >>> > cluster-infrastructure=corosync \
> >> >>> >>> > no-quorum-policy=ignore \
> >> >>> >>> > last-lrm-refresh=1419514875 \
> >> >>> >>> > cluster-name=xxx \
> >> >>> >>> > stonith-enabled=true
> >> >>> >>> > rsc_defaults rsc-options: \
> >> >>> >>> > resource-stickiness=100
> >> >>> >>> >
> >> >>> >>> > --
> >> >>> >>> >>>> import this
> >> >>> >>> > _______________________________________________
> >> >>> >>> > Linux-HA mailing list
> >> >>> >>> > Linux-HA@lists.linux-ha.org
> >> >>> >>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
> >> >>> >>> > See also: http://linux-ha.org/ReportingProblems
> >> >>> >>>
> >> >>> >>>
> >> >>> >>>
> >> >>> >>> --
> >> >>> >>> esta es mi vida e me la vivo hasta que dios quiera
> >> >>> >>> _______________________________________________
> >> >>> >>> Linux-HA mailing list
> >> >>> >>> Linux-HA@lists.linux-ha.org
> >> >>> >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> >> >>> >>> See also: http://linux-ha.org/ReportingProblems
> >> >>> >>>
> >> >>> >>
> >> >>> >>
> >> >>> >>
> >> >>> >> --
> >> >>> >> >>> import this
> >> >>> >>
> >> >>> >
> >> >>> >
> >> >>> >
> >> >>> > --
> >> >>> >>>> import this
> >> >>> > _______________________________________________
> >> >>> > Linux-HA mailing list
> >> >>> > Linux-HA@lists.linux-ha.org
> >> >>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
> >> >>> > See also: http://linux-ha.org/ReportingProblems
> >> >>>
> >> >>>
> >> >>>
> >> >>> --
> >> >>> esta es mi vida e me la vivo hasta que dios quiera
> >> >>> _______________________________________________
> >> >>> Linux-HA mailing list
> >> >>> Linux-HA@lists.linux-ha.org
> >> >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> >> >>> See also: http://linux-ha.org/ReportingProblems
> >> >>>
> >> >>
> >> >>
> >> >>
> >> >> --
> >> >> >>> import this
> >> >>
> >> >
> >> >
> >> >
> >> > --
> >> >>>> import this
> >> > _______________________________________________
> >> > Linux-HA mailing list
> >> > Linux-HA@lists.linux-ha.org
> >> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
> >> > See also: http://linux-ha.org/ReportingProblems
> >>
> >>
> >>
> >> --
> >> esta es mi vida e me la vivo hasta que dios quiera
> >> _______________________________________________
> >> Linux-HA mailing list
> >> Linux-HA@lists.linux-ha.org
> >> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> >> See also: http://linux-ha.org/ReportingProblems
> >>
> >
> >
> >
> > --
> >>>> import this
> > _______________________________________________
> > Linux-HA mailing list
> > Linux-HA@lists.linux-ha.org
> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
> > See also: http://linux-ha.org/ReportingProblems
>
>
>
> --
> esta es mi vida e me la vivo hasta que dios quiera
> _______________________________________________
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
>



-- 
>>> import this
_______________________________________________
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to