you have quorum-policy=ignore, in the thread you posted:"
Nov 24 09:52:10 nebula3 dlm_controld[6263]: 566 datastores wait for fencing
Nov 24 09:52:10 nebula3 dlm_controld[6263]: 566 clvmd wait for fencing
Nov 24 09:55:10 nebula3 dlm_controld[6263]: 747 fence status
1084811078 receive -125 from 1084811079
walltime 1416819310 local 747"

{lvm}->{clvmd}->{dlm}->{fencing} = if fencing isn't working :) your
cluster will be broken.

2014-12-29 15:46 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
> looks like it's similar to this as well.
>
> http://comments.gmane.org/gmane.linux.highavailability.pacemaker/22398
>
> but, could it be because, clvm is not activating the vg on the passive
> node, because it's waiting for quorum?
>
> seeing this on the log as well.
>
> Dec 29 21:18:09 s2 dlm_controld[1776]: 8544 fence work wait for quorum
> Dec 29 21:18:12 s2 dlm_controld[1776]: 8547 clvmd wait for quorum
>
>
>
> On Mon, Dec 29, 2014 at 9:24 PM, Marlon Guao <marlon.g...@gmail.com> wrote:
>
>> interesting, i'm using the newer pacemaker version..
>>
>> pacemaker-1.1.12.git20140904.266d5c2-1.5.x86_64
>>
>>
>> On Mon, Dec 29, 2014 at 8:11 PM, emmanuel segura <emi2f...@gmail.com>
>> wrote:
>>
>>> https://bugzilla.redhat.com/show_bug.cgi?id=1127289#c4
>>> https://bugzilla.redhat.com/show_bug.cgi?id=1127289
>>>
>>> 2014-12-29 11:57 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
>>> > here it is..
>>> >
>>> >
>>> > ==Dumping header on disk /dev/mapper/sbd
>>> > Header version     : 2.1
>>> > UUID               : 36074673-f48e-4da2-b4ee-385e83e6abcc
>>> > Number of slots    : 255
>>> > Sector size        : 512
>>> > Timeout (watchdog) : 5
>>> > Timeout (allocate) : 2
>>> > Timeout (loop)     : 1
>>> > Timeout (msgwait)  : 10
>>> >
>>> > On Mon, Dec 29, 2014 at 6:42 PM, emmanuel segura <emi2f...@gmail.com>
>>> wrote:
>>> >
>>> >> Dlm isn't the problem, but i think is your fencing, when you powered
>>> >> off the active node, the dead remain in unclean state? can you show me
>>> >> your sbd timeouts? sbd -d /dev/path_of_your_device dump
>>> >>
>>> >> Thanks
>>> >>
>>> >> 2014-12-29 11:02 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
>>> >> > Hi,
>>> >> >
>>> >> > ah yeah.. tried to poweroff the active node.. and tried pvscan on the
>>> >> > passive.. and yes.. it didn't worked --- it doesn't return to the
>>> shell.
>>> >> > So, the problem is on DLM?
>>> >> >
>>> >> > On Mon, Dec 29, 2014 at 5:51 PM, emmanuel segura <emi2f...@gmail.com
>>> >
>>> >> wrote:
>>> >> >
>>> >> >> Power off the active node and after one seconde try to use one lvm
>>> >> >> command, for example pvscan, if this command doesn't response is
>>> >> >> because dlm relay on cluster fencing, if the cluster fencing doesn't
>>> >> >> work the dlm state in blocked state.
>>> >> >>
>>> >> >> 2014-12-29 10:43 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
>>> >> >> > perhaps, we need to focus on this message. as mentioned.. the
>>> cluster
>>> >> is
>>> >> >> > working fine under normal circumstances. my only concern is that,
>>> LVM
>>> >> >> > resource agent doesn't try to re-activate the VG on the passive
>>> node
>>> >> when
>>> >> >> > the active node goes down ungracefully (powered off). Hence, it
>>> could
>>> >> not
>>> >> >> > mount the filesystems.. etc.
>>> >> >> >
>>> >> >> >
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: process_lrm_event:
>>> Operation
>>> >> >> > sbd_monitor_0: not running (node=
>>> >> >> > s1, call=5, rc=7, cib-update=35, confirmed=true)
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: te_rsc_command:
>>> Initiating
>>> >> >> action
>>> >> >> > 13: monitor dlm:0_monitor_0
>>> >> >> > on s2
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: te_rsc_command:
>>> Initiating
>>> >> >> action
>>> >> >> > 5: monitor dlm:1_monitor_0 o
>>> >> >> > n s1 (local)
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: process_lrm_event:
>>> Operation
>>> >> >> > dlm_monitor_0: not running (node=
>>> >> >> > s1, call=10, rc=7, cib-update=36, confirmed=true)
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: te_rsc_command:
>>> Initiating
>>> >> >> action
>>> >> >> > 14: monitor clvm:0_monitor_0
>>> >> >> >  on s2
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: te_rsc_command:
>>> Initiating
>>> >> >> action
>>> >> >> > 6: monitor clvm:1_monitor_0
>>> >> >> > on s1 (local)
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: process_lrm_event:
>>> Operation
>>> >> >> > clvm_monitor_0: not running (node
>>> >> >> > =s1, call=15, rc=7, cib-update=37, confirmed=true)
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: te_rsc_command:
>>> Initiating
>>> >> >> action
>>> >> >> > 15: monitor cluIP_monitor_0
>>> >> >> > on s2
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: te_rsc_command:
>>> Initiating
>>> >> >> action
>>> >> >> > 7: monitor cluIP_monitor_0 o
>>> >> >> > n s1 (local)
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: process_lrm_event:
>>> Operation
>>> >> >> > cluIP_monitor_0: not running (nod
>>> >> >> > e=s1, call=19, rc=7, cib-update=38, confirmed=true)
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: te_rsc_command:
>>> Initiating
>>> >> >> action
>>> >> >> > 16: monitor vg1_monitor_0 on
>>> >> >> >  s2
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: te_rsc_command:
>>> Initiating
>>> >> >> action
>>> >> >> > 8: monitor vg1_monitor_0 on
>>> >> >> > s1 (local)
>>> >> >> > Dec 29 17:12:26 s1 LVM(vg1)[1583]: WARNING: LVM Volume cluvg1 is
>>> not
>>> >> >> > available (stopped)
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: process_lrm_event:
>>> Operation
>>> >> >> > vg1_monitor_0: not running (node=
>>> >> >> > s1, call=23, rc=7, cib-update=39, confirmed=true)
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: te_rsc_command:
>>> Initiating
>>> >> >> action
>>> >> >> > 17: monitor fs1_monitor_0 on
>>> >> >> >  s2
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: te_rsc_command:
>>> Initiating
>>> >> >> action
>>> >> >> > 9: monitor fs1_monitor_0 on
>>> >> >> > s1 (local)
>>> >> >> > Dec 29 17:12:26 s1 Filesystem(fs1)[1600]: WARNING: Couldn't find
>>> >> device
>>> >> >> > [/dev/mapper/cluvg1-clulv1]. Ex
>>> >> >> > pected /dev/??? to exist
>>> >> >> > Dec 29 17:12:26 s1 crmd[1495]:   notice: process_lrm_event:
>>> Operation
>>> >> >> > fs1_monitor_0: not running (node=
>>> >> >> > s1, call=27, rc=7, cib-update=40, confirmed=true)
>>> >> >> >
>>> >> >> > On Mon, Dec 29, 2014 at 5:38 PM, emmanuel segura <
>>> emi2f...@gmail.com>
>>> >> >> wrote:
>>> >> >> >
>>> >> >> >> Dec 27 15:38:00 s1 cib[1514]:    error: crm_xml_err: XML Error:
>>> >> >> >> Permission deniedPermission deniedI/O warning : failed to load
>>> >> >> >> external entity "/var/lib/pacemaker/cib/cib.xml"
>>> >> >> >> Dec 27 15:38:00 s1 cib[1514]:    error: write_cib_contents:
>>> Cannot
>>> >> >> >> link /var/lib/pacemaker/cib/cib.xml to
>>> >> >> >> /var/lib/pacemaker/cib/cib-0.raw: Operation not permitted (1)
>>> >> >> >>
>>> >> >> >> 2014-12-29 10:33 GMT+01:00 emmanuel segura <emi2f...@gmail.com>:
>>> >> >> >> > Hi,
>>> >> >> >> >
>>> >> >> >> > You have  a problem with the cluster stonithd:"error:
>>> crm_abort:
>>> >> >> >> > crm_glib_handler: Forked child 6186 to record non-fatal assert
>>> at
>>> >> >> >> > logging.c:73 "
>>> >> >> >> >
>>> >> >> >> > Try to post your cluster version(packages), maybe someone can
>>> tell
>>> >> you
>>> >> >> >> > if this is a known bug or new.
>>> >> >> >> >
>>> >> >> >> >
>>> >> >> >> >
>>> >> >> >> > 2014-12-29 10:29 GMT+01:00 Marlon Guao <marlon.g...@gmail.com
>>> >:
>>> >> >> >> >> ok, sorry for that.. please use this instead.
>>> >> >> >> >>
>>> >> >> >> >> http://pastebin.centos.org/14771/
>>> >> >> >> >>
>>> >> >> >> >> thanks.
>>> >> >> >> >>
>>> >> >> >> >> On Mon, Dec 29, 2014 at 5:25 PM, emmanuel segura <
>>> >> emi2f...@gmail.com
>>> >> >> >
>>> >> >> >> wrote:
>>> >> >> >> >>
>>> >> >> >> >>> Sorry,
>>> >> >> >> >>>
>>> >> >> >> >>> But your paste is empty.
>>> >> >> >> >>>
>>> >> >> >> >>> 2014-12-29 10:19 GMT+01:00 Marlon Guao <
>>> marlon.g...@gmail.com>:
>>> >> >> >> >>> > hi,
>>> >> >> >> >>> >
>>> >> >> >> >>> > uploaded it here.
>>> >> >> >> >>> >
>>> >> >> >> >>> > http://susepaste.org/45413433
>>> >> >> >> >>> >
>>> >> >> >> >>> > thanks.
>>> >> >> >> >>> >
>>> >> >> >> >>> > On Mon, Dec 29, 2014 at 5:09 PM, Marlon Guao <
>>> >> >> marlon.g...@gmail.com>
>>> >> >> >> >>> wrote:
>>> >> >> >> >>> >
>>> >> >> >> >>> >> Ok, i attached the log file of one of the nodes.
>>> >> >> >> >>> >>
>>> >> >> >> >>> >> On Mon, Dec 29, 2014 at 4:42 PM, emmanuel segura <
>>> >> >> >> emi2f...@gmail.com>
>>> >> >> >> >>> >> wrote:
>>> >> >> >> >>> >>
>>> >> >> >> >>> >>> please use pastebin and show your whole logs
>>> >> >> >> >>> >>>
>>> >> >> >> >>> >>> 2014-12-29 9:06 GMT+01:00 Marlon Guao <
>>> marlon.g...@gmail.com
>>> >> >:
>>> >> >> >> >>> >>> > by the way.. just to note that.. for a normal testing
>>> >> (manual
>>> >> >> >> >>> failover,
>>> >> >> >> >>> >>> > rebooting the active node)... the cluster is working
>>> fine.
>>> >> I
>>> >> >> only
>>> >> >> >> >>> >>> encounter
>>> >> >> >> >>> >>> > this error if I try to poweroff/shutoff the active
>>> node.
>>> >> >> >> >>> >>> >
>>> >> >> >> >>> >>> > On Mon, Dec 29, 2014 at 4:05 PM, Marlon Guao <
>>> >> >> >> marlon.g...@gmail.com>
>>> >> >> >> >>> >>> wrote:
>>> >> >> >> >>> >>> >
>>> >> >> >> >>> >>> >> Hi.
>>> >> >> >> >>> >>> >>
>>> >> >> >> >>> >>> >>
>>> >> >> >> >>> >>> >> Dec 29 13:47:16 s1 LVM(vg1)[1601]: WARNING: LVM Volume
>>> >> cluvg1
>>> >> >> >> is not
>>> >> >> >> >>> >>> >> available (stopped)
>>> >> >> >> >>> >>> >> Dec 29 13:47:16 s1 crmd[1515]:   notice:
>>> >> process_lrm_event:
>>> >> >> >> >>> Operation
>>> >> >> >> >>> >>> >> vg1_monitor_0: not running (node=
>>> >> >> >> >>> >>> >> s1, call=23, rc=7, cib-update=40, confirmed=true)
>>> >> >> >> >>> >>> >> Dec 29 13:47:16 s1 crmd[1515]:   notice:
>>> te_rsc_command:
>>> >> >> >> Initiating
>>> >> >> >> >>> >>> action
>>> >> >> >> >>> >>> >> 9: monitor fs1_monitor_0 on
>>> >> >> >> >>> >>> >> s1 (local)
>>> >> >> >> >>> >>> >> Dec 29 13:47:16 s1 crmd[1515]:   notice:
>>> te_rsc_command:
>>> >> >> >> Initiating
>>> >> >> >> >>> >>> action
>>> >> >> >> >>> >>> >> 16: monitor vg1_monitor_0 on
>>> >> >> >> >>> >>> >>  s2
>>> >> >> >> >>> >>> >> Dec 29 13:47:16 s1 Filesystem(fs1)[1618]: WARNING:
>>> >> Couldn't
>>> >> >> find
>>> >> >> >> >>> device
>>> >> >> >> >>> >>> >> [/dev/mapper/cluvg1-clulv1]. Ex
>>> >> >> >> >>> >>> >> pected /dev/??? to exist
>>> >> >> >> >>> >>> >>
>>> >> >> >> >>> >>> >>
>>> >> >> >> >>> >>> >> from the LVM agent, it checked if the volume is
>>> already
>>> >> >> >> available..
>>> >> >> >> >>> and
>>> >> >> >> >>> >>> >> will raise the above error if not. But, I don't see
>>> that
>>> >> it
>>> >> >> >> tries to
>>> >> >> >> >>> >>> >> activate it before raising the VG. Perhaps, it assumes
>>> >> that
>>> >> >> the
>>> >> >> >> VG
>>> >> >> >> >>> is
>>> >> >> >> >>> >>> >> already activated... so, I'm not sure who should be
>>> >> >> activating
>>> >> >> >> it
>>> >> >> >> >>> >>> (should
>>> >> >> >> >>> >>> >> it be LVM?).
>>> >> >> >> >>> >>> >>
>>> >> >> >> >>> >>> >>
>>> >> >> >> >>> >>> >>  if [ $rc -ne 0 ]; then
>>> >> >> >> >>> >>> >>                 ocf_log $loglevel "LVM Volume $1 is
>>> not
>>> >> >> >> available
>>> >> >> >> >>> >>> >> (stopped)"
>>> >> >> >> >>> >>> >>                 rc=$OCF_NOT_RUNNING
>>> >> >> >> >>> >>> >>         else
>>> >> >> >> >>> >>> >>                 case $(get_vg_mode) in
>>> >> >> >> >>> >>> >>                 1) # exclusive with tagging.
>>> >> >> >> >>> >>> >>                         # If vg is running, make sure
>>> the
>>> >> >> >> correct
>>> >> >> >> >>> tag
>>> >> >> >> >>> >>> is
>>> >> >> >> >>> >>> >> present. Otherwise we
>>> >> >> >> >>> >>> >>                         # can not guarantee exclusive
>>> >> >> >> activation.
>>> >> >> >> >>> >>> >>                         if ! check_tags; then
>>> >> >> >> >>> >>> >>                                 ocf_exit_reason
>>> "WARNING:
>>> >> >> >> >>> >>> >> $OCF_RESKEY_volgrpname is active without the cluster
>>> tag,
>>> >> >> >> >>> \"$OUR_TAG\""
>>> >> >> >> >>> >>> >>
>>> >> >> >> >>> >>> >> On Mon, Dec 29, 2014 at 3:36 PM, emmanuel segura <
>>> >> >> >> >>> emi2f...@gmail.com>
>>> >> >> >> >>> >>> >> wrote:
>>> >> >> >> >>> >>> >>
>>> >> >> >> >>> >>> >>> logs?
>>> >> >> >> >>> >>> >>>
>>> >> >> >> >>> >>> >>> 2014-12-29 6:54 GMT+01:00 Marlon Guao <
>>> >> >> marlon.g...@gmail.com>:
>>> >> >> >> >>> >>> >>> > Hi,
>>> >> >> >> >>> >>> >>> >
>>> >> >> >> >>> >>> >>> > just want to ask regarding the LVM resource agent
>>> on
>>> >> >> >> >>> >>> pacemaker/corosync.
>>> >> >> >> >>> >>> >>> >
>>> >> >> >> >>> >>> >>> > I setup 2 nodes cluster (opensuse13.2 -- my config
>>> >> below).
>>> >> >> >> The
>>> >> >> >> >>> >>> cluster
>>> >> >> >> >>> >>> >>> > works as expected, like doing a manual failover
>>> (via
>>> >> crm
>>> >> >> >> resource
>>> >> >> >> >>> >>> move),
>>> >> >> >> >>> >>> >>> > and automatic failover (by rebooting the active
>>> node
>>> >> for
>>> >> >> >> >>> instance).
>>> >> >> >> >>> >>> >>> But, if
>>> >> >> >> >>> >>> >>> > i try to just "shutoff" the active node (it's a
>>> VM, so
>>> >> I
>>> >> >> can
>>> >> >> >> do a
>>> >> >> >> >>> >>> >>> > poweroff). The resources won't be able to failover
>>> to
>>> >> the
>>> >> >> >> passive
>>> >> >> >> >>> >>> node.
>>> >> >> >> >>> >>> >>> > when I did an investigation, it's due to an LVM
>>> >> resource
>>> >> >> not
>>> >> >> >> >>> >>> starting
>>> >> >> >> >>> >>> >>> > (specifically, the VG). I found out that the LVM
>>> >> resource
>>> >> >> >> won't
>>> >> >> >> >>> try
>>> >> >> >> >>> >>> to
>>> >> >> >> >>> >>> >>> > activate the volume group in the passive node. Is
>>> this
>>> >> an
>>> >> >> >> >>> expected
>>> >> >> >> >>> >>> >>> > behaviour?
>>> >> >> >> >>> >>> >>> >
>>> >> >> >> >>> >>> >>> > what I really expect is that, in the event that the
>>> >> active
>>> >> >> >> node
>>> >> >> >> >>> be
>>> >> >> >> >>> >>> >>> shutoff
>>> >> >> >> >>> >>> >>> > (by a power outage for instance), all resources
>>> should
>>> >> be
>>> >> >> >> >>> failover
>>> >> >> >> >>> >>> >>> > automatically to the passive. LVM should
>>> re-activate
>>> >> the
>>> >> >> VG.
>>> >> >> >> >>> >>> >>> >
>>> >> >> >> >>> >>> >>> >
>>> >> >> >> >>> >>> >>> > here's my config.
>>> >> >> >> >>> >>> >>> >
>>> >> >> >> >>> >>> >>> > node 1: s1
>>> >> >> >> >>> >>> >>> > node 2: s2
>>> >> >> >> >>> >>> >>> > primitive cluIP IPaddr2 \
>>> >> >> >> >>> >>> >>> > params ip=192.168.13.200 cidr_netmask=32 \
>>> >> >> >> >>> >>> >>> > op monitor interval=30s
>>> >> >> >> >>> >>> >>> > primitive clvm ocf:lvm2:clvmd \
>>> >> >> >> >>> >>> >>> > params daemon_timeout=30 \
>>> >> >> >> >>> >>> >>> > op monitor timeout=90 interval=30
>>> >> >> >> >>> >>> >>> > primitive dlm ocf:pacemaker:controld \
>>> >> >> >> >>> >>> >>> > op monitor interval=60s timeout=90s on-fail=ignore
>>> \
>>> >> >> >> >>> >>> >>> > op start interval=0 timeout=90
>>> >> >> >> >>> >>> >>> > primitive fs1 Filesystem \
>>> >> >> >> >>> >>> >>> > params device="/dev/mapper/cluvg1-clulv1"
>>> >> >> directory="/data"
>>> >> >> >> >>> >>> fstype=btrfs
>>> >> >> >> >>> >>> >>> > primitive mariadb mysql \
>>> >> >> >> >>> >>> >>> > params config="/etc/my.cnf"
>>> >> >> >> >>> >>> >>> > primitive sbd stonith:external/sbd \
>>> >> >> >> >>> >>> >>> > op monitor interval=15s timeout=60s
>>> >> >> >> >>> >>> >>> > primitive vg1 LVM \
>>> >> >> >> >>> >>> >>> > params volgrpname=cluvg1 exclusive=yes \
>>> >> >> >> >>> >>> >>> > op start timeout=10s interval=0 \
>>> >> >> >> >>> >>> >>> > op stop interval=0 timeout=10 \
>>> >> >> >> >>> >>> >>> > op monitor interval=10 timeout=30 on-fail=restart
>>> >> depth=0
>>> >> >> >> >>> >>> >>> > group base-group dlm clvm
>>> >> >> >> >>> >>> >>> > group rgroup cluIP vg1 fs1 mariadb \
>>> >> >> >> >>> >>> >>> > meta target-role=Started
>>> >> >> >> >>> >>> >>> > clone base-clone base-group \
>>> >> >> >> >>> >>> >>> > meta interleave=true target-role=Started
>>> >> >> >> >>> >>> >>> > property cib-bootstrap-options: \
>>> >> >> >> >>> >>> >>> > dc-version=1.1.12-1.1.12.git20140904.266d5c2 \
>>> >> >> >> >>> >>> >>> > cluster-infrastructure=corosync \
>>> >> >> >> >>> >>> >>> > no-quorum-policy=ignore \
>>> >> >> >> >>> >>> >>> > last-lrm-refresh=1419514875 \
>>> >> >> >> >>> >>> >>> > cluster-name=xxx \
>>> >> >> >> >>> >>> >>> > stonith-enabled=true
>>> >> >> >> >>> >>> >>> > rsc_defaults rsc-options: \
>>> >> >> >> >>> >>> >>> > resource-stickiness=100
>>> >> >> >> >>> >>> >>> >
>>> >> >> >> >>> >>> >>> > --
>>> >> >> >> >>> >>> >>> >>>> import this
>>> >> >> >> >>> >>> >>> > _______________________________________________
>>> >> >> >> >>> >>> >>> > Linux-HA mailing list
>>> >> >> >> >>> >>> >>> > Linux-HA@lists.linux-ha.org
>>> >> >> >> >>> >>> >>> >
>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >> >> >> >>> >>> >>> > See also: http://linux-ha.org/ReportingProblems
>>> >> >> >> >>> >>> >>>
>>> >> >> >> >>> >>> >>>
>>> >> >> >> >>> >>> >>>
>>> >> >> >> >>> >>> >>> --
>>> >> >> >> >>> >>> >>> esta es mi vida e me la vivo hasta que dios quiera
>>> >> >> >> >>> >>> >>> _______________________________________________
>>> >> >> >> >>> >>> >>> Linux-HA mailing list
>>> >> >> >> >>> >>> >>> Linux-HA@lists.linux-ha.org
>>> >> >> >> >>> >>> >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >> >> >> >>> >>> >>> See also: http://linux-ha.org/ReportingProblems
>>> >> >> >> >>> >>> >>>
>>> >> >> >> >>> >>> >>
>>> >> >> >> >>> >>> >>
>>> >> >> >> >>> >>> >>
>>> >> >> >> >>> >>> >> --
>>> >> >> >> >>> >>> >> >>> import this
>>> >> >> >> >>> >>> >>
>>> >> >> >> >>> >>> >
>>> >> >> >> >>> >>> >
>>> >> >> >> >>> >>> >
>>> >> >> >> >>> >>> > --
>>> >> >> >> >>> >>> >>>> import this
>>> >> >> >> >>> >>> > _______________________________________________
>>> >> >> >> >>> >>> > Linux-HA mailing list
>>> >> >> >> >>> >>> > Linux-HA@lists.linux-ha.org
>>> >> >> >> >>> >>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >> >> >> >>> >>> > See also: http://linux-ha.org/ReportingProblems
>>> >> >> >> >>> >>>
>>> >> >> >> >>> >>>
>>> >> >> >> >>> >>>
>>> >> >> >> >>> >>> --
>>> >> >> >> >>> >>> esta es mi vida e me la vivo hasta que dios quiera
>>> >> >> >> >>> >>> _______________________________________________
>>> >> >> >> >>> >>> Linux-HA mailing list
>>> >> >> >> >>> >>> Linux-HA@lists.linux-ha.org
>>> >> >> >> >>> >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >> >> >> >>> >>> See also: http://linux-ha.org/ReportingProblems
>>> >> >> >> >>> >>>
>>> >> >> >> >>> >>
>>> >> >> >> >>> >>
>>> >> >> >> >>> >>
>>> >> >> >> >>> >> --
>>> >> >> >> >>> >> >>> import this
>>> >> >> >> >>> >>
>>> >> >> >> >>> >
>>> >> >> >> >>> >
>>> >> >> >> >>> >
>>> >> >> >> >>> > --
>>> >> >> >> >>> >>>> import this
>>> >> >> >> >>> > _______________________________________________
>>> >> >> >> >>> > Linux-HA mailing list
>>> >> >> >> >>> > Linux-HA@lists.linux-ha.org
>>> >> >> >> >>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >> >> >> >>> > See also: http://linux-ha.org/ReportingProblems
>>> >> >> >> >>>
>>> >> >> >> >>>
>>> >> >> >> >>>
>>> >> >> >> >>> --
>>> >> >> >> >>> esta es mi vida e me la vivo hasta que dios quiera
>>> >> >> >> >>> _______________________________________________
>>> >> >> >> >>> Linux-HA mailing list
>>> >> >> >> >>> Linux-HA@lists.linux-ha.org
>>> >> >> >> >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >> >> >> >>> See also: http://linux-ha.org/ReportingProblems
>>> >> >> >> >>>
>>> >> >> >> >>
>>> >> >> >> >>
>>> >> >> >> >>
>>> >> >> >> >> --
>>> >> >> >> >>>>> import this
>>> >> >> >> >> _______________________________________________
>>> >> >> >> >> Linux-HA mailing list
>>> >> >> >> >> Linux-HA@lists.linux-ha.org
>>> >> >> >> >> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >> >> >> >> See also: http://linux-ha.org/ReportingProblems
>>> >> >> >> >
>>> >> >> >> >
>>> >> >> >> >
>>> >> >> >> > --
>>> >> >> >> > esta es mi vida e me la vivo hasta que dios quiera
>>> >> >> >>
>>> >> >> >>
>>> >> >> >>
>>> >> >> >> --
>>> >> >> >> esta es mi vida e me la vivo hasta que dios quiera
>>> >> >> >> _______________________________________________
>>> >> >> >> Linux-HA mailing list
>>> >> >> >> Linux-HA@lists.linux-ha.org
>>> >> >> >> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >> >> >> See also: http://linux-ha.org/ReportingProblems
>>> >> >> >>
>>> >> >> >
>>> >> >> >
>>> >> >> >
>>> >> >> > --
>>> >> >> >>>> import this
>>> >> >> > _______________________________________________
>>> >> >> > Linux-HA mailing list
>>> >> >> > Linux-HA@lists.linux-ha.org
>>> >> >> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >> >> > See also: http://linux-ha.org/ReportingProblems
>>> >> >>
>>> >> >>
>>> >> >>
>>> >> >> --
>>> >> >> esta es mi vida e me la vivo hasta que dios quiera
>>> >> >> _______________________________________________
>>> >> >> Linux-HA mailing list
>>> >> >> Linux-HA@lists.linux-ha.org
>>> >> >> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >> >> See also: http://linux-ha.org/ReportingProblems
>>> >> >>
>>> >> >
>>> >> >
>>> >> >
>>> >> > --
>>> >> >>>> import this
>>> >> > _______________________________________________
>>> >> > Linux-HA mailing list
>>> >> > Linux-HA@lists.linux-ha.org
>>> >> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >> > See also: http://linux-ha.org/ReportingProblems
>>> >>
>>> >>
>>> >>
>>> >> --
>>> >> esta es mi vida e me la vivo hasta que dios quiera
>>> >> _______________________________________________
>>> >> Linux-HA mailing list
>>> >> Linux-HA@lists.linux-ha.org
>>> >> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >> See also: http://linux-ha.org/ReportingProblems
>>> >>
>>> >
>>> >
>>> >
>>> > --
>>> >>>> import this
>>> > _______________________________________________
>>> > Linux-HA mailing list
>>> > Linux-HA@lists.linux-ha.org
>>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> > See also: http://linux-ha.org/ReportingProblems
>>>
>>>
>>>
>>> --
>>> esta es mi vida e me la vivo hasta que dios quiera
>>> _______________________________________________
>>> Linux-HA mailing list
>>> Linux-HA@lists.linux-ha.org
>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> See also: http://linux-ha.org/ReportingProblems
>>>
>>
>>
>>
>> --
>> >>> import this
>>
>
>
>
> --
>>>> import this
> _______________________________________________
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems



-- 
esta es mi vida e me la vivo hasta que dios quiera
_______________________________________________
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to