Re: [ClusterLabs] Unable to add 'NodeX' to cluster: node is already in a cluster

2017-06-29 Thread Scott Greenlese

Tomas,

Yes, that was it.

[root@zs95KL corosync]# pcs cluster destroy
Shutting down pacemaker/corosync services...
Redirecting to /bin/systemctl stop  pacemaker.service
Redirecting to /bin/systemctl stop  corosync.service
Killing any remaining services...
Removing all cluster configuration files...
[root@zs95KL corosync]#


[root@zs93kl corosync]# pcs cluster node add zs95KLpcs1,zs95KLpcs2
zs95kjpcs1: Corosync updated
zs93KLpcs1: Corosync updated
zs95KLpcs1: Succeeded
Synchronizing pcsd certificates on nodes zs95KLpcs1...
zs95KLpcs1: Success

Restaring pcsd on the nodes in order to reload the certificates...
zs95KLpcs1: Success
[root@zs93kl corosync]#

Thank you very much for this quick fix.

- Scott

Scott Greenlese ... KVM on System Z - Solutions Test, IBM Poughkeepsie,
N.Y.
  INTERNET:  swgre...@us.ibm.com




From:   Tomas Jelinek 
To: users@clusterlabs.org
Date:   06/29/2017 12:13 PM
Subject:Re: [ClusterLabs] Unable to add 'NodeX' to cluster: node is
already in a cluster



Hi Scott,

It looks like some of cluster configuration files still exist on your
node 'zs95KLpcs1'. Try running "pcs cluster destroy" on that node. This
will delete all cluster config files on the node. So make sure it is the
right node before running the command. Then you should be able to add
the node to your cluster.


Regards,
Tomas



Dne 29.6.2017 v 17:32 Scott Greenlese napsal(a):
> Hi all...
>
> When I try to add a previously removed cluster node back into my
> pacemaker cluster, I get the following error:
>
> [root@zs93kl]# pcs cluster node add zs95KLpcs1,zs95KLpcs2
> Error: Unable to add 'zs95KLpcs1' to cluster: node is already in a
cluster
>
> The node I am adding was recently removed from the cluster, but
> apparently the removal
> was incomplete.
>
> I am looking for some help to thoroughly remove zs95KLpcs1 from this (or
> any other)
> cluster that this host may be a part of.
>
>
> Background:
>
> I had removed node ( zs95KLpcs1) from my 3 node, single ring protocol
> pacemaker cluster while that node
> (which happens to be a KVM on System Z Linux host), was deactivated /
> shut down due to
> relentless, unsolicited STONITH events. My thought was that there was
> some issue with the ring0
> interface (on vlan1293) causing the cluster to initiate fence (power
> off) actions, just minutes after
> joining the cluster. That's why I went ahead and deactivated that node.
>
> The first procedure I used to remove zs95KLpcs1 was flawed, because I
> forgot that there's an issue with
> attempting to remove an unreachable cluster node on the older pacemaker
> code:
>
> [root@zs95kj ]# date;pcs cluster node remove zs95KLpcs1
> Tue Jun 27 18:28:23 EDT 2017
> Error: pcsd is not running on zs95KLpcs1
>
> I then followed this procedure (courtesy of Tomasand Ken inthis user
> group):
>
> 1. run 'pcs cluster localnode remove ' on all remaining nodes
> 2. run 'pcs cluster reload corosync' on one node
> 3. run 'crm_node -R  --force' on one node
>
> My execution:
>
> I made the mistake of manually removing the target node (zs95KLpcs1)
> stanza from corosync.conf file before
> executing the above procedure:
>
> [root@zs95kj ]# vi /etc/corosync/corosync.conf
>
> Removed this stanza:
>
> node {
> ring0_addr: zs95KLpcs1
> nodeid: 3
> }
>
> I then followed the recommended steps ...
>
> [root@zs95kj ]# pcs cluster localnode remove zs95KLpcs1
> Error: unable to remove zs95KLpcs1 ### I assume this was because I
> manually removed the stanza (above)
>
> [root@zs93kl ]# pcs cluster localnode remove zs95KLpcs1
> zs95KLpcs1: successfully removed!
> [root@zs93kl ]#
>
> [root@zs95kj ]# pcs cluster reload corosync
> Corosync reloaded
> [root@zs95kj ]#
>
> [root@zs95kj ]# crm_node -R zs95KLpcs1 --force
> [root@zs95kj ]#
>
>
> [root@zs95kj ]# pcs status |less
> Cluster name: test_cluster_2
> Last updated: Tue Jun 27 18:39:14 2017 Last change: Tue Jun 27 18:38:56
> 2017 by root via crm_node on zs95kjpcs1
> Stack: corosync
> Current DC: zs93KLpcs1 (version 1.1.13-10.el7_2.ibm.1-44eb2dd) -
> partition with quorum
> 45 nodes and 227 resources configured
>
> *Online: [ zs93KLpcs1 zs95kjpcs1 ]*
>
>
> This seemed to work well, at least I'm showing only the two cluster
nodes.
>
> Later on, once I was able to activate zs95KLpcs1 (former cluster member)
> ... I did what I thought
> I should do to tell that node that it's no longer a member of the
cluster:
>
> [root@zs95kj ]# cat neuter.sh
> ssh root@zs95KL "/usr/sbin/pcs cluster localnode *remove *zs95KLpcs1"
> ssh root@zs95KL "/usr/sbin/pcs cluster reload corosync"
> ssh root@zs

[ClusterLabs] Unable to add 'NodeX' to cluster: node is already in a cluster

2017-06-29 Thread Scott Greenlese

Hi all...

When I try to add a previously removed cluster node back into my pacemaker
cluster, I get the following error:

[root@zs93kl]# pcs cluster node add zs95KLpcs1,zs95KLpcs2
Error: Unable to add 'zs95KLpcs1' to cluster: node is already in a cluster

The node I am adding was recently removed from the cluster, but apparently
the removal
was incomplete.

I am looking for some help to thoroughly remove zs95KLpcs1 from this (or
any other)
cluster that this host may be a part of.


Background:

I had removed node ( zs95KLpcs1) from my 3 node, single ring protocol
pacemaker cluster while that node
(which happens to be a KVM on System Z Linux host),  was deactivated / shut
down due to
relentless, unsolicited STONITH events.  My thought was that there was some
issue with the ring0
interface (on vlan1293) causing the cluster to initiate fence (power off)
actions, just minutes after
joining the cluster.   That's why I went ahead and deactivated that node.

The first procedure I used to remove zs95KLpcs1 was flawed, because I
forgot that there's an issue with
attempting to remove an unreachable cluster node on the older pacemaker
code:

[root@zs95kj ]# date;pcs cluster node remove zs95KLpcs1
Tue Jun 27 18:28:23 EDT 2017
Error: pcsd is not running on zs95KLpcs1

I then followed this procedure (courtesy of Tomas and Ken in this user
group):

1. run 'pcs cluster localnode remove ' on all remaining nodes
2. run 'pcs cluster reload corosync' on one node
3. run 'crm_node -R  --force' on one node

My execution:

I made the mistake of manually removing the target node (zs95KLpcs1) stanza
from corosync.conf file before
executing the above procedure:

[root@zs95kj ]# vi /etc/corosync/corosync.conf

Removed this stanza:

node {
ring0_addr: zs95KLpcs1
nodeid: 3
}

I then followed the recommended steps ...

[root@zs95kj ]# pcs cluster localnode remove zs95KLpcs1
Error: unable to remove zs95KLpcs1###  I assume this was because I
manually removed the stanza (above)

[root@zs93kl ]# pcs cluster localnode remove zs95KLpcs1
zs95KLpcs1: successfully removed!
[root@zs93kl ]#

[root@zs95kj ]# pcs cluster reload corosync
Corosync reloaded
[root@zs95kj ]#

[root@zs95kj ]# crm_node -R zs95KLpcs1 --force
[root@zs95kj ]#


[root@zs95kj ]# pcs status |less
Cluster name: test_cluster_2
Last updated: Tue Jun 27 18:39:14 2017  Last change: Tue Jun 27
18:38:56 2017 by root via crm_node on zs95kjpcs1
Stack: corosync
Current DC: zs93KLpcs1 (version 1.1.13-10.el7_2.ibm.1-44eb2dd) - partition
with quorum
45 nodes and 227 resources configured

Online: [ zs93KLpcs1 zs95kjpcs1 ]


This seemed to work well, at least I'm showing only the two cluster nodes.

Later on, once I was able to activate zs95KLpcs1 (former cluster
member) ... I did what I thought
I should do to tell that node that it's no longer a member of the cluster:

[root@zs95kj ]# cat neuter.sh
ssh root@zs95KL "/usr/sbin/pcs cluster localnode remove zs95KLpcs1"
ssh root@zs95KL "/usr/sbin/pcs cluster reload corosync"
ssh root@zs95KL "/usr/sbin/crm_node -R zs95KLpcs1 --force"

[root@zs95kj ]# ./neuter.sh
zs95KLpcs1: successfully removed!
Corosync reloaded
[root@zs95kj ]#


Next, I followed a procedure to convert my current 2-node, single ring
cluster to RRP ... which seems to be running
well, and the corosync config looks like this:

[root@zs93kl ]# for host in  zs95kjpcs1 zs93KLpcs1 ; do ssh $host
"hostname;corosync-cfgtool -s"; done
zs95kj
Printing ring status.
Local node ID 2
RING ID 0
id  = 10.20.93.12
status  = ring 0 active with no faults
RING ID 1
id  = 10.20.94.212
status  = ring 1 active with no faults

zs93kl
Printing ring status.
Local node ID 5
RING ID 0
id  = 10.20.93.13
status  = ring 0 active with no faults
RING ID 1
id  = 10.20.94.213
status  = ring 1 active with no faults
[root@zs93kl ]#


So now, when I try to add zs95KLpcs1 (and the second ring interface,
zs95KLpcs2) to the RRP config,
I get the error:

[root@zs93kl]# pcs cluster node add zs95KLpcs1,zs95KLpcs2
Error: Unable to add 'zs95KLpcs1' to cluster: node is already in a cluster


I re-ran the node removal procedures, and also
deleted /etc/corosync/corosync.conf
on the target node zs95KLpcs1, and nothing I've tried resolves my problem.

I checked to see if zs95KLpcs1 exists in any "corosync.conf" file on the 3
nodes, and it does not.

[root@zs95kj corosync]# grep zs95KLpcs1 *
[root@zs95kj corosync]#

[root@zs93kl corosync]# grep zs95KLpcs1 *
[root@zs95kj corosync]#

[root@zs95KL corosync]# grep zs95KLpcs1 *
[root@zs95kj corosync]#

Thanks in advance ..

Scott Greenlese ... KVM on System Z - Solutions Test, IBM Poughkeepsie,
N.Y.
  INTERNET:  swgre...@us.ibm.com
___
Users mailing list: Users@clusterlabs.org
http://lists.clu

Re: [ClusterLabs] How to force remove a cluster node?

2017-04-19 Thread Scott Greenlese
Tomas,

Yes, I have an IBM internal build we're using for KVM on System Z.I
tried the --force option and, while it didn't complain,
it didn't work either (as expected, as per bug
https://bugzilla.redhat.com/show_bug.cgi?id=1225423), so
maybe it is a valid option.

[root@zs95kj VD]# date; pcs cluster node remove zs95KLpcs1 --force
Wed Apr 19 10:14:10 EDT 2017
Error: pcsd is not running on zs95KLpcs1
[root@zs95kj VD]#

Hopefully we'll roll in pcs-0.9.143-15.el7_2.1 with our next release of
KVM.

In the mean time, thanks very much for all the valuable feedback.  I'm good
to go for now with the workaround.

Scott

Scott Greenlese ... KVM on System Z - Solutions Test, IBM Poughkeepsie,
N.Y.
  INTERNET:  swgre...@us.ibm.com




From:   Tomas Jelinek 
To: users@clusterlabs.org
Date:   04/19/2017 03:25 AM
Subject:Re: [ClusterLabs] How to force remove a cluster node?



Dne 18.4.2017 v 19:52 Scott Greenlese napsal(a):
> My thanks to both Ken Gaillot and Tomas Jelinek for the workaround. The
> procedure(s) worked like a champ.
>
> I just have a few side comments / observations ...
>
> First - Tomas, in the bugzilla you show this error message on your
> cluster remove command, directing you to use the --force option:
>
> [root@rh72-node1:~]# pcs cluster node remove rh72-node3
> Error: pcsd is not running on rh72-node3, use --force to override
>
> When I issue the cluster remove, I do not get and reference to the
> --force option in the error message:
>
> [root@zs93kl ]# pcs cluster node remove zs95KLpcs1
> Error: pcsd is not running on zs95KLpcs1
> [root@zs93kl ]#
>
> The man page doesn't mention --force at my level.

The man page doesn't mention --force for most commands in which --force
can be used. One shouldn't really make any conclusions from that.

> Is this a feature added after pcs-0.9.143-15.el7_2.ibm.2.s390x ?

The feature has been backported to pcs-0.9.143-15.el7_2.1. I cannot
really check if it is present in pcs-0.9.143-15.el7_2.ibm.2.s390x
because I don't have access to that particular build. Based on the name
I would say it was build internally at IBM. However, if the error
message doesn't suggest using --force, than the feature is most likely
not present in that build.

>
> Also, in your workaround procedure, you have me do: 'pcs cluster
> *localnode*remove  '.
> However, wondering why the 'localnode' option is not in the pcs man page
> for the pcs cluster command?
> The command / option worked great, just curious why it's not
documented ...

It's an internal pcs command which is not meant to be run by users. It
exists mostly for the sake of the current pcs/pcsd architecture ("pcs
cluster node" calls pcsd instance on all nodes over network and pcsd
instance runs "pcs cluster localnode" to do the actual job) and is
likely to be removed in the future. It is useful for the workaround as
the check whether all nodes are running is done in the "pcs cluster
node" command.

Regards,
Tomas

>
> [root@zs93kl # pcs cluster localnode remove zs93kjpcs1
> zs93kjpcs1: successfully removed!
>
> My man page level:
>
> [root@zs93kl VD]# rpm -q --whatprovides /usr/share/man/man8/pcs.8.gz
> pcs-0.9.143-15.el7_2.ibm.2.s390x
> [root@zs93kl VD]#
>
> Thanks again,
>
> Scott G.
>
> Scott Greenlese ... KVM on System Z - Solutions Test, IBM Poughkeepsie,
N.Y.
> INTERNET: swgre...@us.ibm.com
>
>
> Inactive hide details for Tomas Jelinek ---04/18/2017 09:04:59 AM---Dne
> 17.4.2017 v 17:28 Ken Gaillot napsal(a): > On 04/13/201Tomas Jelinek
> ---04/18/2017 09:04:59 AM---Dne 17.4.2017 v 17:28 Ken Gaillot napsal(a):
>> On 04/13/2017 01:11 PM, Scott Greenlese wrote:
>
> From: Tomas Jelinek 
> To: users@clusterlabs.org
> Date: 04/18/2017 09:04 AM
> Subject: Re: [ClusterLabs] How to force remove a cluster node?
>
> 
>
>
>
> Dne 17.4.2017 v 17:28 Ken Gaillot napsal(a):
>> On 04/13/2017 01:11 PM, Scott Greenlese wrote:
>>> Hi,
>>>
>>> I need to remove some nodes from my existing pacemaker cluster which
are
>>> currently unbootable / unreachable.
>>>
>>> Referenced
>>>
>
https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Linux/7/html/High_Availability_Add-On_Reference/s1-clusternodemanage-HAAR.html#s2-noderemove-HAAR

>>>
>>> *4.4.4. Removing Cluster Nodes*
>>> The following command shuts down the specified node and removes it from
>>> the cluster configuration file, corosync.conf, on all of the other
nodes
>>> in the cluster. For information on removing all information about the
>>> cluster from the cluster nodes entirely, 

Re: [ClusterLabs] How to force remove a cluster node?

2017-04-18 Thread Scott Greenlese
My thanks to both Ken Gaillot and Tomas Jelinek for the workaround.   The
procedure(s) worked like a champ.

I just have a few side comments / observations ...

First - Tomas,  in the bugzilla you show this error message on your cluster
remove command, directing you to use the --force option:

[root@rh72-node1:~]# pcs cluster node remove rh72-node3
Error: pcsd is not running on rh72-node3, use --force to override

When I issue the cluster remove, I do not get and reference to the --force
option in the error message:

[root@zs93kl ]# pcs cluster node remove  zs95KLpcs1
Error: pcsd is not running on zs95KLpcs1
[root@zs93kl ]#

The man page doesn't mention --force at my level.  Is this a feature added
after pcs-0.9.143-15.el7_2.ibm.2.s390x ?

Also, in your workaround procedure,  you have me do: 'pcs cluster localnode
remove  '.
However, wondering why the 'localnode' option is not in the pcs man page
for the pcs cluster command?
The command / option worked great, just curious why it's not documented ...

[root@zs93kl #  pcs cluster localnode remove zs93kjpcs1
zs93kjpcs1: successfully removed!

My man page level:

[root@zs93kl VD]# rpm -q --whatprovides /usr/share/man/man8/pcs.8.gz
pcs-0.9.143-15.el7_2.ibm.2.s390x
[root@zs93kl VD]#

Thanks again,

Scott G.

Scott Greenlese ... KVM on System Z - Solutions Test, IBM Poughkeepsie,
N.Y.
  INTERNET:  swgre...@us.ibm.com




From:   Tomas Jelinek 
To: users@clusterlabs.org
Date:   04/18/2017 09:04 AM
Subject:Re: [ClusterLabs] How to force remove a cluster node?



Dne 17.4.2017 v 17:28 Ken Gaillot napsal(a):
> On 04/13/2017 01:11 PM, Scott Greenlese wrote:
>> Hi,
>>
>> I need to remove some nodes from my existing pacemaker cluster which are
>> currently unbootable / unreachable.
>>
>> Referenced
>>
https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Linux/7/html/High_Availability_Add-On_Reference/s1-clusternodemanage-HAAR.html#s2-noderemove-HAAR

>>
>> *4.4.4. Removing Cluster Nodes*
>> The following command shuts down the specified node and removes it from
>> the cluster configuration file, corosync.conf, on all of the other nodes
>> in the cluster. For information on removing all information about the
>> cluster from the cluster nodes entirely, thereby destroying the cluster
>> permanently, refer to _Section 4.6, “Removing the Cluster
>> Configuration”_
>> <
https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Linux/7/html/High_Availability_Add-On_Reference/s1-clusterremove-HAAR.html#s2-noderemove-HAAR
>.
>>
>> pcs cluster node remove /node/
>>
>> I ran the command with the cluster active on 3 of the 5 available
>> cluster nodes (with quorum). The command fails with:
>>
>> [root@zs90KP VD]# date;*pcs cluster node remove zs93kjpcs1*
>> Thu Apr 13 13:40:59 EDT 2017
>> *Error: pcsd is not running on zs93kjpcs1*
>>
>>
>> The node was not removed:
>>
>> [root@zs90KP VD]# pcs status |less
>> Cluster name: test_cluster_2
>> Last updated: Thu Apr 13 14:08:15 2017 Last change: Wed Apr 12 16:40:26
>> 2017 by root via cibadmin on zs93KLpcs1
>> Stack: corosync
>> Current DC: zs90kppcs1 (version 1.1.13-10.el7_2.ibm.1-44eb2dd) -
>> partition with quorum
>> 45 nodes and 180 resources configured
>>
>> Node zs95KLpcs1: UNCLEAN (offline)
>> Online: [ zs90kppcs1 zs93KLpcs1 zs95kjpcs1 ]
>> *OFFLINE: [ zs93kjpcs1 ]*
>>
>>
>> Is there a way to force remove a node that's no longer bootable? If not,
>> what's the procedure for removing a rogue cluster node?
>>
>> Thank you...
>>
>> Scott Greenlese ... KVM on System Z - Solutions Test, IBM Poughkeepsie,
N.Y.
>> INTERNET: swgre...@us.ibm.com
>
> Yes, the pcs command is just a convenient shorthand for a series of
> commands. You want to ensure pacemaker and corosync are stopped on the
> node to be removed (in the general case, obviously already done in this
> case), remove the node from corosync.conf and restart corosync on all
> other nodes, then run "crm_node -R " on any one active node.

Hi Scott,

It is possible to remove an offline node from a cluster with upstream
pcs 0.9.154 or RHEL pcs-0.9.152-5 (available in RHEL7.3) or newer.

If you have an older version, here's a workaround:
1. run 'pcs cluster localnode remove ' on all remaining nodes
2. run 'pcs cluster reload corosync' on one node
3. run 'crm_node -R  --force' on one node
It's basically the same procedure Ken described.

See https://bugzilla.redhat.com/show_bug.cgi?id=1225423 for more details.

Regards,
Tomas

___
Users mailing list: Users@clusterlabs.org
http://lists.clust

[ClusterLabs] How to force remove a cluster node?

2017-04-13 Thread Scott Greenlese

Hi,

I need to remove some nodes from my existing pacemaker cluster which are
currently unbootable / unreachable.

Referenced
https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Linux/7/html/High_Availability_Add-On_Reference/s1-clusternodemanage-HAAR.html#s2-noderemove-HAAR

4.4.4. Removing Cluster Nodes
The following command shuts down the specified node and removes it from the
cluster configuration file, corosync.conf, on all of the other nodes in the
cluster. For information on removing all information about the cluster from
the cluster nodes entirely, thereby destroying the cluster permanently,
refer to Section 4.6, “Removing the Cluster Configuration”.
pcs cluster node remove node

I ran the command with the cluster active on 3 of the 5 available cluster
nodes (with quorum).  The command fails with:

[root@zs90KP VD]# date;pcs cluster node remove zs93kjpcs1
Thu Apr 13 13:40:59 EDT 2017
Error: pcsd is not running on zs93kjpcs1


The node was not removed:

[root@zs90KP VD]# pcs status |less
Cluster name: test_cluster_2
Last updated: Thu Apr 13 14:08:15 2017  Last change: Wed Apr 12
16:40:26 2017 by root via cibadmin on zs93KLpcs1
Stack: corosync
Current DC: zs90kppcs1 (version 1.1.13-10.el7_2.ibm.1-44eb2dd) - partition
with quorum
45 nodes and 180 resources configured

Node zs95KLpcs1: UNCLEAN (offline)
Online: [ zs90kppcs1 zs93KLpcs1 zs95kjpcs1 ]
OFFLINE: [ zs93kjpcs1 ]


Is there a way to force remove a node that's no longer bootable?If not,
what's the procedure for removing a rogue cluster node?

Thank you...

Scott Greenlese ... KVM on System Z - Solutions Test, IBM Poughkeepsie,
N.Y.
  INTERNET:  swgre...@us.ibm.com

___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] Antw: Re: Antw: Expected recovery behavior of remote-node guest when corosync ring0 is lost in a passive mode RRP config?

2017-03-08 Thread Scott Greenlese

Ulrich,

Can you share with me some details how your pacemaker, corosync and LGM is
configured?   I should
consider implementing a similar approach and this would be very helpful to
me.   Thanks...

Scott


Scott Greenlese ... KVM on System Z - Solutions Test, IBM Poughkeepsie,
N.Y.
  INTERNET:  swgre...@us.ibm.com
  PHONE:  8/293-7301 (845-433-7301)




From:   "Ulrich Windl" 
To: 
Date:   03/08/2017 03:20 AM
Subject:[ClusterLabs] Antw: Re: Antw: Expected recovery behavior of
remote-node guest when corosync ring0 is lost in a passive mode
RRP config?



>>> "Scott Greenlese"  schrieb am 07.03.2017 um 17:28
in
Nachricht
:
[...]
>
> Maybe my question is... is there any way to facilitate an alternate Live
> Guest Migration path in the event of a ring0_addr failure?
> This might also apply to a single ring protocol as well.
[...]
I don't know the answer, but what we have here is that every network is
connected using two NICs in a bonding interface, and we have two of these
networks for cluster communication, one being exclusively for cluster
communication. And the network being used for migration is separate from
the two networks used for cluster communication. As said, we are USING
SLES11 here, but we never had problems that you have. One problem we had
was that massive parallel VM migrations would overload the network, so we
limited the number of parallel migrations. Maybe this is still helpful...

Regards,
Ulrich




___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org



___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] Antw: Expected recovery behavior of remote-node guest when corosync ring0 is lost in a passive mode RRP config?

2017-03-07 Thread Scott Greenlese

Ulrich,

Thank you very much for your feedback.

You wrote, "Could it be you forgot "allow-migrate=true" at the resource
level or some migration IP address at the node level?
I only have SLES11 here..."

I know for sure that the pacemaker remote node (zs95kjg110102) I mentioned
below is configured correctly for pacemaker Live Guest Migration.
I can demonstrate this using the 'pcs resource move' CLI :

I will migrate this "remote node" guest (zs95kjg110102) and resource
"zs95kjg110102_res" to another cluster node
(e.g. zs95kjpcs1 / 10.20.93.12) , using the 'pcs1' hostname / IP which is
currently running on zs93kjpcs1 (10.20.93.11):

[root@zs95kj ~]# pcs resource show |grep zs95kjg110102_res
 zs95kjg110102_res  (ocf::heartbeat:VirtualDomain): Started zs93kjpcs1

[root@zs93kj ~]# pcs resource move zs95kjg110102_res zs95kjpcs1

[root@zs93kj ~]# pcs resource show |grep zs95kjg110102_res
 zs95kjg110102_res  (ocf::heartbeat:VirtualDomain): Started zs95kjpcs1

## On zs95kjpcs1,  you can see that the guest is actually running there...

[root@zs95kj ~]# virsh list |grep zs95kjg110102
 63zs95kjg110102  running

[root@zs95kj ~]# ping 10.20.110.102
PING 10.20.110.102 (10.20.110.102) 56(84) bytes of data.
64 bytes from 10.20.110.102: icmp_seq=1 ttl=63 time=0.775 ms

So, everything seems set up correctly for live guest migration of this
VirtualDomain resource.

What I am really looking for is a way to ensure 100% availability of a
"live guest migratable" pacemaker remote node guest
in a situation where the interface (in this case vlan1293) ring0_addr goes
down.  I thought that maybe configuring
Redundant Ring Protocol (RRP) for corosync would provide this, but from
what I've seen so far it doesn't
look that way.If the ring0_addr interface is lost in an RRP
configuration while the remote guest is connected
to the host using that ring0_addr, the guest gets rebooted and
reestablishes the  "remote-node-to-host" connection over the ring1_addr,
which is great as long as you don't care if the guest gets rebooted.
Corosync is doing its job of preventing the
cluster node from being fenced by failing over its heartbeat messaging to
ring1, however the remote_node guests take
a short term hit due to the remote-node-to-host reconnect.

In the event of a ring0_addr failure, I don't see any attempt by pacemaker
to migrate the remote_node to another cluster node,
but maybe this is by design, since there is no alternate path for the guest
to use for LGM (i.e. ring0 is a single point of failure).
If the guest could be migrated over an alternate route, it would prevent
the guest outage.

Maybe my question is... is there any way to facilitate an alternate Live
Guest Migration path in the event of a ring0_addr failure?
This might also apply to a single ring protocol as well.

Thanks,

Scott Greenlese ... KVM on System Z - Solutions Test, IBM Poughkeepsie,
N.Y.
  INTERNET:  swgre...@us.ibm.com




From:   "Ulrich Windl" 
To: 
Date:   03/02/2017 02:39 AM
Subject:[ClusterLabs] Antw: Expected recovery behavior of remote-node
guest when corosync ring0 is lost in a passive mode RRP config?



>>> "Scott Greenlese"  schrieb am 01.03.2017 um 22:07
in
Nachricht
:

> Hi..
>
> I am running a few corosync "passive mode" Redundant Ring Protocol (RRP)
> failure scenarios, where
> my cluster has several remote-node VirtualDomain resources running on
each
> node in the cluster,
> which have been configured to allow Live Guest Migration (LGM)
operations.
>
> While both corosync rings are active, if I drop ring0 on a given node
where
> I have remote node (guests) running,
> I noticed that the guest will be shutdown / re-started on the same host,
> after which the connection is re-established
> and the guest proceeds to run on that same cluster node.

Could it be you forgot "allow-migrate=true" at the resource level or some
migration IP address at the node level?
I only have SLES11 here...

>
> I am wondering why pacemaker doesn't try to "live" migrate the remote
node
> (guest) to a different node, instead
> of rebooting the guest?  Is there some way to configure the remote nodes
> such that the recovery action is
> LGM instead of reboot when the host-to-remote_node connect is lost in an
> RRP situation?   I guess the
> next question is, is it even possible to LGM a remote node guest if the
> corosync ring fails over from ring0 to ring1
> (or vise-versa)?
>
> # For example, here's a remote node's VirtualDomain resource definition.
>
> [root@zs95kj]# pcs resource show  zs95kjg110102_res
>  Resource: zs95kjg110102_res (class=ocf provider=heartbeat
> type=VirtualDomain)
>   Attributes: config=/guestxml/nfs1/zs95kjg110102.xml
> hy

[ClusterLabs] Expected recovery behavior of remote-node guest when corosync ring0 is lost in a passive mode RRP config?

2017-03-01 Thread Scott Greenlese
ame: stonith-ng
Feb 28 15:55:37 [942] zs95kjg110102 pacemaker_remoted: info:
qb_ipcs_us_publish:server name: crmd
Feb 28 15:55:37 [942] zs95kjg110102 pacemaker_remoted: info: main:
Starting
Feb 28 15:55:38 [942] zs95kjg110102 pacemaker_remoted:   notice:
lrmd_remote_listen:LRMD client connection established. 0xbed1ab50 id:
b19ed532-6f61-4d9c-9439-ffb836eea34f
zs95kjg110102:~ #



zs95kjg110102:~ # netstat -anp |less
Active Internet connections (servers and established)
Proto Recv-Q Send-Q Local Address   Foreign Address State
PID/Program name
tcp0  0 0.0.0.0:22  0.0.0.0:*   LISTEN
961/sshd
tcp0  0 127.0.0.1:250.0.0.0:*   LISTEN
1065/master
tcp0  0 0.0.0.0:56660.0.0.0:*   LISTEN
946/xinetd
tcp0  0 0.0.0.0:58010.0.0.0:*   LISTEN
946/xinetd
tcp0  0 0.0.0.0:59010.0.0.0:*   LISTEN
946/xinetd
tcp0  0 10.20.110.102:2210.20.94.32:57749
ESTABLISHED 1134/0
tcp0  0 :::21   :::*LISTEN
941/vsftpd
tcp0  0 :::22   :::*LISTEN
961/sshd
tcp0  0 ::1:25  :::*LISTEN
1065/master
tcp0  0 :::80   :::*LISTEN
944/httpd-prefork
tcp0  0 :::3121 :::*LISTEN
942/pacemaker_remot
tcp0  0 :::34836:::*LISTEN
1070/xdm
tcp0  0 10.20.110.102:3121  10.20.94.212:49666
ESTABLISHED 942/pacemaker_remot
udp0  0 :::177  :::*
1070/xdm



## On host node, zs95kj show system messages indicating remote node (guest)
shutdown / start ...  (but no attempt to LGM).

[root@zs95kj ~]# grep "Feb 28" /var/log/messages |grep zs95kjg110102

Feb 28 15:55:07 zs95kj crmd[121380]:   error: Operation
zs95kjg110102_monitor_3: Timed Out (node=zs95kjpcs1, call=2,
timeout=3ms)
Feb 28 15:55:07 zs95kj crmd[121380]:   error: Unexpected disconnect on
remote-node zs95kjg110102
Feb 28 15:55:17 zs95kj crmd[121380]:  notice: Operation
zs95kjg110102_stop_0: ok (node=zs95kjpcs1, call=38, rc=0, cib-update=370,
confirmed=true)
Feb 28 15:55:17 zs95kj attrd[121378]:  notice: Removing all zs95kjg110102
attributes for zs95kjpcs1
Feb 28 15:55:17 zs95kj VirtualDomain(zs95kjg110102_res)[173127]: INFO:
Issuing graceful shutdown request for domain zs95kjg110102.
Feb 28 15:55:23 zs95kj systemd-machined: Machine qemu-38-zs95kjg110102
terminated.
Feb 28 15:55:23 zs95kj crmd[121380]:  notice: Operation
zs95kjg110102_res_stop_0: ok (node=zs95kjpcs1, call=858, rc=0,
cib-update=378, confirmed=true)
Feb 28 15:55:24 zs95kj systemd-machined: New machine qemu-64-zs95kjg110102.
Feb 28 15:55:24 zs95kj systemd: Started Virtual Machine
qemu-64-zs95kjg110102.
Feb 28 15:55:24 zs95kj systemd: Starting Virtual Machine
qemu-64-zs95kjg110102.
Feb 28 15:55:25 zs95kj crmd[121380]:  notice: Operation
zs95kjg110102_res_start_0: ok (node=zs95kjpcs1, call=859, rc=0,
cib-update=385, confirmed=true)
Feb 28 15:55:38 zs95kj crmd[121380]:  notice: Operation
zs95kjg110102_start_0: ok (node=zs95kjpcs1, call=44, rc=0, cib-update=387,
confirmed=true)
[root@zs95kj ~]#


Once the remote node established re-connection, there was no further remote
node / resource instability.

Anyway, just wondering why there was no attempt to migrate this remote node
guest as opposed to a reboot?   Is it necessary to reboot the guest in
order to be managed
by pacemaker and corosync over the ring1 interface if ring0 goes down?
Is live guest migration even possible if ring0 goes away and ring1 takes
over?

Thanks in advance..

Scott Greenlese ... KVM on System Z - Solutions Test, IBM Poughkeepsie,
N.Y.
  INTERNET:  swgre...@us.ibm.com

___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] How to configure my ocf_heartbeat_iscsi resource(s) such that I have both paths to the LUN?

2017-02-23 Thread Scott Greenlese

Refreshing this post in case anyone might have missed it...   still trying
to figure out some way to have multiple iscsi paths managed
by a single ocf_heartbeat_iscsi resource.  Any ideas?

Thanks!

Scott Greenlese ... KVM on System Z - Solutions Test, IBM Poughkeepsie,
N.Y.
  INTERNET:  swgre...@us.ibm.com




From:   Scott Greenlese/Poughkeepsie/IBM@IBMUS
To: users@clusterlabs.org
Cc: Si Bo Niu , Michael
Tebolt/Poughkeepsie/IBM@IBMUS
Date:   02/15/2017 12:18 PM
Subject:[ClusterLabs] How to configure my ocf_heartbeat_iscsi resource
(s) such that I have both paths to the LUN?



Hi folks,

I'm running some test scenarios with an ocf_heartbeat_iscsi pacemaker
resource,
using the following XIV multipath'ed configuration:

I created a single XIV iscsi host definition containing all the pacemaker
host (cluster node) 'Initiator's:

XIV 7812475>>host_list_ports host=pacemaker_iscsi
Host Type Port Name
pacemaker_iscsi iSCSI iqn.2005-03.org.open-iscsi:6539c3daf095
pacemaker_iscsi iSCSI iqn.2005-03.org.open-iscsi:11d639c0976c
pacemaker_iscsi iSCSI iqn.1994-05.com.redhat:74ca24d6476
pacemaker_iscsi iSCSI iqn.1994-05.com.redhat:ea17bebd09a
pacemaker_iscsi iSCSI iqn.1994-05.com.redhat:b852a67852c

Here is the XIV Target IP:port / IQN info:

10.20.92.108:3260
Target: iqn.2005-10.com.xivstorage:012475

10.20.92.109:3260
Target: iqn.2005-10.com.xivstorage:012475

I mapped a single 17Gb Lun to the XIV host: 20017380030BB110F


>From ocf_heartbeat_iscsi man page, it didn't seem obvious to me how I might
specify multiple
IP Portals (i.e. both the .108 and .109 IPs) to a single
ocf_heartbeat_iscsi resource. So,
I went ahead and configured my resource using just the .108 IP path, as
follows:

Using the above XIV definitions, I created a single iscsi pacemaker
resource using only one of the two IP
paths to the XIV LUN:

[root@zs95kj VD]# pcs resource show iscsi_r4

Resource: iscsi_r4 (class=ocf provider=heartbeat type=iscsi)
Attributes: portal=10.20.92.108:3260
target=iqn.2005-10.com.xivstorage:012475
Operations: start interval=0s timeout=120 (iscsi_r4-start-interval-0s)
stop interval=0s timeout=120 (iscsi_r4-stop-interval-0s)
monitor interval=120 timeout=30 (iscsi_r4-monitor-interval-120)

I'm looking for suggestions as to how I should configure my iscsi resource
(s) such that I have both paths (.108 and .109)
to the LUN available to my application? Do I need to create a second iscsi
resource for the .109 path, and colocate
them so that they move about the cluster together?

As an aside, I have run into situations where the second IP (.109) comes
online to where my iscsi_r4 resource is running (how or why,
I'm not sure), which introduces problems because iscsi_r4 only manages
the .108 connection.

Thanks in advance...

Scott Greenlese ... KVM on System Z - Solutions Test, IBM Poughkeepsie,
N.Y.
INTERNET: swgre...@us.ibm.com

___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[ClusterLabs] How to configure my ocf_heartbeat_iscsi resource(s) such that I have both paths to the LUN?

2017-02-15 Thread Scott Greenlese

Hi folks,

I'm running some test scenarios with an ocf_heartbeat_iscsi pacemaker
resource,
using the following XIV multipath'ed configuration:

I created a single XIV iscsi host definition containing all the pacemaker
host (cluster node) 'Initiator's:

XIV 7812475>>host_list_ports host=pacemaker_iscsi
Host  TypePort Name
pacemaker_iscsi   iSCSI   iqn.2005-03.org.open-iscsi:6539c3daf095
pacemaker_iscsi   iSCSI   iqn.2005-03.org.open-iscsi:11d639c0976c
pacemaker_iscsi   iSCSI   iqn.1994-05.com.redhat:74ca24d6476
pacemaker_iscsi   iSCSI   iqn.1994-05.com.redhat:ea17bebd09a
pacemaker_iscsi   iSCSI   iqn.1994-05.com.redhat:b852a67852c

Here is the XIV Target IP:port / IQN info:

10.20.92.108:3260
Target: iqn.2005-10.com.xivstorage:012475

10.20.92.109:3260
Target: iqn.2005-10.com.xivstorage:012475

I mapped a single 17Gb Lun to the XIV host: 20017380030BB110F


>From ocf_heartbeat_iscsi man page, it didn't seem obvious to me how I might
specify multiple
IP Portals (i.e.  both the .108 and .109 IPs) to a single
ocf_heartbeat_iscsi resource.   So,
I went ahead and configured my resource using just the .108 IP path, as
follows:

Using the above XIV definitions, I created a single iscsi pacemaker
resource using only one of the two IP
paths to the XIV LUN:

[root@zs95kj VD]# pcs resource show iscsi_r4

 Resource: iscsi_r4 (class=ocf provider=heartbeat type=iscsi)
  Attributes: portal=10.20.92.108:3260
target=iqn.2005-10.com.xivstorage:012475
  Operations: start interval=0s timeout=120 (iscsi_r4-start-interval-0s)
  stop interval=0s timeout=120 (iscsi_r4-stop-interval-0s)
  monitor interval=120 timeout=30
(iscsi_r4-monitor-interval-120)

I'm looking for suggestions as to how I should configure my iscsi resource
(s) such that I have both paths (.108 and .109)
to the LUN available to my application? Do I need to create a second
iscsi resource for the .109 path, and colocate
them so that they move about the cluster together?

As an aside, I have run into situations where the second IP (.109) comes
online to where my iscsi_r4 resource is running (how or why,
I'm not sure),  which introduces problems because iscsi_r4 only manages
the .108 connection.

Thanks in advance...

Scott Greenlese ... KVM on System Z - Solutions Test, IBM Poughkeepsie,
N.Y.
  INTERNET:  swgre...@us.ibm.com

___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] pcsd 99% CPU

2017-02-06 Thread Scott Greenlese
 /lib64/libbz2.so.1
Saved corefile /guestimages/nfs1/zs95kj_cib_core_Feb6.27225



In the pcsd log, I see "pcs status nodes corosync" requests (approximately)
once per 64 seconds.

[root@zs95kj cluster]# grep "pcs status nodes
corosync" /var/log/pcsd/pcsd.log |wc -l
1861

[root@zs95kj ~]# grep "pcs status nodes corosync" /var/log/pcsd/pcsd.log |
tail
I, [2017-02-06T16:18:53.468885 #23277]  INFO -- : Running: /usr/sbin/pcs
status nodes corosync
I, [2017-02-06T16:19:57.139796 #23277]  INFO -- : Running: /usr/sbin/pcs
status nodes corosync
I, [2017-02-06T16:21:00.833517 #23277]  INFO -- : Running: /usr/sbin/pcs
status nodes corosync
I, [2017-02-06T16:22:04.151613 #23277]  INFO -- : Running: /usr/sbin/pcs
status nodes corosync
I, [2017-02-06T16:23:07.803426 #23277]  INFO -- : Running: /usr/sbin/pcs
status nodes corosync
I, [2017-02-06T16:24:11.360107 #23277]  INFO -- : Running: /usr/sbin/pcs
status nodes corosync
I, [2017-02-06T16:25:14.828617 #23277]  INFO -- : Running: /usr/sbin/pcs
status nodes corosync
I, [2017-02-06T16:26:18.553637 #23277]  INFO -- : Running: /usr/sbin/pcs
status nodes corosync
I, [2017-02-06T16:27:22.363723 #23277]  INFO -- : Running: /usr/sbin/pcs
status nodes corosync
I, [2017-02-06T16:28:26.330005 #23277]  INFO -- : Running: /usr/sbin/pcs
status nodes corosync


A typical, repeating set of entries in the pcsd.log look like this:

I, [2017-02-06T16:29:26.367339 #23277]  INFO -- :
Running: /usr/sbin/corosync-cmapctl totem.cluster_name
I, [2017-02-06T16:29:26.367458 #23277]  INFO -- : CIB USER: hacluster,
groups:
I, [2017-02-06T16:29:27.438925 #23277]  INFO -- : Return Value: 0
:::10.20.93.13 - - [06/Feb/2017:16:29:27 -0500]
"GET /remote/get_configs HTTP/1.1" 200 1077 1.0768
:::10.20.93.13 - - [06/Feb/2017:16:29:27 -0500]
"GET /remote/get_configs HTTP/1.1" 200 1077 1.0770
zs93KLpcs1.pokprv.stglabs.ibm.com - - [06/Feb/2017:16:29:26 EST]
"GET /remote/get_configs HTTP/1.1" 200 1077
- -> /remote/get_configs
I, [2017-02-06T16:29:28.816100 #23277]  INFO -- :
Running: /usr/sbin/corosync-cmapctl totem.cluster_name
I, [2017-02-06T16:29:28.816288 #23277]  INFO -- : CIB USER: hacluster,
groups:
I, [2017-02-06T16:29:29.890788 #23277]  INFO -- : Return Value: 0
I, [2017-02-06T16:29:29.890958 #23277]  INFO -- : Running: /usr/sbin/pcs
status nodes corosync
I, [2017-02-06T16:29:29.890990 #23277]  INFO -- : CIB USER: hacluster,
groups:
I, [2017-02-06T16:29:31.082885 #23277]  INFO -- : Return Value: 0
I, [2017-02-06T16:29:31.084041 #23277]  INFO -- : SRWT Node: zs90kppcs1
Request: get_configs
I, [2017-02-06T16:29:31.085012 #23277]  INFO -- : SRWT Node: zs93kjpcs1
Request: get_configs
I, [2017-02-06T16:29:31.086098 #23277]  INFO -- : SRWT Node: zs93KLpcs1
Request: get_configs
I, [2017-02-06T16:29:31.088295 #23277]  INFO -- : SRWT Node: zs95kjpcs1
Request: get_configs
I, [2017-02-06T16:29:31.088515 #23277]  INFO -- : SRWT Node: zs95KLpcs1
Request: get_configs
I, [2017-02-06T16:29:31.157357 #23277]  INFO -- :
Running: /usr/sbin/corosync-cmapctl totem.cluster_name
I, [2017-02-06T16:29:31.157480 #23277]  INFO -- : CIB USER: hacluster,
groups:
I, [2017-02-06T16:29:32.271072 #23277]  INFO -- : Return Value: 0
:::10.20.93.12 - - [06/Feb/2017:16:29:32 -0500]
"GET /remote/get_configs HTTP/1.1" 200 1077 1.1190
:::10.20.93.12 - - [06/Feb/2017:16:29:32 -0500]
"GET /remote/get_configs HTTP/1.1" 200 1077 1.1193
zs95kjpcs1.pokprv.stglabs.ibm.com - - [06/Feb/2017:16:29:31 EST]
"GET /remote/get_configs HTTP/1.1" 200 1077
- -> /remote/get_configs


Again, the pcsd query commands are returning quickly now, so all this data
may not be of any use now.

Comments / suggestions welcome.  Thanks...



Scott Greenlese ... IBM KVM on System Z Solutions Test,  Poughkeepsie, N.Y.
  INTERNET:  swgre...@us.ibm.com





From:   Tomas Jelinek 
To: users@clusterlabs.org
Date:   02/06/2017 04:55 AM
Subject:Re: [ClusterLabs] pcsd 99% CPU



Dne 3.2.2017 v 22:08 Scott Greenlese napsal(a):
> Hi all..
>
> Over the past few days, I noticed that pcsd and ruby process is pegged
> at 99% CPU, and commands such as
> pcs status pcsd take up to 5 minutes to complete. On all active cluster
> nodes, top shows:
>
> PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
> 27225 haclust+ 20 0 116324 91600 23136 R 99.3 0.1 1943:40 cib
> 23277 root 20 0 12.868g 8.176g 8460 S 99.7 13.0 407:44.18 ruby
>
> The system log indicates High CIB load detected over the past 2 days:
>
> [root@zs95kj ~]# grep "High CIB load detected" /var/log/messages |grep
> "Feb 3" |wc -l
> 1655
> [root@zs95kj ~]# grep "High CIB load detected" /var/log/messages |grep
> "Feb 2" |wc -l
> 1658
> [root@zs95kj ~]# grep "High CIB load detected" /var/log/messages |grep
> "Feb 1" |wc -l
> 147
> [root@zs95kj ~]# grep "High CIB load detected

Re: [ClusterLabs] Failure to configure iface-bridge resource causes cluster node fence action.

2017-02-06 Thread Scott Greenlese

Further explanation for my concern about --disabled not taking effect until
after the iface-bridge was configured  ...

The reason I wanted to create the iface-bridge resource "disabled", was to
allow me the opportunity to impose
a location constraint / rule  on the resource to prevent it from being
started on certain cluster nodes,
where the specified slave vlan did not exist.

In my case, pacemaker assigned the resource to a cluster node where the
specified slave vlan did not exist, which in turn
triggered a fenced (off) action against that node (apparently, because the
device could not be stopped, per Ken's reply earlier).

Again, my cluster is configured as "symmetric" , so I would have to "opt
out" my new resource from
certain cluster nodes via location constraint.

So, if this really is how --disable is designed to work, is there any way
to impose a location constraint rule BEFORE
the iface-bridge resource gets assigned. configured and started on a
cluster node in a symmetrical cluster?

Thanks,

Scott Greenlese ... IBM KVM on System Z -  Solutions Test,  Poughkeepsie,
N.Y.
  INTERNET:  swgre...@us.ibm.com





From:   Scott Greenlese/Poughkeepsie/IBM@IBMUS
To: kgail...@redhat.com, Cluster Labs - All topics related to
open-source clustering welcomed 
Date:   02/03/2017 03:23 PM
Subject:Re: [ClusterLabs] Failure to configure iface-bridge resource
causes cluster node fence action.



Ken,

Thanks for the explanation.

One other thing, relating to the iface-bridge resource creation. I
specified --disabled flag:

> [root@zs95kj VD]# date;pcs resource create br0_r1
> ocf:heartbeat:iface-bridge bridge_name=br0 bridge_slaves=vlan1292 op
> monitor timeout="20s" interval="10s" --disabled

Does the bridge device have to be successfully configured by pacemaker
before disabling the resource? It seems
that that was the behavior, since it failed the resource and fenced the
node instead of disabling the resource.
Just checking with you to be sure.

Thanks again..

Scott Greenlese ... IBM KVM on System Z Solutions Test, Poughkeepsie, N.Y.
INTERNET: swgre...@us.ibm.com



Inactive hide details for Ken Gaillot ---02/02/2017 03:29:12 PM---On
02/02/2017 02:14 PM, Scott Greenlese wrote: > Hi folks,Ken Gaillot
---02/02/2017 03:29:12 PM---On 02/02/2017 02:14 PM, Scott Greenlese wrote:
> Hi folks,

From: Ken Gaillot 
To: users@clusterlabs.org
Date: 02/02/2017 03:29 PM
Subject: Re: [ClusterLabs] Failure to configure iface-bridge resource
causes cluster node fence action.



On 02/02/2017 02:14 PM, Scott Greenlese wrote:
> Hi folks,
>
> I'm testing iface-bridge resource support on a Linux KVM on System Z
> pacemaker cluster.
>
> pacemaker-1.1.13-10.el7_2.ibm.1.s390x
> corosync-2.3.4-7.el7_2.ibm.1.s390x
>
> I created an iface-bridge resource, but specified a non-existent
> bridge_slaves value, vlan1292 (i.e. vlan1292 doesn't exist).
>
> [root@zs95kj VD]# date;pcs resource create br0_r1
> ocf:heartbeat:iface-bridge bridge_name=br0 bridge_slaves=vlan1292 op
> monitor timeout="20s" interval="10s" --disabled
> Wed Feb 1 17:49:16 EST 2017
> [root@zs95kj VD]#
>
> [root@zs95kj VD]# pcs resource show |grep br0
> br0_r1 (ocf::heartbeat:iface-bridge): FAILED zs93kjpcs1
> [root@zs95kj VD]#
>
> As you can see, the resource was created, but failed to start on the
> target node zs93kppcs1.
>
> To my surprise, the target node zs93kppcs1 was unceremoniously fenced.
>
> pacemaker.log shows a fence (off) action initiated against that target
> node, "because of resource failure(s)" :
>
> Feb 01 17:55:56 [52941] zs95kj crm_resource: ( unpack.c:2719 ) debug:
> determine_op_status: br0_r1_stop_0 on zs93kjpcs1 returned 'not
> configured' (6) instead of the expected value: 'ok' (0)
> Feb 01 17:55:56 [52941] zs95kj crm_resource: ( unpack.c:2602 ) warning:
> unpack_rsc_op_failure: Processing failed op stop for br0_r1 on
> zs93kjpcs1: not configured (6)
> Feb 01 17:55:56 [52941] zs95kj crm_resource: ( unpack.c:3244 ) error:
> unpack_rsc_op: Preventing br0_r1 from re-starting anywhere: operation
> stop failed 'not configured' (6)
> Feb 01 17:55:56 [52941] zs95kj crm_resource: ( unpack.c:2719 ) debug:
> determine_op_status: br0_r1_stop_0 on zs93kjpcs1 returned 'not
> configured' (6) instead of the expected value: 'ok' (0)
> Feb 01 17:55:56 [52941] zs95kj crm_resource: ( unpack.c:2602 ) warning:
> unpack_rsc_op_failure: Processing failed op stop for br0_r1 on
> zs93kjpcs1: not configured (6)
> Feb 01 17:55:56 [52941] zs95kj crm_resource: ( unpack.c:3244 ) error:
> unpack_rsc_op: Preventing br0_r1 from re-starting anywhere: operation
> stop failed 'not configured' (6)
> Feb 01 17:55:56 [52941] zs95kj crm_resourc

[ClusterLabs] pcsd 99% CPU

2017-02-03 Thread Scott Greenlese

Hi all..

Over the past few days, I noticed that pcsd and ruby process is pegged at
99% CPU, and commands such as
pcs status pcsd  take up to 5 minutes to complete.  On all active cluster
nodes, top shows:

PID USER PR NI  VIRT  RES SHRS  %CPU %MEM  TIME+
COMMAND
27225   haclust+ 20 0   116324   91600 23136 R  99.3
0.1  1943:40cib
23277   root   200  12.868g  8.176g   8460   S  99.7
13.0407:44.18   ruby

The system log indicates High CIB load detected over the past 2 days:

[root@zs95kj ~]# grep "High CIB load detected" /var/log/messages |grep "Feb
3" |wc -l
1655
[root@zs95kj ~]# grep "High CIB load detected" /var/log/messages |grep "Feb
2" |wc -l
1658
[root@zs95kj ~]# grep "High CIB load detected" /var/log/messages |grep "Feb
1" |wc -l
147
[root@zs95kj ~]# grep "High CIB load detected" /var/log/messages |grep "Jan
31" |wc -l
444
[root@zs95kj ~]# grep "High CIB load detected" /var/log/messages |grep "Jan
30" |wc -l
352


The first entries logged on Feb 2 started around 8:42am ...

Feb  2 08:42:12 zs95kj crmd[27233]:  notice: High CIB load detected:
0.974333

This happens to coincide with the time that I had caused a node fence (off)
action by creating a iface-bridge resources and specified
a non-existent vlan slave interface (reported to the group yesterday in a
separate email thread).   It also happened to cause me to lose
quorum in the cluster, because 2 of my 5 cluster nodes were already
offline.

My cluster currently has just over 200 VirtualDomain resources to manage,
plus one iface-bridge resource and one iface-vlan resource.
Both of which are currently configured properly and operational.

I would appreciate some guidance how to proceed with debugging this issue.
I have not taken any recovery actions yet.
I considered stopping the cluster, recycling pcsd.service on all nodes,
restarting cluster... and also, reboot the nodes, if
necessary.  But, didn't want to clear it yet in case there's anything I can
capture while in this state.

Thanks..

Scott Greenlese ... KVM on System Z -  Solutions Test,  Poughkeepsie, N.Y.
  INTERNET:  swgre...@us.ibm.com
___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] Failure to configure iface-bridge resource causes cluster node fence action.

2017-02-03 Thread Scott Greenlese

Ken,

Thanks for the explanation.

One other thing, relating to the iface-bridge resource creation.   I
specified --disabled flag:

> [root@zs95kj VD]# date;pcs resource create br0_r1
> ocf:heartbeat:iface-bridge bridge_name=br0 bridge_slaves=vlan1292 op
> monitor timeout="20s" interval="10s" --disabled

Does the bridge device have to be successfully configured by pacemaker
before disabling the resource?   It seems
that that was the behavior, since it failed the resource and fenced the
node instead of disabling the resource.
Just checking with you to be sure.

Thanks again..

Scott Greenlese ... IBM KVM on System Z Solutions Test,  Poughkeepsie, N.Y.
  INTERNET:  swgre...@us.ibm.com





From:   Ken Gaillot 
To: users@clusterlabs.org
Date:   02/02/2017 03:29 PM
Subject:Re: [ClusterLabs] Failure to configure iface-bridge resource
causes cluster node fence action.



On 02/02/2017 02:14 PM, Scott Greenlese wrote:
> Hi folks,
>
> I'm testing iface-bridge resource support on a Linux KVM on System Z
> pacemaker cluster.
>
> pacemaker-1.1.13-10.el7_2.ibm.1.s390x
> corosync-2.3.4-7.el7_2.ibm.1.s390x
>
> I created an iface-bridge resource, but specified a non-existent
> bridge_slaves value, vlan1292 (i.e. vlan1292 doesn't exist).
>
> [root@zs95kj VD]# date;pcs resource create br0_r1
> ocf:heartbeat:iface-bridge bridge_name=br0 bridge_slaves=vlan1292 op
> monitor timeout="20s" interval="10s" --disabled
> Wed Feb 1 17:49:16 EST 2017
> [root@zs95kj VD]#
>
> [root@zs95kj VD]# pcs resource show |grep br0
> br0_r1 (ocf::heartbeat:iface-bridge): FAILED zs93kjpcs1
> [root@zs95kj VD]#
>
> As you can see, the resource was created, but failed to start on the
> target node zs93kppcs1.
>
> To my surprise, the target node zs93kppcs1 was unceremoniously fenced.
>
> pacemaker.log shows a fence (off) action initiated against that target
> node, "because of resource failure(s)" :
>
> Feb 01 17:55:56 [52941] zs95kj crm_resource: ( unpack.c:2719 ) debug:
> determine_op_status: br0_r1_stop_0 on zs93kjpcs1 returned 'not
> configured' (6) instead of the expected value: 'ok' (0)
> Feb 01 17:55:56 [52941] zs95kj crm_resource: ( unpack.c:2602 ) warning:
> unpack_rsc_op_failure: Processing failed op stop for br0_r1 on
> zs93kjpcs1: not configured (6)
> Feb 01 17:55:56 [52941] zs95kj crm_resource: ( unpack.c:3244 ) error:
> unpack_rsc_op: Preventing br0_r1 from re-starting anywhere: operation
> stop failed 'not configured' (6)
> Feb 01 17:55:56 [52941] zs95kj crm_resource: ( unpack.c:2719 ) debug:
> determine_op_status: br0_r1_stop_0 on zs93kjpcs1 returned 'not
> configured' (6) instead of the expected value: 'ok' (0)
> Feb 01 17:55:56 [52941] zs95kj crm_resource: ( unpack.c:2602 ) warning:
> unpack_rsc_op_failure: Processing failed op stop for br0_r1 on
> zs93kjpcs1: not configured (6)
> Feb 01 17:55:56 [52941] zs95kj crm_resource: ( unpack.c:3244 ) error:
> unpack_rsc_op: Preventing br0_r1 from re-starting anywhere: operation
> stop failed 'not configured' (6)
> Feb 01 17:55:56 [52941] zs95kj crm_resource: ( unpack.c:96 ) warning:
> pe_fence_node: Node zs93kjpcs1 will be fenced because of resource failure
(s)
>
>
> Thankfully, I was able to successfully create a iface-bridge resource
> when I changed the bridge_slaves value to an existent vlan interface.
>
> My main concern is, why would the response to a failed bridge config
> operation warrant a node fence (off) action? Isn't it enough to just
> fail the resource and try another cluster node,
> or at most, give up if it can't be started / configured on any node?
>
> Is there any way to control this harsh recovery action in the cluster?
>
> Thanks much..
>
>
> Scott Greenlese ... IBM KVM on System Z Solutions Test, Poughkeepsie,
N.Y.
> INTERNET: swgre...@us.ibm.com

It's actually the stop operation failure that leads to the fence.

If a resource fails to stop, fencing is the only way pacemaker can
recover the resource elsewhere. Consider a database master -- if it
doesn't stop, starting the master elsewhere could lead to severe data
inconsistency.

You can tell pacemaker to not attempt recovery, by setting on-fail=block
on the stop operation, so it doesn't need to fence. Obviously, that
prevents high availability, as manual intervention is required to do
anything further with the service.

___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org



___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[ClusterLabs] Failure to configure iface-bridge resource causes cluster node fence action.

2017-02-02 Thread Scott Greenlese

Hi folks,

I'm testing iface-bridge resource support on a Linux KVM on System Z
pacemaker cluster.

pacemaker-1.1.13-10.el7_2.ibm.1.s390x
corosync-2.3.4-7.el7_2.ibm.1.s390x

I created an iface-bridge resource, but specified a non-existent
bridge_slaves value, vlan1292  (i.e. vlan1292 doesn't exist).

[root@zs95kj VD]# date;pcs resource create br0_r1
ocf:heartbeat:iface-bridge bridge_name=br0 bridge_slaves=vlan1292 op
monitor timeout="20s"  interval="10s" --disabled
Wed Feb  1 17:49:16 EST 2017
[root@zs95kj VD]#

[root@zs95kj VD]# pcs resource show |grep br0
 br0_r1 (ocf::heartbeat:iface-bridge):  FAILED zs93kjpcs1
[root@zs95kj VD]#

As you can see, the resource was created, but failed to start on the target
node zs93kppcs1.

To my surprise, the target node zs93kppcs1 was unceremoniously fenced.

pacemaker.log shows a fence (off) action initiated against that target
node, "because of resource failure(s)" :

Feb 01 17:55:56 [52941] zs95kj crm_resource: (unpack.c:2719  )   debug:
determine_op_status:br0_r1_stop_0 on zs93kjpcs1 returned 'not
configured' (6) instead of the expected value: 'ok' (0)
Feb 01 17:55:56 [52941] zs95kj crm_resource: (unpack.c:2602  ) warning:
unpack_rsc_op_failure:  Processing failed op stop for br0_r1 on
zs93kjpcs1: not configured (6)
Feb 01 17:55:56 [52941] zs95kj crm_resource: (unpack.c:3244  )   error:
unpack_rsc_op:  Preventing br0_r1 from re-starting anywhere: operation
stop failed 'not configured' (6)
Feb 01 17:55:56 [52941] zs95kj crm_resource: (unpack.c:2719  )   debug:
determine_op_status:br0_r1_stop_0 on zs93kjpcs1 returned 'not
configured' (6) instead of the expected value: 'ok' (0)
Feb 01 17:55:56 [52941] zs95kj crm_resource: (unpack.c:2602  ) warning:
unpack_rsc_op_failure:  Processing failed op stop for br0_r1 on
zs93kjpcs1: not configured (6)
Feb 01 17:55:56 [52941] zs95kj crm_resource: (unpack.c:3244  )   error:
unpack_rsc_op:  Preventing br0_r1 from re-starting anywhere: operation
stop failed 'not configured' (6)
Feb 01 17:55:56 [52941] zs95kj crm_resource: (unpack.c:96) warning:
pe_fence_node:  Node zs93kjpcs1 will be fenced because of resource
failure(s)


Thankfully, I was able to successfully create a iface-bridge resource when
I changed the bridge_slaves value to an existent vlan interface.

My main concern is, why would the response to a failed bridge config
operation warrant a node fence (off) action?  Isn't it enough to just fail
the resource and try another cluster node,
or at most, give up if it can't be started / configured on any node?

Is there any way to control this harsh recovery action in the cluster?

Thanks much..


Scott Greenlese ... IBM KVM on System Z Solutions Test,  Poughkeepsie, N.Y.
  INTERNET:  swgre...@us.ibm.com

___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] Live Guest Migration timeouts for VirtualDomain resources

2017-02-01 Thread Scott Greenlese

Ken (and Ulrich),

Found it!  You're right, we do deliver a man page...

man]# find . -name *Virtual* -print
./man7/ocf_heartbeat_VirtualDomain.7.gz

# rpm -q
--whatprovides /usr/share/man/man7/ocf_heartbeat_VirtualDomain.7.gz
resource-agents-3.9.7-4.el7_2.kvmibm1_1_3.1.s390x

Much obliged, sir(s).

Scott Greenlese ... IBM z/BX Solutions Test,  Poughkeepsie, N.Y.
  INTERNET:  swgre...@us.ibm.com
  PHONE:  8/293-7301 (845-433-7301)M/S:  POK 42HA/P966




From:   Ken Gaillot 
To: users@clusterlabs.org
Date:   02/01/2017 10:33 AM
Subject:Re: [ClusterLabs] Live Guest Migration timeouts for
VirtualDomain resources



On 02/01/2017 09:15 AM, Scott Greenlese wrote:
> Hi all...
>
> Just a quick follow-up.
>
> Thought I should come clean and share with you that the incorrect
> "migrate-to" operation name defined in my VirtualDomain
> resource was my mistake. It was mis-coded in the virtual guest
> provisioning script. I have since changed it to "migrate_to"
> and of course, the specified live migration timeout value is working
> effectively now. (For some reason, I assumed we were letting that
> operation meta value default).
>
> I was wondering if someone could refer me to the definitive online link
> for pacemaker resource man pages? I don't see any resource man pages
> installed
> on my system anywhere. I found this one online:
> https://www.mankier.com/7/ocf_heartbeat_VirtualDomain but is there a
> more 'official' page I should refer our
> Linux KVM on System z customers to?

All distributions that I know of include the man pages with the packages
they distribute. Are you building from source? They are named like "man
ocf_heartbeat_IPaddr2".

FYI after following this thread, the pcs developers are making a change
so that pcs refuses to add an unrecognized operation unless the user
uses --force. Thanks for being involved in the community; this is how we
learn to improve!

> Thanks again for your assistance.
>
> Scott Greenlese ...IBM KVM on System Z Solution Test Poughkeepsie, N.Y.
> INTERNET: swgre...@us.ibm.com
>
>
> Inactive hide details for "Ulrich Windl" ---01/27/2017 02:32:43 AM--->>>
> "Scott Greenlese"  schrieb am 27."Ulrich Windl"
> ---01/27/2017 02:32:43 AM--->>> "Scott Greenlese" 
> schrieb am 27.01.2017 um 02:47 in Nachricht
>
> From: "Ulrich Windl" 
> To: , Scott Greenlese/Poughkeepsie/IBM@IBMUS
> Cc: "Si Bo Niu" , Michael
Tebolt/Poughkeepsie/IBM@IBMUS
> Date: 01/27/2017 02:32 AM
> Subject: Antw: Re: [ClusterLabs] Antw: Re: Live Guest Migration timeouts
> for VirtualDomain resources
>
> 
>
>
>
>>>> "Scott Greenlese"  schrieb am 27.01.2017 um
> 02:47 in
> Nachricht
>

> m>:
>
>> Hi guys..
>>
>> Well, today I confirmed that what Ulrich said is correct.  If I update
the
>> VirtualDomain resource with the operation name  "migrate_to" instead of
>> "migrate-to",  it effectively overrides and enforces the 1200ms default
>> value to the new value.
>>
>> I am wondering how I would have known that I was using the wrong
operation
>> name, when the initial operation name is already incorrect
>> when the resource is created?
>
> For SLES 11, I made a quick (portable non-portable unstable) try (print
> the operations known to an RA):
> # crm ra info VirtualDomain |sed -n -e "/Operations' defaults/,\$p"
> Operations' defaults (advisory minimum):
>
>start timeout=90
>stop  timeout=90
>statustimeout=30 interval=10
>monitor   timeout=30 interval=10
>migrate_from  timeout=60
>migrate_totimeout=120
>
> Regards,
> Ulrich
>
>>
>> This is what the meta data for my resource looked like after making the
>> update:
>>
>> [root@zs95kj VD]# date;pcs resource update zs95kjg110065_res op
migrate_to
>> timeout="360s"
>> Thu Jan 26 16:43:11 EST 2017
>> You have new mail in /var/spool/mail/root
>>
>> [root@zs95kj VD]# date;pcs resource show zs95kjg110065_res
>> Thu Jan 26 16:43:46 EST 2017
>>  Resource: zs95kjg110065_res (class=ocf provider=heartbeat
>> type=VirtualDomain)
>>   Attributes: config=/guestxml/nfs1/zs95kjg110065.xml
>> hypervisor=qemu:///system migration_transport=ssh
>>   Meta Attrs: allow-migrate=true
>>   Operations: start interval=0s timeout=120
>> (zs95kjg110065_res-start-interval-0s)
>>   stop interval=0s timeout=120
>> (zs95kjg110065_res-stop-interval-0s)
>>   monitor 

Re: [ClusterLabs] Live Guest Migration timeouts for VirtualDomain resources

2017-02-01 Thread Scott Greenlese

Hi all...

Just a quick follow-up.

Thought I should come clean and share with you that the incorrect
"migrate-to" operation name defined in my VirtualDomain
resource was my mistake.  It was mis-coded in the virtual guest
provisioning script.  I have since changed it to "migrate_to"
and of course, the specified live migration timeout value is working
effectively now.  (For some reason, I assumed we were letting that
operation meta value default).

I was wondering if someone could refer me to the definitive online link for
pacemaker resource man pages?  I don't see any resource man pages installed
on my system anywhere.   I found this one online:
https://www.mankier.com/7/ocf_heartbeat_VirtualDomain  but is there a more
'official' page I should refer our
Linux KVM on System z customers to?

Thanks again for your assistance.

Scott Greenlese ...IBM KVM on System Z Solution Test Poughkeepsie, N.Y.
  INTERNET:  swgre...@us.ibm.com




From:   "Ulrich Windl" 
To: , Scott Greenlese/Poughkeepsie/IBM@IBMUS
Cc: "Si Bo Niu" , Michael
Tebolt/Poughkeepsie/IBM@IBMUS
Date:   01/27/2017 02:32 AM
Subject:Antw: Re: [ClusterLabs] Antw: Re: Live Guest Migration timeouts
for VirtualDomain resources



>>> "Scott Greenlese"  schrieb am 27.01.2017 um 02:47
in
Nachricht
:

> Hi guys..
>
> Well, today I confirmed that what Ulrich said is correct.  If I update
the
> VirtualDomain resource with the operation name  "migrate_to" instead of
> "migrate-to",  it effectively overrides and enforces the 1200ms default
> value to the new value.
>
> I am wondering how I would have known that I was using the wrong
operation
> name, when the initial operation name is already incorrect
> when the resource is created?

For SLES 11, I made a quick (portable non-portable unstable) try (print the
operations known to an RA):
 # crm ra info VirtualDomain |sed -n -e "/Operations' defaults/,\$p"
Operations' defaults (advisory minimum):

start timeout=90
stop  timeout=90
statustimeout=30 interval=10
monitor   timeout=30 interval=10
migrate_from  timeout=60
migrate_totimeout=120

Regards,
Ulrich

>
> This is what the meta data for my resource looked like after making the
> update:
>
> [root@zs95kj VD]# date;pcs resource update zs95kjg110065_res op
migrate_to
> timeout="360s"
> Thu Jan 26 16:43:11 EST 2017
> You have new mail in /var/spool/mail/root
>
> [root@zs95kj VD]# date;pcs resource show zs95kjg110065_res
> Thu Jan 26 16:43:46 EST 2017
>  Resource: zs95kjg110065_res (class=ocf provider=heartbeat
> type=VirtualDomain)
>   Attributes: config=/guestxml/nfs1/zs95kjg110065.xml
> hypervisor=qemu:///system migration_transport=ssh
>   Meta Attrs: allow-migrate=true
>   Operations: start interval=0s timeout=120
> (zs95kjg110065_res-start-interval-0s)
>   stop interval=0s timeout=120
> (zs95kjg110065_res-stop-interval-0s)
>   monitor interval=30s
(zs95kjg110065_res-monitor-interval-30s)
>   migrate-from interval=0s timeout=1200
> (zs95kjg110065_res-migrate-from-interval-0s)
>   migrate-to interval=0s timeout=1200
> (zs95kjg110065_res-migrate-to-interval-0s)   <<< Original op name / value
>   migrate_to interval=0s timeout=360s
> (zs95kjg110065_res-migrate_to-interval-0s)  <<< New op name / value
>
>
> Where does that original op name come from in the VirtualDomain resource
> definition?  How can we get the initial meta value changed and shipped
with
> a valid operation name (i.e. migrate_to), and
> maybe a more reasonable migrate_to timeout value... something
significantly
> higher than 1200ms , i.e. 1.2 seconds ?  Can I report this request as a
> bugzilla on the RHEL side, or should this go to my internal IBM bugzilla
> for KVM on System Z development?
>
> Anyway, thanks so much for identifying my issue.  I can reconfigure my
> resources to make them tolerate longer migration execution times.
>
>
> Scott Greenlese ... IBM KVM on System Z Solution Test
>   INTERNET:  swgre...@us.ibm.com
>
>
>
>
> From:  Ken Gaillot 
> To:Ulrich Windl ,
> users@clusterlabs.org
> Date:  01/19/2017 10:26 AM
> Subject:   Re: [ClusterLabs] Antw: Re: Live Guest Migration
timeouts for
> VirtualDomain resources
>
>
>
> On 01/19/2017 01:36 AM, Ulrich Windl wrote:
>>>>> Ken Gaillot  schrieb am 18.01.2017 um 16:32 in
> Nachricht
>> <4b02d3fa-4693-473b-8bed-dc98f9e3f...@redhat.com>:
>>> On 01/17/2017 04:45 PM, Scott Greenlese wrote:
>>>> Ken and Co,
>>>>
>>>&g

Re: [ClusterLabs] Antw: Re: Live Guest Migration timeouts for VirtualDomain resources

2017-01-26 Thread Scott Greenlese

Hi guys..

Well, today I confirmed that what Ulrich said is correct.  If I update the
VirtualDomain resource with the operation name  "migrate_to" instead of
"migrate-to",  it effectively overrides and enforces the 1200ms default
value to the new value.

I am wondering how I would have known that I was using the wrong operation
name, when the initial operation name is already incorrect
when the resource is created?

This is what the meta data for my resource looked like after making the
update:

[root@zs95kj VD]# date;pcs resource update zs95kjg110065_res op migrate_to
timeout="360s"
Thu Jan 26 16:43:11 EST 2017
You have new mail in /var/spool/mail/root

[root@zs95kj VD]# date;pcs resource show zs95kjg110065_res
Thu Jan 26 16:43:46 EST 2017
 Resource: zs95kjg110065_res (class=ocf provider=heartbeat
type=VirtualDomain)
  Attributes: config=/guestxml/nfs1/zs95kjg110065.xml
hypervisor=qemu:///system migration_transport=ssh
  Meta Attrs: allow-migrate=true
  Operations: start interval=0s timeout=120
(zs95kjg110065_res-start-interval-0s)
  stop interval=0s timeout=120
(zs95kjg110065_res-stop-interval-0s)
  monitor interval=30s (zs95kjg110065_res-monitor-interval-30s)
  migrate-from interval=0s timeout=1200
(zs95kjg110065_res-migrate-from-interval-0s)
  migrate-to interval=0s timeout=1200
(zs95kjg110065_res-migrate-to-interval-0s)   <<< Original op name / value
  migrate_to interval=0s timeout=360s
(zs95kjg110065_res-migrate_to-interval-0s)  <<< New op name / value


Where does that original op name come from in the VirtualDomain resource
definition?  How can we get the initial meta value changed and shipped with
a valid operation name (i.e. migrate_to), and
maybe a more reasonable migrate_to timeout value... something significantly
higher than 1200ms , i.e. 1.2 seconds ?  Can I report this request as a
bugzilla on the RHEL side, or should this go to my internal IBM bugzilla
for KVM on System Z development?

Anyway, thanks so much for identifying my issue.  I can reconfigure my
resources to make them tolerate longer migration execution times.


Scott Greenlese ... IBM KVM on System Z Solution Test
  INTERNET:  swgre...@us.ibm.com




From:   Ken Gaillot 
To: Ulrich Windl ,
users@clusterlabs.org
Date:   01/19/2017 10:26 AM
Subject:Re: [ClusterLabs] Antw: Re: Live Guest Migration timeouts for
VirtualDomain resources



On 01/19/2017 01:36 AM, Ulrich Windl wrote:
>>>> Ken Gaillot  schrieb am 18.01.2017 um 16:32 in
Nachricht
> <4b02d3fa-4693-473b-8bed-dc98f9e3f...@redhat.com>:
>> On 01/17/2017 04:45 PM, Scott Greenlese wrote:
>>> Ken and Co,
>>>
>>> Thanks for the useful information.
>>>
>
> [...]
>>>
>>> Is this internally coded within the class=ocf provider=heartbeat
>>> type=VirtualDomain resource agent?
>>
>> Aha, I just realized what the issue is: the operation name is
>> migrate_to, not migrate-to.
>>
>> For technical reasons, pacemaker can't validate operation names (at the
>> time that the configuration is edited, it does not necessarily have
>> access to the agent metadata).
>
> BUT the set of operations is finite, right? So if those were in some XML
schema, the names could be verified at least (not meaning that the
operation is actually supported).
> BTW: Would a "crm configure verify" detect this kijnd of problem?
>
> [...]
>
> Ulrich

Yes, it's in the resource agent meta-data. While pacemaker itself uses a
small set of well-defined actions, the agent may define any arbitrarily
named actions it desires, and the user could configure one of these as a
recurring action in pacemaker.

Pacemaker itself has to be liberal about where its configuration comes
from -- the configuration can be edited on a separate machine, which
doesn't have resource agents, and then uploaded to the cluster. So
Pacemaker can't do that validation at configuration time. (It could
theoretically do some checking after the fact when the configuration is
loaded, but this could be a lot of overhead, and there are
implementation issues at the moment.)

Higher-level tools like crmsh and pcs, on the other hand, can make
simplifying assumptions. They can require access to the resource agents
so that they can do extra validation.

___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org



___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] Live Guest Migration timeouts for VirtualDomain resources

2017-01-17 Thread Scott Greenlese
 out, which occurred at 13:55:14 :

Jan 17 13:54:54 [27555] zs95kj   crmd: (   lrm.c:2005  )   debug:
do_lrm_rsc_op:Stopped 0 recurring operations in preparation for
zs95kjg110061_res_migrate_to_0
Jan 17 13:54:54 [27555] zs95kj   crmd: (   lrm.c:2009  )info:
do_lrm_rsc_op:Performing
key=337:574:0:43c68253-0e7d-4bb5-a78b-9b3c9af677dd
op=zs95kjg110061_res_migrate_to_0
Jan 17 13:54:54 [27555] zs95kj   crmd: (lrmd_client.:817   )   trace:
lrmd_send_command:sending lrmd_rsc_exec op to lrmd
Jan 17 13:54:54 [27555] zs95kj   crmd: (lrmd_client.:488   )   trace:
lrmd_create_op:   Sending call options: 0004, 4
Jan 17 13:54:54 [27552] zs95kj   lrmd: (  lrmd.c:1644  )   trace:
process_lrmd_message: Processing lrmd_rsc_exec operation from
a2a90d95-a766-42ba-84a0-54574c10adee
Jan 17 13:54:54 [27552] zs95kj   lrmd: (  lrmd.c:347   )   trace:
schedule_lrmd_cmd:Scheduling migrate_to on zs95kjg110061_res
Jan 17 13:54:54 [27552] zs95kj   lrmd: (  lrmd.c:1684  )   debug:
process_lrmd_message: Processed lrmd_rsc_exec operation from
a2a90d95-a766-42ba-84a0-54574c10adee: rc=941, reply=1, notify=0,
exit=-2147471352
Jan 17 13:54:54 [27552] zs95kj   lrmd: (  lrmd.c:131   )info:
log_execute:  executing - rsc:zs95kjg110061_res action:migrate_to
call_id:941
Jan 17 13:54:54 [27552] zs95kj   lrmd: (  lrmd.c:1184  )   trace:
lrmd_rsc_execute_service_lib: Creating action, resource:zs95kjg110061_res
action:migrate_to class:ocf provider:heartbeat agent:VirtualDomain
Jan 17 13:54:54 [27555] zs95kj   crmd: (lrmd_client.:846   )   trace:
lrmd_send_command:lrmd_rsc_exec op reply received
Jan 17 13:54:54 [27555] zs95kj   crmd: (lrmd_client.:852   )   trace:
lrmd_send_command:Reply   
Jan 17 13:54:54 [27555] zs95kj   crmd: (lrmd_client.:300   )   trace:
lrmd_dispatch_internal:   op lrmd_rsc_exec notify event received
Jan 17 13:54:54 [27555] zs95kj   crmd: (   lrm.c:2385  )info:
process_lrm_event:Operation zs95kjg110061_res_monitor_3: Cancelled
(node=zs95kjpcs1, call=936, confirmed=true)
Jan 17 13:54:54 [27555] zs95kj   crmd: (   lrm.c:196   )   debug:
update_history_cache: Updating history for 'zs95kjg110061_res' with monitor
op
VirtualDomain(zs95kjg110061_res)[135045]:   2017/01/17_13:54:54 DEBUG:
Virtual domain zs95kjg110061 is currently running.
VirtualDomain(zs95kjg110061_res)[135045]:   2017/01/17_13:54:54 INFO:
zs95kjg110061: Starting live migration to zs90kppcs1 (using: virsh
--connect=qemu:///system --quiet migrate --live  zs95kjg110061 qemu
+ssh://zs90kppcs1/system ).

Jan 17 13:54:54 [27550] zs95kjcib: (   xml.c:1728  )info:
cib_perform_op:   +  /cib/status/node_state[@id='2']/lrm
[@id='2']/lrm_resources/lr
m_resource[@id='zs95kjg110061_res']/lrm_rsc_op
[@id='zs95kjg110061_res_last_0']:
@operation_key=zs95kjg110061_res_migrate_to_0, @operation=migrate_to,
@crm-debu
g-origin=cib_action_update,
@transition-key=337:574:0:43c68253-0e7d-4bb5-a78b-9b3c9af677dd,
@transition-magic=-1:193;337:574:0:43c68253-0e7d-4bb5-a78b-9b3c9af67
7dd, @call-id=-1, @rc-code=193, @op-s


### 20 seconds after the live migration was initiated... we get the time
out:

Jan 17 13:55:14 [27552] zs95kj   lrmd: (  mainloop.c:949   ) warning:
child_timeout_callback:   zs95kjg110061_res_migrate_to_0 process (PID
135045) timed out
Jan 17 13:55:14 [27552] zs95kj   lrmd: (services_lin:314   ) warning:
operation_finished:   zs95kjg110061_res_migrate_to_0:135045 - timed out
after 2ms
Jan 17 13:55:14 [27552] zs95kj   lrmd: (services_lin:332   )   debug:
operation_finished:   zs95kjg110061_res_migrate_to_0:135045:stderr [ --
empty -- ]
Jan 17 13:55:14 [27552] zs95kj   lrmd: (services_lin:336   )   debug:
operation_finished:   zs95kjg110061_res_migrate_to_0:135045:stdout [ --
empty -- ]
Jan 17 13:55:14 [27552] zs95kj   lrmd: (  lrmd.c:592   )   trace:
cmd_finalize: Resource operation rsc:zs95kjg110061_res action:migrate_to
completed (0xa5f1edd0 0xa5f1edd0)
Jan 17 13:55:14 [27552] zs95kj   lrmd: (  lrmd.c:113   )info:
log_finished: finished - rsc:zs95kjg110061_res action:migrate_to
call_id:941 pid:135045 exit-code:1 exec-time:20003ms queue-time:0ms
Jan 17 13:55:14 [27552] zs95kj   lrmd: (  lrmd.c:1292  )   trace:
lrmd_rsc_execute: Nothing further to do for zs95kjg110061_res
Jan 17 13:55:14 [27555] zs95kj   crmd: ( utils.c:1942  )   debug:
create_operation_update:  do_update_resource: Updating resource
zs95kjg110061_res after migrate_to op Timed Out (interval=0)
Jan 17 13:55:14 [27555] zs95kj   crmd: (   lrm.c:2397  )   error:
process_lrm_event:Operation zs95kjg110061_res_migrate_to_0: Timed Out
(node=zs95kjpcs1, call=941, timeout=2ms)
Jan 17 13:55:14 [27555] zs95kj   crmd: (   lrm.c:196   )   debug:
update_history_cache: Updating history for 'zs95kjg110061_res&

[ClusterLabs] Live Guest Migration timeouts for VirtualDomain resources

2017-01-17 Thread Scott Greenlese
would really
prefer to have a good understanding of timeout configuration and recovery
behavior first.

Thanks!


Scott Greenlese ... IBM KVM on System z - Solution Test,  Poughkeepsie,
N.Y.
  INTERNET:  swgre...@us.ibm.com

___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] Pacemaker quorum behavior

2016-09-28 Thread Scott Greenlese

A quick addendum...

After sending this post, I decided to stop pacemaker on the single, Online
node in the cluster,
and this effectively killed the corosync daemon:

[root@zs93kl VD]# date;pcs cluster stop
Wed Sep 28 16:39:22 EDT 2016
Stopping Cluster (pacemaker)... Stopping Cluster (corosync)...


[root@zs93kl VD]# date;ps -ef |grep coro|grep -v grep
Wed Sep 28 16:46:19 EDT 2016
[root@zs93kl VD]#



Next, I went to a node in "Pending" state, and sure enough... the pcs
cluster stop killed the daemon there, too:

[root@zs95kj VD]# date;pcs cluster stop
Wed Sep 28 16:48:15 EDT 2016
Stopping Cluster (pacemaker)... Stopping Cluster (corosync)...

[root@zs95kj VD]# date;ps -ef |grep coro |grep -v grep
Wed Sep 28 16:48:38 EDT 2016
[root@zs95kj VD]#

So, this answers my own question...  cluster stop should kill corosync.
So, why isn't the `pcs cluster stop --all` failing to
kill corosync?

Thanks...


Scott Greenlese ... IBM KVM on System Z Test,  Poughkeepsie, N.Y.
  INTERNET:  swgre...@us.ibm.com





From:   Scott Greenlese/Poughkeepsie/IBM
To: kgail...@redhat.com, Cluster Labs - All topics related to
open-source clustering welcomed 
Date:   09/28/2016 04:30 PM
Subject:Re: [ClusterLabs] Pacemaker quorum behavior


Hi folks..

I have some follow-up questions about corosync daemon status after cluster
shutdown.

Basically, what should happen to corosync on a cluster node when pacemaker
is shutdown on that node?
On my 5 node cluster, when I do a global shutdown, the pacemaker processes
exit, but corosync processes remain active.

Here's an example of where this led me into some trouble...

My cluster is still configured to use the "symmetric" resource
distribution.   I don't have any location constraints in place, so
pacemaker tries to evenly distribute resources across all Online nodes.

With one cluster node (KVM host) powered off, I did the global cluster
stop:

[root@zs90KP VD]# date;pcs cluster stop --all
Wed Sep 28 15:07:40 EDT 2016
zs93KLpcs1: Unable to connect to zs93KLpcs1 ([Errno 113] No route to host)
zs90kppcs1: Stopping Cluster (pacemaker)...
zs95KLpcs1: Stopping Cluster (pacemaker)...
zs95kjpcs1: Stopping Cluster (pacemaker)...
zs93kjpcs1: Stopping Cluster (pacemaker)...
Error: unable to stop all nodes
zs93KLpcs1: Unable to connect to zs93KLpcs1 ([Errno 113] No route to host)

Note:  The "No route to host" messages are expected because that node /
LPAR is powered down.

(I don't show it here, but the corosync daemon is still running on the 4
active nodes. I do show it later).

I then powered on the one zs93KLpcs1 LPAR,  so in theory I should not have
quorum when it comes up and activates
pacemaker, which is enabled to autostart at boot time on all 5 cluster
nodes.  At this point, only 1 out of 5
nodes should be Online to the cluster, and therefore ... no quorum.

I login to zs93KLpcs1, and pcs status shows those 4 nodes as 'pending'
Online, and "partition with quorum":

[root@zs93kl ~]# date;pcs status |less
Wed Sep 28 15:25:13 EDT 2016
Cluster name: test_cluster_2
Last updated: Wed Sep 28 15:25:13 2016  Last change: Mon Sep 26
16:15:08 2016 by root via crm_resource on zs95kjpcs1
Stack: corosync
Current DC: zs93KLpcs1 (version 1.1.13-10.el7_2.ibm.1-44eb2dd) - partition
with quorum
106 nodes and 304 resources configured

Node zs90kppcs1: pending
Node zs93kjpcs1: pending
Node zs95KLpcs1: pending
Node zs95kjpcs1: pending
Online: [ zs93KLpcs1 ]

Full list of resources:

 zs95kjg109062_res  (ocf::heartbeat:VirtualDomain): Started zs93KLpcs1
 zs95kjg109063_res  (ocf::heartbeat:VirtualDomain): Started zs93KLpcs1
.
.
.


Here you can see that corosync is up on all 5 nodes:

[root@zs95kj VD]# date;for host in zs90kppcs1 zs95KLpcs1 zs95kjpcs1
zs93kjpcs1 zs93KLpcs1 ; do ssh $host "hostname;ps -ef |grep corosync |grep
-v grep"; done
Wed Sep 28 15:22:21 EDT 2016
zs90KP
root 155374  1  0 Sep26 ?00:10:17 corosync
zs95KL
root  22933  1  0 11:51 ?00:00:54 corosync
zs95kj
root  19382  1  0 Sep26 ?00:10:15 corosync
zs93kj
root 129102  1  0 Sep26 ?00:12:10 corosync
zs93kl
root  21894  1  0 15:19 ?00:00:00 corosync


But, pacemaker is only running on the one, online node:

[root@zs95kj VD]# date;for host in zs90kppcs1 zs95KLpcs1 zs95kjpcs1
zs93kjpcs1 zs93KLpcs1 ; do ssh $host "hostname;ps -ef |grep pacemakerd |
grep -v grep"; done
Wed Sep 28 15:23:29 EDT 2016
zs90KP
zs95KL
zs95kj
zs93kj
zs93kl
root  23005  1  0 15:19 ?00:00:00 /usr/sbin/pacemakerd -f
You have new mail in /var/spool/mail/root
[root@zs95kj VD]#


This situation wreaks havoc on my VirtualDomain resources, as the majority
of them are in FAILED or Stopped state, and to my
surprise... many of them show as Started:

[root@zs93kl VD]# date;pcs resource show |grep zs93KL
Wed Sep 28 15:55:29 EDT 2016
 zs95kjg109062_res  (ocf::heartbea

Re: [ClusterLabs] Pacemaker quorum behavior

2016-09-28 Thread Scott Greenlese
::heartbeat:VirtualDomain): Started zs93KLpcs1
 zs95kjg110122_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110123_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110124_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110125_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110126_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110128_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110129_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110130_res  (ocf::heartbeat:VirtualDomain): Started zs93KLpcs1
 zs95kjg110131_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110132_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110133_res  (ocf::heartbeat:VirtualDomain): Started zs93KLpcs1
 zs95kjg110134_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110135_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110137_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110138_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110139_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110140_res  (ocf::heartbeat:VirtualDomain): Started zs93KLpcs1
 zs95kjg110141_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110142_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110143_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110144_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110145_res  (ocf::heartbeat:VirtualDomain): Started zs93KLpcs1
 zs95kjg110146_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110148_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110149_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110150_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110152_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110154_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110155_res  (ocf::heartbeat:VirtualDomain): Started zs93KLpcs1
 zs95kjg110156_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110159_res  (ocf::heartbeat:VirtualDomain): Started zs93KLpcs1
 zs95kjg110160_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110161_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110164_res  (ocf::heartbeat:VirtualDomain): Started zs93KLpcs1
 zs95kjg110165_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
 zs95kjg110166_res  (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1


Pacemaker is attempting to activate all VirtualDomain resources on the one
cluster node.

So back to my original question... what should happen when I do a cluster
stop?
If it should be deactivating, what would prevent this?

Also,  I have tried simulating a failed cluster node (to trigger a STONITH
action) by killing the
corosync daemon on one node, but all that does is respawn the daemon ...
causing a temporary / transient
failure condition, and no fence takes place.   Is there a way to kill
corosync in such a way
that it stays down?   Is there a best practice for STONITH testing?

As usual, thanks in advance for your advice.

Scott Greenlese ... IBM KVM on System Z -  Solutions Test,  Poughkeepsie,
N.Y.
  INTERNET:  swgre...@us.ibm.com





From:   Ken Gaillot 
To: users@clusterlabs.org
Date:   09/09/2016 06:23 PM
Subject:Re: [ClusterLabs] Pacemaker quorum behavior



On 09/09/2016 04:27 AM, Klaus Wenninger wrote:
> On 09/08/2016 07:31 PM, Scott Greenlese wrote:
>>
>> Hi Klaus, thanks for your prompt and thoughtful feedback...
>>
>> Please see my answers nested below (sections entitled, "Scott's
>> Reply"). Thanks!
>>
>> - Scott
>>
>>
>> Scott Greenlese ... IBM Solutions Test, Poughkeepsie, N.Y.
>> INTERNET: swgre...@us.ibm.com
>> PHONE: 8/293-7301 (845-433-7301) M/S: POK 42HA/P966
>>
>>
>> Inactive hide details for Klaus Wenninger ---09/08/2016 10:59:27
>> AM---On 09/08/2016 03:55 PM, Scott Greenlese wrote: >Klaus Wenninger
>> ---09/08/2016 10:59:27 AM---On 09/08/2016 03:55 PM, Scott Greenlese
>> wrote: >
>>
>> From: Klaus Wenninger 
>> To: users@clusterlabs.org
>> Date: 09/08/2016 10:59 AM
>> Subject: Re: [ClusterLabs] Pacemaker quorum behavior
>>
>> 
>>
>>
>>
>> On 09/08/2016 03:55 PM, Scott Greenlese wrote:
>> >
>> > Hi all...
>> >
>> > I have a few very basic questions for the group.
>> >
>> > I have a 5 node (Linux on Z LPARs) pacemaker cluster with 100
>> > VirtualDomain pacemaker-remote nodes

Re: [ClusterLabs] "VirtualDomain is active on 2 nodes" due to transient network failure

2016-09-09 Thread Scott Greenlese

Hi Ken ,

Below where you commented,

"It's considered good practice to stop
pacemaker+corosync before rebooting a node intentionally (for even more
safety, you can put the node into standby first)."

.. is this something that we document anywhere?

Our 'reboot' action performs a halt (deactivate lpar) and then activate.
Do I run the risk
of guest instances running on multiple hosts in my case?  I'm performing
various recovery
scenarios and want to avoid this procedure (reboot without first stopping
cluster), if it's not supported.

By the way, I always put the node in cluster standby before an intentional
reboot.

Thanks!

Scott Greenlese ... IBM Solutions Test,  Poughkeepsie, N.Y.
  INTERNET:  swgre...@us.ibm.com
  PHONE:  8/293-7301 (845-433-7301)M/S:  POK 42HA/P966




From:   Ken Gaillot 
To: users@clusterlabs.org
Date:   09/02/2016 10:01 AM
Subject:Re: [ClusterLabs] "VirtualDomain is active on 2 nodes" due to
transient network failure



On 09/01/2016 09:39 AM, Scott Greenlese wrote:
> Andreas,
>
> You wrote:
>
> /"Would be good to see your full cluster configuration (corosync.conf
> and cib) - but first guess is: no fencing at all  and what is your
> "no-quorum-policy" in Pacemaker?/
>
> /Regards,/
> /Andreas"/
>
> Thanks for your interest. I actually do have a stonith device configured
> which maps all 5 cluster nodes in the cluster:
>
> [root@zs95kj ~]# date;pcs stonith show fence_S90HMC1
> Thu Sep 1 10:11:25 EDT 2016
> Resource: fence_S90HMC1 (class=stonith type=fence_ibmz)
> Attributes: ipaddr=9.12.35.134 login=stonith passwd=lnx4ltic
>
pcmk_host_map=zs95KLpcs1:S95/KVL;zs93KLpcs1:S93/KVL;zs93kjpcs1:S93/KVJ;zs95kjpcs1:S95/KVJ;zs90kppcs1:S90/PACEMAKER

> pcmk_host_list="zs95KLpcs1 zs93KLpcs1 zs93kjpcs1 zs95kjpcs1 zs90kppcs1"
> pcmk_list_timeout=300 pcmk_off_timeout=600 pcmk_reboot_action=off
> pcmk_reboot_timeout=600
> Operations: monitor interval=60s (fence_S90HMC1-monitor-interval-60s)
>
> This fencing device works, too well actually. It seems extremely
> sensitive to node "failures", and I'm not sure how to tune that. Stonith
> reboot actoin is 'off', and the general stonith action (cluster config)
> is also 'off'. In fact, often if I reboot a cluster node (i.e. reboot
> command) that is an active member in the cluster... stonith will power
> off that node while it's on its wait back up. (perhaps requires a
> separate issue thread on this forum?).

That depends on what a reboot does in your OS ... if it shuts down the
cluster services cleanly, you shouldn't get a fence, but if it kills
anything still running, then the cluster will see the node as failed,
and fencing is appropriate. It's considered good practice to stop
pacemaker+corosync before rebooting a node intentionally (for even more
safety, you can put the node into standby first).

>
> My no-quorum-policy is: no-quorum-policy: stop
>
> I don't think I should have lost quorum, only two of the five cluster
> nodes lost their corosync ring connection.

Those two nodes lost quorum, so they should have stopped all their
resources. And the three remaining nodes should have fenced them.

I'd check the logs around the time of the incident. Do the two affected
nodes detect the loss of quorum? Do they attempt to stop their
resources? Do those stops succeed? Do the other three nodes detect the
loss of the two nodes? Does the DC attempt to fence them? Do the fence
attempts succeed?

> Here's the full configuration:
>
>
> [root@zs95kj ~]# cat /etc/corosync/corosync.conf
> totem {
> version: 2
> secauth: off
> cluster_name: test_cluster_2
> transport: udpu
> }
>
> nodelist {
> node {
> ring0_addr: zs93kjpcs1
> nodeid: 1
> }
>
> node {
> ring0_addr: zs95kjpcs1
> nodeid: 2
> }
>
> node {
> ring0_addr: zs95KLpcs1
> nodeid: 3
> }
>
> node {
> ring0_addr: zs90kppcs1
> nodeid: 4
> }
>
> node {
> ring0_addr: zs93KLpcs1
> nodeid: 5
> }
> }
>
> quorum {
> provider: corosync_votequorum
> }
>
> logging {
> #Log to a specified file
> to_logfile: yes
> logfile: /var/log/corosync/corosync.log
> #Log timestamp as well
> timestamp: on
>
> #Facility in syslog
> syslog_facility: daemon
>
> logger_subsys {
> #Enable debug for this logger.
>
> debug: off
>
> #This specifies the subsystem identity (name) for which logging is
specified
>
> subsys: QUORUM
>
> }
> #Log to syslog
> to_syslog: yes
>
> #Whether or not turning on the debug information in the log
> debug: on
> }
> [root@zs95kj ~]#
>
>
>
> The full CIB (see attachment)
>
> [root@zs95kj ~]# pcs cluster cib > /tm

Re: [ClusterLabs] Pacemaker quorum behavior

2016-09-09 Thread Scott Greenlese
ualDomain): Started zs95kjpcs1
 zs95kjg109064_res  (ocf::heartbeat:VirtualDomain): Started zs95kjpcs1
 zs95kjg109065_res  (ocf::heartbeat:VirtualDomain): Started zs95kjpcs1
 zs95kjg109066_res  (ocf::heartbeat:VirtualDomain): Started zs95kjpcs1
 zs95kjg109067_res  (ocf::heartbeat:VirtualDomain): Started zs95kjpcs1
 zs95kjg109068_res  (ocf::heartbeat:VirtualDomain): Started zs95kjpcs1
.,
.
.
PCSD Status:
  zs93kjpcs1: Online
  zs95kjpcs1: Online
  zs95KLpcs1: Online
  zs90kppcs1: Offline
  zs93KLpcs1: Online


Check resources again:

Wed Sep  7 16:09:52 EDT 2016

 ### VirtualDomain Resource Statistics: ###

"_res" Virtual Domain resources:
  Started on zs95kj: 199
  Started on zs93kj: 0
  Started on zs95KL: 0
  Started on zs93KL: 0
  Started on zs90KP: 0
  Total Started: 199
  Total NOT Started: 1


I have since isolated all the corrupted virtual domain images and disabled
their VirtualDomain resources.
We already rebooted all five cluster nodes, after installing a new KVM
driver on them.

Now,  the quorum calculation and behavior seems to be working perfectly as
expected.

I started pacemaker on the nodes, one at a time... and, after 3 of the 5
nodes had pacemaker "Online" ...
resources activated and were evenly distributed across them.

In summary,  a lesson learned here is to check status of the pcs process to
be certain pacemaker and corosync
are indeed "offline" and that all threads to that process have terminated.
You had mentioned this command:

pstree -p | grep -A5 $(pidof -x pcs)

I'm not quite sure what the $(pidof -x pcs) represents??

On an "Online" cluster node, I see:

[root@zs93kj ~]# ps -ef |grep pcs |grep -v grep
root  18876  1  0 Sep07 ?
00:00:00 /bin/sh /usr/lib/pcsd/pcsd start
root  18905  18876  0 Sep07 ?00:00:00 /bin/bash -c ulimit -S -c
0 >/dev/null 2>&1 ; /usr/bin/ruby -I/usr/lib/pcsd /usr/lib/pcsd/ssl.rb
root  18906  18905  0 Sep07 ?00:04:22 /usr/bin/ruby
-I/usr/lib/pcsd /usr/lib/pcsd/ssl.rb
[root@zs93kj ~]#

If I use the 18876 PID on a healthy node, I get..

[root@zs93kj ~]# pstree -p |grep -A5 18876
   |-pcsd(18876)---bash(18905)---ruby(18906)-+-{ruby}(19102)
   | |-{ruby}(20212)
   | `-{ruby}(224258)
   |-pkcsslotd(18851)
   |-polkitd(19091)-+-{polkitd}(19100)
   ||-{polkitd}(19101)


Is this what you meant for me to do?If so, I'll be sure to do that next
time I suspect processes are not exiting on cluster kill or stop.

Thanks


Scott Greenlese ... IBM z/BX Solutions Test,  Poughkeepsie, N.Y.
  INTERNET:  swgre...@us.ibm.com
  PHONE:  8/293-7301 (845-433-7301)M/S:  POK 42HA/P966




From:   Jan Pokorný 
To: Cluster Labs - All topics related to open-source clustering
welcomed 
Cc: Si Bo Niu , Scott
Loveland/Poughkeepsie/IBM@IBMUS, Michael
Tebolt/Poughkeepsie/IBM@IBMUS
Date:   09/08/2016 02:43 PM
Subject:Re: [ClusterLabs] Pacemaker quorum behavior



On 08/09/16 10:20 -0400, Scott Greenlese wrote:
> Correction...
>
> When I stopped pacemaker/corosync on the four (powered on / active)
> cluster node hosts,  I was having an issue with the gentle method of
> stopping the cluster (pcs cluster stop --all),

Can you elaborate on what went wrong with this gentle method, please?

If it seemed to have stuck, you can perhaps run some diagnostics like:

  pstree -p | grep -A5 $(pidof -x pcs)

across the nodes to see if what process(es) pcs waits on, next time.

> so I ended up doing individual (pcs cluster kill ) on
> each of the four cluster nodes.   I then had to stop the virtual
> domains manually via 'virsh destroy ' on each host.
> Perhaps there was some residual node status affecting my quorum?

Hardly if corosync processes were indeed dead.

--
Jan (Poki)
[attachment "attyopgs.dat" deleted by Scott Greenlese/Poughkeepsie/IBM]
___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] Pacemaker quorum behavior

2016-09-08 Thread Scott Greenlese

Hi Klaus, thanks for your prompt and thoughtful feedback...

Please see my answers nested below (sections entitled, "Scott's Reply").
Thanks!

- Scott


Scott Greenlese ... IBM Solutions Test,  Poughkeepsie, N.Y.
  INTERNET:  swgre...@us.ibm.com
  PHONE:  8/293-7301 (845-433-7301)M/S:  POK 42HA/P966




From:   Klaus Wenninger 
To: users@clusterlabs.org
Date:   09/08/2016 10:59 AM
Subject:Re: [ClusterLabs] Pacemaker quorum behavior



On 09/08/2016 03:55 PM, Scott Greenlese wrote:
>
> Hi all...
>
> I have a few very basic questions for the group.
>
> I have a 5 node (Linux on Z LPARs) pacemaker cluster with 100
> VirtualDomain pacemaker-remote nodes
> plus 100 "opaque" VirtualDomain resources. The cluster is configured
> to be 'symmetric' and I have no
> location constraints on the 200 VirtualDomain resources (other than to
> prevent the opaque guests
> from running on the pacemaker remote node resources). My quorum is set
> as:
>
> quorum {
> provider: corosync_votequorum
> }
>
> As an experiment, I powered down one LPAR in the cluster, leaving 4
> powered up with the pcsd service up on the 4 survivors
> but corosync/pacemaker down (pcs cluster stop --all) on the 4
> survivors. I then started pacemaker/corosync on a single cluster
>

"pcs cluster stop" shuts down pacemaker & corosync on my test-cluster but
did you check the status of the individual services?

Scott's reply:

No, I only assumed that pacemaker was down because I got this back on my
pcs status
command from each cluster node:

[root@zs95kj VD]# date;for host in zs93KLpcs1 zs95KLpcs1 zs95kjpcs1
zs93kjpcs1 ; do ssh $host pcs status; done
Wed Sep  7 15:49:27 EDT 2016
Error: cluster is not currently running on this node
Error: cluster is not currently running on this node
Error: cluster is not currently running on this node
Error: cluster is not currently running on this node


What else should I check?  The pcsd.service service was still up, since I
didn't not stop that
anywhere. Should I have done,  ps -ef |grep -e pacemaker -e corosync  to
check the state before
assuming it was really down?




> node (pcs cluster start), and this resulted in the 200 VirtualDomain
> resources activating on the single node.
> This was not what I was expecting. I assumed that no resources would
> activate / start on any cluster nodes
> until 3 out of the 5 total cluster nodes had pacemaker/corosync running.
>
> After starting pacemaker/corosync on the single host (zs95kjpcs1),
> this is what I see :
>
> [root@zs95kj VD]# date;pcs status |less
> Wed Sep 7 15:51:17 EDT 2016
> Cluster name: test_cluster_2
> Last updated: Wed Sep 7 15:51:18 2016 Last change: Wed Sep 7 15:30:12
> 2016 by hacluster via crmd on zs93kjpcs1
> Stack: corosync
> Current DC: zs95kjpcs1 (version 1.1.13-10.el7_2.ibm.1-44eb2dd) -
> partition with quorum
> 106 nodes and 304 resources configured
>
> Node zs93KLpcs1: pending
> Node zs93kjpcs1: pending
> Node zs95KLpcs1: pending
> Online: [ zs95kjpcs1 ]
> OFFLINE: [ zs90kppcs1 ]
>
> .
> .
> .
> PCSD Status:
> zs93kjpcs1: Online
> zs95kjpcs1: Online
> zs95KLpcs1: Online
> zs90kppcs1: Offline
> zs93KLpcs1: Online
>
> So, what exactly constitutes an "Online" vs. "Offline" cluster node
> w.r.t. quorum calculation? Seems like in my case, it's "pending" on 3
> nodes,
> so where does that fall? Any why "pending"? What does that mean?
>
> Also, what exactly is the cluster's expected reaction to quorum loss?
> Cluster resources will be stopped or something else?
>
Depends on how you configure it using cluster property no-quorum-policy
(default: stop).

Scott's reply:

This is how the policy is configured:

[root@zs95kj VD]# date;pcs config |grep quorum
Thu Sep  8 13:18:33 EDT 2016
 no-quorum-policy: stop

What should I expect with the 'stop' setting?


>
>
> Where can I find this documentation?
>
http://clusterlabs.org/doc/en-US/Pacemaker/1.1/html/Pacemaker_Explained/

Scott's reply:

OK, I'll keep looking thru this doc, but I don't easily find the
no-quorum-policy explained.

Thanks..


>
>
> Thanks!
>
> Scott Greenlese - IBM Solution Test Team.
>
>
>
> Scott Greenlese ... IBM Solutions Test, Poughkeepsie, N.Y.
> INTERNET: swgre...@us.ibm.com
> PHONE: 8/293-7301 (845-433-7301) M/S: POK 42HA/P966
>
>
>
> ___
> Users mailing list: Users@clusterlabs.org
> http://clusterlabs.org/mailman/listinfo/users
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org



___

Re: [ClusterLabs] Pacemaker quorum behavior

2016-09-08 Thread Scott Greenlese

Correction...

When I stopped pacemaker/corosync on the four (powered on / active) cluster
node hosts,  I was having an issue with
the gentle method of stopping the cluster (pcs cluster stop --all), so I
ended up doing individual (pcs cluster kill )
on each of the four cluster nodes.   I then had to stop the virtual domains
manually via 'virsh destroy ' on each host.
Perhaps there was some residual node status affecting my quorum?

Thanks...

Scott Greenlese ... IBM Solutions Test,  Poughkeepsie, N.Y.
  INTERNET:  swgre...@us.ibm.com
  PHONE:  8/293-7301 (845-433-7301)M/S:  POK 42HA/P966




From:   Scott Greenlese/Poughkeepsie/IBM@IBMUS
To: users@clusterlabs.org
Cc: Si Bo Niu , Scott
Loveland/Poughkeepsie/IBM@IBMUS, Michael
Tebolt/Poughkeepsie/IBM@IBMUS
Date:   09/08/2016 10:01 AM
Subject:[ClusterLabs] Pacemaker quorum behavior



Hi all...

I have a few very basic questions for the group.

I have a 5 node (Linux on Z LPARs) pacemaker cluster with 100 VirtualDomain
pacemaker-remote nodes
plus 100 "opaque" VirtualDomain resources. The cluster is configured to be
'symmetric' and I have no
location constraints on the 200 VirtualDomain resources (other than to
prevent the opaque guests
from running on the pacemaker remote node resources). My quorum is set as:

quorum {
provider: corosync_votequorum
}

As an experiment, I powered down one LPAR in the cluster, leaving 4 powered
up with the pcsd service up on the 4 survivors
but corosync/pacemaker down (pcs cluster stop --all) on the 4 survivors. I
then started pacemaker/corosync on a single cluster
node (pcs cluster start), and this resulted in the 200 VirtualDomain
resources activating on the single node.
This was not what I was expecting. I assumed that no resources would
activate / start on any cluster nodes
until 3 out of the 5 total cluster nodes had pacemaker/corosync running.

After starting pacemaker/corosync on the single host (zs95kjpcs1), this is
what I see :

[root@zs95kj VD]# date;pcs status |less
Wed Sep 7 15:51:17 EDT 2016
Cluster name: test_cluster_2
Last updated: Wed Sep 7 15:51:18 2016 Last change: Wed Sep 7 15:30:12 2016
by hacluster via crmd on zs93kjpcs1
Stack: corosync
Current DC: zs95kjpcs1 (version 1.1.13-10.el7_2.ibm.1-44eb2dd) - partition
with quorum
106 nodes and 304 resources configured

Node zs93KLpcs1: pending
Node zs93kjpcs1: pending
Node zs95KLpcs1: pending
Online: [ zs95kjpcs1 ]
OFFLINE: [ zs90kppcs1 ]

.
.
.
PCSD Status:
zs93kjpcs1: Online
zs95kjpcs1: Online
zs95KLpcs1: Online
zs90kppcs1: Offline
zs93KLpcs1: Online

So, what exactly constitutes an "Online" vs. "Offline" cluster node w.r.t.
quorum calculation? Seems like in my case, it's "pending" on 3 nodes,
so where does that fall? Any why "pending"? What does that mean?

Also, what exactly is the cluster's expected reaction to quorum loss?
Cluster resources will be stopped or something else?

Where can I find this documentation?

Thanks!

Scott Greenlese - IBM Solution Test Team.



Scott Greenlese ... IBM Solutions Test, Poughkeepsie, N.Y.
INTERNET: swgre...@us.ibm.com
PHONE: 8/293-7301 (845-433-7301) M/S: POK 42HA/P966
___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[ClusterLabs] Pacemaker quorum behavior

2016-09-08 Thread Scott Greenlese

Hi all...

I have a few very basic questions for the group.

I have a 5 node (Linux on Z LPARs) pacemaker cluster with 100 VirtualDomain
pacemaker-remote nodes
plus 100 "opaque" VirtualDomain resources. The cluster is configured to be
'symmetric' and I have no
location constraints on the 200 VirtualDomain resources (other than to
prevent the opaque guests
from running on the pacemaker remote node resources).  My quorum is set as:

quorum {
provider: corosync_votequorum
}

As an experiment, I powered down one LPAR in the cluster, leaving 4 powered
up with the pcsd service up on the 4 survivors
but corosync/pacemaker down (pcs cluster stop --all) on the 4 survivors.
I then started pacemaker/corosync on a single cluster
node (pcs cluster start), and this resulted in the 200 VirtualDomain
resources activating on the single node.
This was not what I was expecting.  I assumed that no resources would
activate / start on any cluster nodes
until 3 out of the 5 total cluster nodes had pacemaker/corosync running.

After starting pacemaker/corosync on the single host (zs95kjpcs1), this is
what I see :

[root@zs95kj VD]# date;pcs status |less
Wed Sep  7 15:51:17 EDT 2016
Cluster name: test_cluster_2
Last updated: Wed Sep  7 15:51:18 2016  Last change: Wed Sep  7
15:30:12 2016 by hacluster via crmd on zs93kjpcs1
Stack: corosync
Current DC: zs95kjpcs1 (version 1.1.13-10.el7_2.ibm.1-44eb2dd) - partition
with quorum
106 nodes and 304 resources configured

Node zs93KLpcs1: pending
Node zs93kjpcs1: pending
Node zs95KLpcs1: pending
Online: [ zs95kjpcs1 ]
OFFLINE: [ zs90kppcs1 ]

.
.
.
PCSD Status:
  zs93kjpcs1: Online
  zs95kjpcs1: Online
  zs95KLpcs1: Online
  zs90kppcs1: Offline
  zs93KLpcs1: Online

So, what exactly constitutes an "Online" vs. "Offline" cluster node w.r.t.
quorum calculation?   Seems like in my case, it's "pending" on 3 nodes,
so where does that fall?   Any why "pending"?  What does that mean?

Also, what exactly is the cluster's expected reaction to quorum loss?
Cluster resources will be stopped or something else?

Where can I find this documentation?

Thanks!

Scott Greenlese -  IBM Solution Test Team.



Scott Greenlese ... IBM Solutions Test,  Poughkeepsie, N.Y.
  INTERNET:  swgre...@us.ibm.com
  PHONE:  8/293-7301 (845-433-7301)M/S:  POK 42HA/P966
___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] "VirtualDomain is active on 2 nodes" due to transient network failure

2016-08-31 Thread Scott Greenlese

Andreas Kurz,

I see that you replied to my post, but I can't seem to find any text
written from you??   Can you please
re-post your reply, or message me offline at swgre...@us.ibm.com?   I'm new
to the process, so
maybe something I did wrong in my post.

Thanks for your guidance.

Scott G.


Scott Greenlese ... IBM z/BX Solutions Test,  Poughkeepsie, N.Y.
  INTERNET:  swgre...@us.ibm.com
  PHONE:  8/293-7301 (845-433-7301)M/S:  POK 42HA/P966




From:   Andreas Kurz 
To: Cluster Labs - All topics related to open-source clustering
welcomed 
Date:   08/30/2016 05:06 PM
Subject:Re: [ClusterLabs] "VirtualDomain is active on 2 nodes" due to
transient network failure



Hi,

On Tue, Aug 30, 2016 at 10:03 PM, Scott Greenlese 
wrote:
  Added an appropriate subject line (was blank). Thanks...


  Scott Greenlese ... IBM z/BX Solutions Test, Poughkeepsie, N.Y.
  INTERNET: swgre...@us.ibm.com
  PHONE: 8/293-7301 (845-433-7301) M/S: POK 42HA/P966

  - Forwarded by Scott Greenlese/Poughkeepsie/IBM on 08/30/2016 03:59
  PM -

  From: Scott Greenlese/Poughkeepsie/IBM@IBMUS
  To: Cluster Labs - All topics related to open-source clustering welcomed
  
  Date: 08/29/2016 06:36 PM
  Subject: [ClusterLabs] (no subject)



  Hi folks,

  I'm assigned to system test Pacemaker/Corosync on the KVM on System Z
  platform
  with pacemaker-1.1.13-10 and corosync-2.3.4-7 .

Would be good to see your full cluster configuration (corosync.conf and
cib) - but first guess is: no fencing at all  and what is your
"no-quorum-policy" in Pacemaker?

Regards,
Andreas


  I have a cluster with 5 KVM hosts, and a total of 200
  ocf:pacemakerVirtualDomain resources defined to run
  across the 5 cluster nodes (symmertical is true for this cluster).

  The heartbeat network is communicating over vlan1293, which is hung off a
  network device, 0230 .

  In general, pacemaker does a good job of distributing my virtual guest
  resources evenly across the hypervisors
  in the cluster. These resource are a mixed bag:

  - "opaque" and remote "guest nodes" managed by the cluster.
  - allow-migrate=false and allow-migrate=true
  - qcow2 (file based) guests and LUN based guests
  - Sles and Ubuntu OS

  [root@zs95kj ]# pcs status |less
  Cluster name: test_cluster_2
  Last updated: Mon Aug 29 17:02:08 2016 Last change: Mon Aug 29 16:37:31
  2016 by root via crm_resource on zs93kjpcs1
  Stack: corosync
  Current DC: zs95kjpcs1 (version 1.1.13-10.el7_2.ibm.1-44eb2dd) -
  partition with quorum
  103 nodes and 300 resources configured

  Node zs90kppcs1: standby
  Online: [ zs93KLpcs1 zs93kjpcs1 zs95KLpcs1 zs95kjpcs1 ]

  This morning, our system admin team performed a
  "non-disruptive" (concurrent) microcode code load on the OSA, which
  (to our surprise) dropped the network connection for 13 seconds on the
  S93 CEC, from 11:18:34am to 11:18:47am , to be exact.
  This temporary outage caused the two cluster nodes on S93 (zs93kjpcs1 and
  zs93KLpcs1) to drop out of the cluster,
  as expected.

  However, pacemaker didn't handle this too well. The end result was
  numerous VirtualDomain resources in FAILED state:

  [root@zs95kj log]# date;pcs status |grep VirtualD |grep zs93 |grep FAILED
  Mon Aug 29 12:33:32 EDT 2016
  zs95kjg110104_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110092_res (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
  zs95kjg110099_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110102_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110106_res (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
  zs95kjg110112_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110115_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110118_res (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
  zs95kjg110124_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110127_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110130_res (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
  zs95kjg110136_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110139_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110142_res (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
  zs95kjg110148_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110152_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110155_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110161_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110164_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110167_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110173_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110176_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110179_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
  zs95kjg110185_res (ocf

[ClusterLabs] "VirtualDomain is active on 2 nodes" due to transient network failure

2016-08-30 Thread Scott Greenlese


Added an appropriate subject line (was blank).  Thanks...


Scott Greenlese ... IBM z/BX Solutions Test,  Poughkeepsie, N.Y.
  INTERNET:  swgre...@us.ibm.com
  PHONE:  8/293-7301 (845-433-7301)M/S:  POK 42HA/P966

- Forwarded by Scott Greenlese/Poughkeepsie/IBM on 08/30/2016 03:59 PM
-

From:   Scott Greenlese/Poughkeepsie/IBM@IBMUS
To: Cluster Labs - All topics related to open-source clustering
welcomed 
Date:   08/29/2016 06:36 PM
Subject:[ClusterLabs] (no subject)



Hi folks,

I'm assigned to system test Pacemaker/Corosync on the KVM on System Z
platform
with pacemaker-1.1.13-10 and corosync-2.3.4-7 .

I have a cluster with 5 KVM hosts, and a total of 200
ocf:pacemakerVirtualDomain resources defined to run
across the 5 cluster nodes (symmertical is true for this cluster).

The heartbeat network is communicating over vlan1293, which is hung off a
network device, 0230 .

In general, pacemaker does a good job of distributing my virtual guest
resources evenly across the hypervisors
in the cluster. These resource are a mixed bag:

- "opaque" and remote "guest nodes" managed by the cluster.
- allow-migrate=false and allow-migrate=true
- qcow2 (file based) guests and LUN based guests
- Sles and Ubuntu OS

[root@zs95kj ]# pcs status |less
Cluster name: test_cluster_2
Last updated: Mon Aug 29 17:02:08 2016 Last change: Mon Aug 29 16:37:31
2016 by root via crm_resource on zs93kjpcs1
Stack: corosync
Current DC: zs95kjpcs1 (version 1.1.13-10.el7_2.ibm.1-44eb2dd) - partition
with quorum
103 nodes and 300 resources configured

Node zs90kppcs1: standby
Online: [ zs93KLpcs1 zs93kjpcs1 zs95KLpcs1 zs95kjpcs1 ]

This morning, our system admin team performed a
"non-disruptive" (concurrent) microcode code load on the OSA, which
(to our surprise) dropped the network connection for 13 seconds on the S93
CEC, from 11:18:34am to 11:18:47am , to be exact.
This temporary outage caused the two cluster nodes on S93 (zs93kjpcs1 and
zs93KLpcs1) to drop out of the cluster,
as expected.

However, pacemaker didn't handle this too well. The end result was numerous
VirtualDomain resources in FAILED state:

[root@zs95kj log]# date;pcs status |grep VirtualD |grep zs93 |grep FAILED
Mon Aug 29 12:33:32 EDT 2016
zs95kjg110104_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110092_res (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
zs95kjg110099_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110102_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110106_res (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
zs95kjg110112_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110115_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110118_res (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
zs95kjg110124_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110127_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110130_res (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
zs95kjg110136_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110139_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110142_res (ocf::heartbeat:VirtualDomain): FAILED zs93KLpcs1
zs95kjg110148_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110152_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110155_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110161_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110164_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110167_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110173_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110176_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110179_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg110185_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1
zs95kjg109106_res (ocf::heartbeat:VirtualDomain): FAILED zs93kjpcs1


As well as, several VirtualDomain resources showing "Started" on two
cluster nodes:

zs95kjg110079_res (ocf::heartbeat:VirtualDomain): Started[ zs93kjpcs1
zs93KLpcs1 ]
zs95kjg110108_res (ocf::heartbeat:VirtualDomain): Started[ zs93kjpcs1
zs93KLpcs1 ]
zs95kjg110186_res (ocf::heartbeat:VirtualDomain): Started[ zs93kjpcs1
zs93KLpcs1 ]
zs95kjg110188_res (ocf::heartbeat:VirtualDomain): Started[ zs93kjpcs1
zs93KLpcs1 ]
zs95kjg110198_res (ocf::heartbeat:VirtualDomain): Started[ zs93kjpcs1
zs93KLpcs1 ]


The virtual machines themselves were in fact, "running" on both hosts. For
example:

[root@zs93kl ~]# virsh list |grep zs95kjg110079
70 zs95kjg110079 running

[root@zs93kj cli]# virsh list |grep zs95kjg110079
18 zs95kjg110079 running


On this particular VM, here was file corruption of this file-based qcow2
guest's image, such that you could not ping or ssh,
and if you open a virsh console, you get "initramfs" prompt.

To recover, we had to mount the volume on a

[ClusterLabs] (no subject)

2016-08-29 Thread Scott Greenlese
otice: [Action  429]: Pending pseudo
op zs95kjg110079_res_stop_0 on N/A (priority: 0, waiting:  460 491
517 542 567)
Aug 29 13:00:04 zs93kl crmd[20001]:  notice: [Action  428]: Completed rsc
op zs95kjg110079_res_stop_0  on zs93kjpcs1 (priority: 0, waiting:
none)




System log on zs93kjpcs1:


Aug 29 11:20:48 zs93kj crmd[21402]:  notice: Recurring action
zs95kjg110079_res:817 (zs95kjg110079_res_monitor_3) incomplete at
shutdown

Aug 29 11:22:07 zs93kj crmd[259639]:  notice: Operation
zs95kjg110079_res_monitor_0: ok (node=zs93kjpcs1, call=1223, rc=0,
cib-update=104, confirmed=true)

Aug 29 12:59:42 zs93kj VirtualDomain(zs95kjg110079_res)[9148]: INFO:
Issuing graceful shutdown request for domain zs95kjg110079.

Finally zs95kjg110079 shuts down on ZS93KJ at 12:59


===

Does this "active on two nodes" recovery process look right?

What is the recommended procedure to "undo" the resource failures and dual
host assignments?   It took several hours (short of stopping/starting the
entire cluster)
to recover them... resource disable, cleanup, enable was the basis ... but
it seemed that I would fix one resource and two more would fall out.

This seems to be one of the pitfalls of configuring resources in
symmetrical mode.

I would appreciate any best practice guidelines you have to offer.  I saved
the system logs on all hosts in case anyone needs more detailed
information.
I also have pacemaker.log logs.

Thanks in advance!



Scott Greenlese ... IBM z/BX Solutions Test,  Poughkeepsie, N.Y.
  INTERNET:  swgre...@us.ibm.com
  PHONE:  8/293-7301 (845-433-7301)M/S:  POK 42HA/P966
___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org