Re: [ClusterLabs] RES: RES: Can't get a group of IP address up when moving to a new version of Pacemaker/Corosync

2015-09-03 Thread Ken Gaillot
On 09/03/2015 01:48 PM, Carlos Xavier wrote:
> Tank you very much Ken.
> 
> I'll take a close look at that.
> Is there any documentation where I can learn more about the 
> Pacemaker/IPtables integration?
> The chain that needs to be created has the same name of the resource I'm 
> trying to clone?

Pacemaker itself is ignorant of the details. All the intelligence is
inside the IPadd2 RA. The RA's help text is the only documentation, but
the script itself isn't hard to read and is probably better for
understanding details.

For general information about the iptables feature, search online for
"iptables clusterip"

> Shouldn't the IPaddr2 RA solve this issue by itself? When it sees that there 
> is no Chain useful to add the rule, instead of giving a error create that 
> chain?

Yes, which is why I think the problem is at the iptables level and not
from the RA itself. I suspect if you figure out the exact commands
IPaddr2 would generate and run them manually, you'll run into a more
descriptive error.

> Regards,
> Carlos.
> 
> 
>> -Mensagem original-
>> De: Ken Gaillot [mailto:kgail...@redhat.com]
>> Enviada em: quarta-feira, 2 de setembro de 2015 16:16
>> Para: users@clusterlabs.org
>> Assunto: Re: [ClusterLabs] RES: Can't get a group of IP address up when 
>> moving to a new version of
>> Pacemaker/Corosync
>>
>> On 09/02/2015 01:44 PM, Carlos Xavier wrote:
>>> Hi Kristoffer.
>>>
>>> Tank you very much for fast reply.
>>>
>>> I did a cleanup of the resource and took a look at the log and could see 
>>> that the issue has
>> something to do with the RA IPaddr2 trying to set some iptables rule, 
>> although we are not using any
>> iptables rule set.
>>>
>>> crm(live)resource# cleanup c-ip-httpd
>>> Cleaning up ip_ccardbusiness:0 on apolo Cleaning up ip_ccardbusiness:0
>>> on diana Cleaning up ip_ccardgift:0 on apolo Cleaning up
>>> ip_ccardgift:0 on diana Cleaning up ip_intranet:0 on apolo Cleaning up
>>> ip_intranet:0 on diana Cleaning up ip_ccardbusiness:1 on apolo
>>> Cleaning up ip_ccardbusiness:1 on diana Cleaning up ip_ccardgift:1 on
>>> apolo Cleaning up ip_ccardgift:1 on diana Cleaning up ip_intranet:1 on
>>> apolo Cleaning up ip_intranet:1 on diana Waiting for 12 replies from
>>> the CRMd OK
>>>
>>> And on the log we have
>>>
>>> 2015-09-02T14:40:54.074834-03:00 apolo crmd[5185]:  warning:
>>> update_failcount: Updating failcount for ip_ccardbusiness on diana
>>> after failed start: rc=1 (update=INFINITY, time=1441215654)
>>> 2015-09-02T14:40:54.075034-03:00 apolo crmd[5185]:  warning: 
>>> status_from_rc: Action 33
>> (ip_ccardbusiness:0_start_0) on diana failed (target: 0 vs. rc: 1): Error 
>> 2015-09-02T14:40:54.075230-
>> 03:00 apolo crmd[5185]:  warning: update_failcount: Updating failcount for 
>> ip_ccardbusiness on diana
>> after failed start: rc=1 (update=INFINITY, time=1441215654) 
>> 2015-09-02T14:40:54.075427-03:00 apolo
>> crmd[5185]:  warning: update_failcount: Updating failcount for 
>> ip_ccardbusiness on diana after failed
>> start: rc=1 (update=INFINITY, time=1441215654)
>>> 2015-09-02T14:40:54.078344-03:00 apolo crmd[5185]:   notice: 
>>> abort_transition_graph: Transition
>> aborted by status-168427778-fail-count-ip_ccardbusiness, 
>> fail-count-ip_ccardbusiness=INFINITY:
>> Transient attribute change (create cib=0.378.14, source=te_update_diff:391,
>> path=/cib/status/node_state[@id='168427778']/transient_attributes[@id='168427778']/instance_attributes
>> [@id='status-168427778'], 0)
>>> 2015-09-02T14:40:54.184995-03:00 apolo IPaddr2(ip_ccardbusiness)[8360]: 
>>> ERROR: iptables failed
>>> 2015-09-02T14:40:54.187651-03:00 apolo lrmd[5182]:   notice: 
>>> operation_finished:
>> ip_ccardbusiness_start_0:8360:stderr [ iptables: No chain/target/match by 
>> that name. ]
>>> 2015-09-02T14:40:54.187978-03:00 apolo lrmd[5182]:   notice: 
>>> operation_finished:
>> ip_ccardbusiness_start_0:8360:stderr [ ocf-exit-reason:iptables failed ]
>>> 2015-09-02T14:40:54.203780-03:00 apolo crmd[5185]:   notice: 
>>> process_lrm_event: Operation
>> ip_ccardbusiness_start_0: unknown error (node=apolo, call=88, rc=1, 
>> cib-update=1321, confirmed=true)
>>> 2015-09-02T14:40:54.204026-03:00 apolo crmd[5185]:   notice: 
>>> process_lrm_event: apolo-
>> ip_ccardbusiness_start_0:88 [ iptables: No chain/target/match by that 
>> name.\nocf-exit-reason:iptables
>> failed\n ]
>>> 2015-09-02T14:40:54.206111-03:00 apolo crmd[5185]:  warning:
>>> status_from_rc: Action 43 (ip_ccardbusiness:1_start_0) on apolo failed
>>> (target: 0 vs. rc: 1): Error 2015-09-02T14:40:54.206442-03:00 apolo
>>> crmd[5185]:  warning: update_failcount: Updating failcount for
>>> ip_ccardbusiness on apolo after failed start: rc=1 (update=INFINITY,
>>> time=1441215654) 2015-09-02T14:40:54.206663-03:00 apolo crmd[5185]:  
>>> warning: update_failcount:
>> Updating failcount for ip_ccardbusiness on apolo after failed start: rc=1 
>> (update=INFINITY,
>> time=1441215654) 2015-09-02T14:40:54.206859-03:00 apolo crmd[518

Re: [ClusterLabs] RES: RES: Can't get a group of IP address up when moving to a new version of Pacemaker/Corosync

2015-09-03 Thread Victor José Acosta Domínguez
Thanks for answer, can you please tell me how much cost every extra IP?

Thanks

Regards

Victor Acosta

RHCE - RHCSA - VCA-DCV


2015-09-03 14:48 GMT-04:00 Carlos Xavier :

> Tank you very much Ken.
>
> I'll take a close look at that.
> Is there any documentation where I can learn more about the
> Pacemaker/IPtables integration?
> The chain that needs to be created has the same name of the resource I'm
> trying to clone?
>
> Shouldn't the IPaddr2 RA solve this issue by itself? When it sees that
> there is no Chain useful to add the rule, instead of giving a error create
> that chain?
>
> Regards,
> Carlos.
>
>
> > -Mensagem original-
> > De: Ken Gaillot [mailto:kgail...@redhat.com]
> > Enviada em: quarta-feira, 2 de setembro de 2015 16:16
> > Para: users@clusterlabs.org
> > Assunto: Re: [ClusterLabs] RES: Can't get a group of IP address up when
> moving to a new version of
> > Pacemaker/Corosync
> >
> > On 09/02/2015 01:44 PM, Carlos Xavier wrote:
> > > Hi Kristoffer.
> > >
> > > Tank you very much for fast reply.
> > >
> > > I did a cleanup of the resource and took a look at the log and could
> see that the issue has
> > something to do with the RA IPaddr2 trying to set some iptables rule,
> although we are not using any
> > iptables rule set.
> > >
> > > crm(live)resource# cleanup c-ip-httpd
> > > Cleaning up ip_ccardbusiness:0 on apolo Cleaning up ip_ccardbusiness:0
> > > on diana Cleaning up ip_ccardgift:0 on apolo Cleaning up
> > > ip_ccardgift:0 on diana Cleaning up ip_intranet:0 on apolo Cleaning up
> > > ip_intranet:0 on diana Cleaning up ip_ccardbusiness:1 on apolo
> > > Cleaning up ip_ccardbusiness:1 on diana Cleaning up ip_ccardgift:1 on
> > > apolo Cleaning up ip_ccardgift:1 on diana Cleaning up ip_intranet:1 on
> > > apolo Cleaning up ip_intranet:1 on diana Waiting for 12 replies from
> > > the CRMd OK
> > >
> > > And on the log we have
> > >
> > > 2015-09-02T14:40:54.074834-03:00 apolo crmd[5185]:  warning:
> > > update_failcount: Updating failcount for ip_ccardbusiness on diana
> > > after failed start: rc=1 (update=INFINITY, time=1441215654)
> > > 2015-09-02T14:40:54.075034-03:00 apolo crmd[5185]:  warning:
> status_from_rc: Action 33
> > (ip_ccardbusiness:0_start_0) on diana failed (target: 0 vs. rc: 1):
> Error 2015-09-02T14:40:54.075230-
> > 03:00 apolo crmd[5185]:  warning: update_failcount: Updating failcount
> for ip_ccardbusiness on diana
> > after failed start: rc=1 (update=INFINITY, time=1441215654)
> 2015-09-02T14:40:54.075427-03:00 apolo
> > crmd[5185]:  warning: update_failcount: Updating failcount for
> ip_ccardbusiness on diana after failed
> > start: rc=1 (update=INFINITY, time=1441215654)
> > > 2015-09-02T14:40:54.078344-03:00 apolo crmd[5185]:   notice:
> abort_transition_graph: Transition
> > aborted by status-168427778-fail-count-ip_ccardbusiness,
> fail-count-ip_ccardbusiness=INFINITY:
> > Transient attribute change (create cib=0.378.14,
> source=te_update_diff:391,
> >
> path=/cib/status/node_state[@id='168427778']/transient_attributes[@id='168427778']/instance_attributes
> > [@id='status-168427778'], 0)
> > > 2015-09-02T14:40:54.184995-03:00 apolo
> IPaddr2(ip_ccardbusiness)[8360]: ERROR: iptables failed
> > > 2015-09-02T14:40:54.187651-03:00 apolo lrmd[5182]:   notice:
> operation_finished:
> > ip_ccardbusiness_start_0:8360:stderr [ iptables: No chain/target/match
> by that name. ]
> > > 2015-09-02T14:40:54.187978-03:00 apolo lrmd[5182]:   notice:
> operation_finished:
> > ip_ccardbusiness_start_0:8360:stderr [ ocf-exit-reason:iptables failed ]
> > > 2015-09-02T14:40:54.203780-03:00 apolo crmd[5185]:   notice:
> process_lrm_event: Operation
> > ip_ccardbusiness_start_0: unknown error (node=apolo, call=88, rc=1,
> cib-update=1321, confirmed=true)
> > > 2015-09-02T14:40:54.204026-03:00 apolo crmd[5185]:   notice:
> process_lrm_event: apolo-
> > ip_ccardbusiness_start_0:88 [ iptables: No chain/target/match by that
> name.\nocf-exit-reason:iptables
> > failed\n ]
> > > 2015-09-02T14:40:54.206111-03:00 apolo crmd[5185]:  warning:
> > > status_from_rc: Action 43 (ip_ccardbusiness:1_start_0) on apolo failed
> > > (target: 0 vs. rc: 1): Error 2015-09-02T14:40:54.206442-03:00 apolo
> > > crmd[5185]:  warning: update_failcount: Updating failcount for
> > > ip_ccardbusiness on apolo after failed start: rc=1 (update=INFINITY,
> > > time=1441215654) 2015-09-02T14:40:54.206663-03:00 apolo crmd[5185]:
> warning: update_failcount:
> > Updating failcount for ip_ccardbusiness on apolo after failed start:
> rc=1 (update=INFINITY,
> > time=1441215654) 2015-09-02T14:40:54.206859-03:00 apolo crmd[5185]:
> warning: status_from_rc: Action
> > 43 (ip_ccardbusiness:1_start_0) on apolo failed (target: 0 vs. rc: 1):
> Error 2015-09-
> > 02T14:40:54.207109-03:00 apolo crmd[5185]:  warning: update_failcount:
> Updating failcount for
> > ip_ccardbusiness on apolo after failed start: rc=1 (update=INFINITY,
> time=1441215654) 2015-09-
> > 02T14:40:54.207489-03:00 apolo crmd[5185]

[ClusterLabs] RES: RES: Can't get a group of IP address up when moving to a new version of Pacemaker/Corosync

2015-09-03 Thread Carlos Xavier
Tank you very much Ken.

I'll take a close look at that.
Is there any documentation where I can learn more about the Pacemaker/IPtables 
integration?
The chain that needs to be created has the same name of the resource I'm trying 
to clone?

Shouldn't the IPaddr2 RA solve this issue by itself? When it sees that there is 
no Chain useful to add the rule, instead of giving a error create that chain?

Regards,
Carlos.


> -Mensagem original-
> De: Ken Gaillot [mailto:kgail...@redhat.com]
> Enviada em: quarta-feira, 2 de setembro de 2015 16:16
> Para: users@clusterlabs.org
> Assunto: Re: [ClusterLabs] RES: Can't get a group of IP address up when 
> moving to a new version of
> Pacemaker/Corosync
> 
> On 09/02/2015 01:44 PM, Carlos Xavier wrote:
> > Hi Kristoffer.
> >
> > Tank you very much for fast reply.
> >
> > I did a cleanup of the resource and took a look at the log and could see 
> > that the issue has
> something to do with the RA IPaddr2 trying to set some iptables rule, 
> although we are not using any
> iptables rule set.
> >
> > crm(live)resource# cleanup c-ip-httpd
> > Cleaning up ip_ccardbusiness:0 on apolo Cleaning up ip_ccardbusiness:0
> > on diana Cleaning up ip_ccardgift:0 on apolo Cleaning up
> > ip_ccardgift:0 on diana Cleaning up ip_intranet:0 on apolo Cleaning up
> > ip_intranet:0 on diana Cleaning up ip_ccardbusiness:1 on apolo
> > Cleaning up ip_ccardbusiness:1 on diana Cleaning up ip_ccardgift:1 on
> > apolo Cleaning up ip_ccardgift:1 on diana Cleaning up ip_intranet:1 on
> > apolo Cleaning up ip_intranet:1 on diana Waiting for 12 replies from
> > the CRMd OK
> >
> > And on the log we have
> >
> > 2015-09-02T14:40:54.074834-03:00 apolo crmd[5185]:  warning:
> > update_failcount: Updating failcount for ip_ccardbusiness on diana
> > after failed start: rc=1 (update=INFINITY, time=1441215654)
> > 2015-09-02T14:40:54.075034-03:00 apolo crmd[5185]:  warning: 
> > status_from_rc: Action 33
> (ip_ccardbusiness:0_start_0) on diana failed (target: 0 vs. rc: 1): Error 
> 2015-09-02T14:40:54.075230-
> 03:00 apolo crmd[5185]:  warning: update_failcount: Updating failcount for 
> ip_ccardbusiness on diana
> after failed start: rc=1 (update=INFINITY, time=1441215654) 
> 2015-09-02T14:40:54.075427-03:00 apolo
> crmd[5185]:  warning: update_failcount: Updating failcount for 
> ip_ccardbusiness on diana after failed
> start: rc=1 (update=INFINITY, time=1441215654)
> > 2015-09-02T14:40:54.078344-03:00 apolo crmd[5185]:   notice: 
> > abort_transition_graph: Transition
> aborted by status-168427778-fail-count-ip_ccardbusiness, 
> fail-count-ip_ccardbusiness=INFINITY:
> Transient attribute change (create cib=0.378.14, source=te_update_diff:391,
> path=/cib/status/node_state[@id='168427778']/transient_attributes[@id='168427778']/instance_attributes
> [@id='status-168427778'], 0)
> > 2015-09-02T14:40:54.184995-03:00 apolo IPaddr2(ip_ccardbusiness)[8360]: 
> > ERROR: iptables failed
> > 2015-09-02T14:40:54.187651-03:00 apolo lrmd[5182]:   notice: 
> > operation_finished:
> ip_ccardbusiness_start_0:8360:stderr [ iptables: No chain/target/match by 
> that name. ]
> > 2015-09-02T14:40:54.187978-03:00 apolo lrmd[5182]:   notice: 
> > operation_finished:
> ip_ccardbusiness_start_0:8360:stderr [ ocf-exit-reason:iptables failed ]
> > 2015-09-02T14:40:54.203780-03:00 apolo crmd[5185]:   notice: 
> > process_lrm_event: Operation
> ip_ccardbusiness_start_0: unknown error (node=apolo, call=88, rc=1, 
> cib-update=1321, confirmed=true)
> > 2015-09-02T14:40:54.204026-03:00 apolo crmd[5185]:   notice: 
> > process_lrm_event: apolo-
> ip_ccardbusiness_start_0:88 [ iptables: No chain/target/match by that 
> name.\nocf-exit-reason:iptables
> failed\n ]
> > 2015-09-02T14:40:54.206111-03:00 apolo crmd[5185]:  warning:
> > status_from_rc: Action 43 (ip_ccardbusiness:1_start_0) on apolo failed
> > (target: 0 vs. rc: 1): Error 2015-09-02T14:40:54.206442-03:00 apolo
> > crmd[5185]:  warning: update_failcount: Updating failcount for
> > ip_ccardbusiness on apolo after failed start: rc=1 (update=INFINITY,
> > time=1441215654) 2015-09-02T14:40:54.206663-03:00 apolo crmd[5185]:  
> > warning: update_failcount:
> Updating failcount for ip_ccardbusiness on apolo after failed start: rc=1 
> (update=INFINITY,
> time=1441215654) 2015-09-02T14:40:54.206859-03:00 apolo crmd[5185]:  warning: 
> status_from_rc: Action
> 43 (ip_ccardbusiness:1_start_0) on apolo failed (target: 0 vs. rc: 1): Error 
> 2015-09-
> 02T14:40:54.207109-03:00 apolo crmd[5185]:  warning: update_failcount: 
> Updating failcount for
> ip_ccardbusiness on apolo after failed start: rc=1 (update=INFINITY, 
> time=1441215654) 2015-09-
> 02T14:40:54.207489-03:00 apolo crmd[5185]:  warning: update_failcount: 
> Updating failcount for
> ip_ccardbusiness on apolo after failed start: rc=1 (update=INFINITY, 
> time=1441215654)
> > 2015-09-02T14:40:54.207829-03:00 apolo crmd[5185]:   notice: run_graph: 
> > Transition 1166
> (Complete=14, Pending=0, Fired=0, Skipp

Re: [ClusterLabs] HA Cluster and Fencing

2015-09-03 Thread Ken Gaillot
On 09/03/2015 11:44 AM, Streeter, Michelle N wrote:
> I was trying to get a HA Cluster working but it was not failing over.   In 
> past posts, someone kept asking me to get the fencing working and make it a 
> priority.  So I finally got the fencing to work with VBox.  And the fail over 
> finally started working for my HA cluster.   When I tried to explain this to 
> my lead, he didn't believe me that the fencing was the issue with the fail 
> over.   So, would someone help me understand why this happened so I can 
> explain it to my lead.   Also, when I was trying to get Pacemaker 1.1.11 
> working, it was failing over fine without the fencing but when I added more 
> than one drive to be serviced by the cluster via NFS.   The drives were being 
> serviced by  both nodes almost as if it was load balancing.  It was suggested 
> back then to get the fencing working.   So I take it if I went back to that 
> version, this would have fixed the issue.  Would you also help me explain why 
> this is true?
> 
> Michelle Streeter
> ASC2 MCS - SDE/ACL/SDL/EDL OKC Software Engineer
> The Boeing Company

Hi Michelle,

Congratulations on getting fencing working.

There's not enough information about your configuration to answer your
questions, but fencing is more a requirement for general cluster
stability rather than a solution to the specific problems you were facing.

Regarding load-balancing, I'm not sure whether you mean that a single
resource was started on multiple nodes, or different resources were
spread out on multiple nodes.

If one resource is active on multiple nodes, that means it was defined
as a clone or master-slave resource in your configuration. Clones are
used for active-active HA. If you want active-passive, where the
resource is only active on one node, don't clone it.

If instead you mean that multiple resources were spread out among nodes,
that's Pacemaker's default behavior. If you want two resources to always
be started together on the same node, you need to define a colocation
constraint for them (as well as an ordering constraint if one has to
start before the other), or put them in a resource group.

___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] HA Cluster and Fencing

2015-09-03 Thread Digimer
On 03/09/15 12:44 PM, Streeter, Michelle N wrote:
> I was trying to get a HA Cluster working but it was not failing over.  
> In past posts, someone kept asking me to get the fencing working and
> make it a priority.  So I finally got the fencing to work with VBox. 
> And the fail over finally started working for my HA cluster.   When I
> tried to explain this to my lead, he didn’t believe me that the fencing
> was the issue with the fail over.   So, would someone help me understand
> why this happened so I can explain it to my lead.   Also, when I was
> trying to get Pacemaker 1.1.11 working, it was failing over fine without
> the fencing but when I added more than one drive to be serviced by the
> cluster via NFS.   The drives were being serviced by  both nodes almost
> as if it was load balancing.  It was suggested back then to get the
> fencing working.   So I take it if I went back to that version, this
> would have fixed the issue.  Would you also help me explain why this is
> true?

That person was me.

It boils down to this;

If a service can safely run in two places at once, you don't need an HA
cluster. So if that is not the case, you need a resource manager to make
sure actions are coordinated between nodes. This fundamentally requires
understanding the state of each node.

If a node stops responding, no assumptions about its state are allowed.
The node must be put into a known state and that is where fencing comes
in. After a successful fence, all nodes enter a known state (with the
lost one being in an "off" state with power fencing and isolated in
fabric fencing).

With known states, the resource manager, pacemaker, can determine what
was lost (if anything), come up with a new configuration based on your
fail-over configuration, plot a course to reach that state and begin
recovery.

-- 
Digimer
Papers and Projects: https://alteeve.ca/w/
What if the cure for cancer is trapped in the mind of a person without
access to education?

___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] HA Cluster and Fencing

2015-09-03 Thread Kai Dupke
On 09/03/2015 06:44 PM, Streeter, Michelle N wrote:
>  Would you also help me explain why this is true?

Mind to give us more details? What exact setup? What was in the logs?

Just in case, clustering without fencing is a pretty bad idea in
general, and in special when data is not only read-only accessed by the
nodes.

greetings
Kai Dupke
Senior Product Manager
Server Product Line
-- 
Sell not virtue to purchase wealth, nor liberty to purchase power.
Phone:  +49-(0)5102-9310828 Mail: kdu...@suse.com
Mobile: +49-(0)173-5876766  WWW:  www.suse.com

SUSE Linux GmbH - Maxfeldstr. 5 - 90409 Nuernberg (Germany)
GF:Felix Imendörffer,Jane Smithard,Graham Norton,HRB 21284 (AG Nürnberg)

___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] HA Cluster and Fencing

2015-09-03 Thread David Gersic
What wasn't working about your failover?


Fencing is required for clustering, because it's how the cluster handles a 
failed node.



From: Streeter, Michelle N 
Sent: Thursday, September 03, 2015 11:44 AM
To: users@clusterlabs.org
Subject: [ClusterLabs] HA Cluster and Fencing


I was trying to get a HA Cluster working but it was not failing over.   In past 
posts, someone kept asking me to get the fencing working and make it a 
priority.  So I finally got the fencing to work with VBox.  And the fail over 
finally started working for my HA cluster.   When I tried to explain this to my 
lead, he didn't believe me that the fencing was the issue with the fail over.   
So, would someone help me understand why this happened so I can explain it to 
my lead.   Also, when I was trying to get Pacemaker 1.1.11 working, it was 
failing over fine without the fencing but when I added more than one drive to 
be serviced by the cluster via NFS.   The drives were being serviced by  both 
nodes almost as if it was load balancing.  It was suggested back then to get 
the fencing working.   So I take it if I went back to that version, this would 
have fixed the issue.  Would you also help me explain why this is true?



Michelle Streeter

ASC2 MCS - SDE/ACL/SDL/EDL OKC Software Engineer
The Boeing Company
___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[ClusterLabs] HA Cluster and Fencing

2015-09-03 Thread Streeter, Michelle N
I was trying to get a HA Cluster working but it was not failing over.   In past 
posts, someone kept asking me to get the fencing working and make it a 
priority.  So I finally got the fencing to work with VBox.  And the fail over 
finally started working for my HA cluster.   When I tried to explain this to my 
lead, he didn't believe me that the fencing was the issue with the fail over.   
So, would someone help me understand why this happened so I can explain it to 
my lead.   Also, when I was trying to get Pacemaker 1.1.11 working, it was 
failing over fine without the fencing but when I added more than one drive to 
be serviced by the cluster via NFS.   The drives were being serviced by  both 
nodes almost as if it was load balancing.  It was suggested back then to get 
the fencing working.   So I take it if I went back to that version, this would 
have fixed the issue.  Would you also help me explain why this is true?

Michelle Streeter
ASC2 MCS - SDE/ACL/SDL/EDL OKC Software Engineer
The Boeing Company
___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] fence-agents 4.0.21 release

2015-09-03 Thread Digimer
On 03/09/15 11:08 AM, Marek "marx" Grac wrote:
>Join us on IRC (irc.freenode.net #linux-cluster) and share your

Hi Marek,

  Might want to update this footer to #clusterlabs. :)

-- 
Digimer
Papers and Projects: https://alteeve.ca/w/
What if the cure for cancer is trapped in the mind of a person without
access to education?

___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[ClusterLabs] fence-agents 4.0.21 release

2015-09-03 Thread Marek "marx" Grac

Welcome to the fence-agents 4.0.21 release

This release includes several bugfixes:

* Documentation of new actions in manual pages is improved
* Tests were improved and Travis CI is now used for every push
* fence_virsh has new option --missing-as-off to support older XEN hosts
* fence_ilo(2) now automatically attempts to use TLS1.0 if it is not 
possible to connect via SSLv3
* fence_mpath re-create SCSI keys after reboot, monitoring of devices 
were improved too


Git repository can be found at https://github.com/ClusterLabs/fence-agents/

The new source tarball can be downloaded here:

https://github.com/ClusterLabs/fence-agents/archive/v4.0.21.tar.gz

To report bugs or issues:

https://bugzilla.redhat.com/

Would you like to meet the cluster team or members of its community?

   Join us on IRC (irc.freenode.net #linux-cluster) and share your
   experience  with other sysadministrators or power users.

Thanks/congratulations to all people that contributed to achieve this
great milestone.

m,

___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[ClusterLabs] [Announce] clufter-0.50.3 released

2015-09-03 Thread Jan Pokorný
I am happy to announce that clufter-0.50.3, a tool/library for
transforming/analyzing cluster configuration formats, has been
released and published (incl. signature using my 60BCBB4F5CD7F9EF key):


or alternative (original) location:



Changelog highlights:
- this is a bugfix release as previously instances of SAPDatabase and
  SAPInstance agents within cluster.conf were omitted in ccs2pcs*
  subcommands and *2pcscmd-needle subcommands might choke in
  corosync.conf part of the conversion if it involved its
  interface directive; both should be fixed now

Note that Github is no longer considered a primary origin and venue
for development/issue discussion.  It will still be available and kept
in sync, but the authoritative forge now is ,
service that offers a crucial advantage of free software infrastructure.

 * * *

The public repository (notably master and next branches) is currently at

(rather than ).

Official, signed releases can be found at
 or, alternatively, at

(also beware, automatic archives by GitHub preserve a "dev structure").

Natively packaged in Fedora (python-clufter, clufter-cli).

Issues & suggestions can be reported at either of (regardless if Fedora)
,

(rather than ).


Happy clustering/high-availing :)

-- 
Jan (Poki)


pgpSXiSscMObq.pgp
Description: PGP signature
___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org