Re: [Pacemaker] [Question]About "sequential" designation of resource_set.

2013-03-21 Thread renayama19661014
Hi Andrew,

Thank you for comments.

We demand time and the same movement that appointed ordered=false of the group 
resource.

 * Case 0 - group : orderded=false
  * At the time of orderded=false, it takes start of vip-rep. > We 
demand!!
{{{
(snip)

  

  


  



  


  



  

  
 
(snip)
[root@rh64-heartbeat1 ~]# grep "Initiating action" /var/log/ha-log
Mar 22 21:45:01 rh64-heartbeat1 crmd: [2625]: info: te_rsc_command: Initiating 
action 2: probe_complete probe_complete on rh64-heartbeat1 (local) - no waiting
Mar 22 21:46:36 rh64-heartbeat1 crmd: [2625]: info: te_rsc_command: Initiating 
action 4: monitor vip-master_monitor_0 on rh64-heartbeat1 (local)
Mar 22 21:46:36 rh64-heartbeat1 crmd: [2625]: info: te_rsc_command: Initiating 
action 5: monitor vip-rep_monitor_0 on rh64-heartbeat1 (local)
Mar 22 21:46:36 rh64-heartbeat1 crmd: [2625]: info: te_rsc_command: Initiating 
action 3: probe_complete probe_complete on rh64-heartbeat1 (local) - no waiting
Mar 22 21:46:36 rh64-heartbeat1 crmd: [2625]: info: te_rsc_command: Initiating 
action 6: start vip-master_start_0 on rh64-heartbeat1 (local)
Mar 22 21:46:36 rh64-heartbeat1 crmd: [2625]: info: te_rsc_command: Initiating 
action 8: start vip-rep_start_0 on rh64-heartbeat1 (local)
Mar 22 21:46:37 rh64-heartbeat1 crmd: [2625]: info: te_rsc_command: Initiating 
action 1: stop vip-master_stop_0 on rh64-heartbeat1 (local)
Mar 22 21:46:37 rh64-heartbeat1 crmd: [2625]: info: te_rsc_command: Initiating 
action 5: stop vip-rep_stop_0 on rh64-heartbeat1 (local)
}}}


I tried an all combination of ordering_set and colocation_set.
However, start of vip-rep was not carried out by all combinations.
 * I do not do the "ordered=false" designation of the group resource.

 * Case 1 : true/true
{{{
(snip)

  

  
(snip)
















[root@rh64-heartbeat1 ~]# grep "Initiating action" /var/log/ha-log
Mar 22 21:30:26 rh64-heartbeat1 crmd: [2076]: info: te_rsc_command: Initiating 
action 2: probe_complete probe_complete on rh64-heartbeat1 (local) - no waiting
Mar 22 21:33:36 rh64-heartbeat1 crmd: [2076]: info: te_rsc_command: Initiating 
action 8: monitor vip-master_monitor_0 on rh64-heartbeat1 (local)
Mar 22 21:33:36 rh64-heartbeat1 crmd: [2076]: info: te_rsc_command: Initiating 
action 9: monitor vip-rep_monitor_0 on rh64-heartbeat1 (local)
Mar 22 21:33:36 rh64-heartbeat1 crmd: [2076]: info: te_rsc_command: Initiating 
action 7: probe_complete probe_complete on rh64-heartbeat1 (local) - no waiting
Mar 22 21:33:36 rh64-heartbeat1 crmd: [2076]: info: te_rsc_command: Initiating 
action 10: start vip-master_start_0 on rh64-heartbeat1 (local)
Mar 22 21:33:36 rh64-heartbeat1 crmd: [2076]: info: te_rsc_command: Initiating 
action 1: stop vip-master_stop_0 on rh64-heartbeat1 (local)
}}}
 * Case 2 : true/false
{{{
(snip)

  

  
(snip)
















[root@rh64-heartbeat1 ~]# grep "Initiating action" /var/log/ha-log
Mar 22 21:36:38 rh64-heartbeat1 crmd: []: info: te_rsc_command: Initiating 
action 2: probe_complete probe_complete on rh64-heartbeat1 (local) - no waiting
Mar 22 21:36:49 rh64-heartbeat1 crmd: []: info: te_rsc_command: Initiating 
action 8: monitor vip-master_monitor_0 on rh64-heartbeat1 (local)
Mar 22 21:36:49 rh64-heartbeat1 crmd: []: info: te_rsc_command: Initiating 
action 9: monitor vip-rep_monitor_0 on rh64-heartbeat1 (local)
Mar 22 21:36:49 rh64-heartbeat1 crmd: []: info: te_rsc_command: Initiating 
action 7: probe_complete probe_complete on rh64-heartbeat1 (local) - no waiting
Mar 22 21:36:49 rh64-heartbeat1 crmd: []: info: te_rsc_command: Initiating 
action 10: start vip-master_start_0 on rh64-heartbeat1 (local)
Mar 22 21:36:50 rh64-heartbeat1 crmd: []: info: te_rsc_command: Initiating 
action 1: stop vip-master_stop_0 on rh64-heartbeat1 (local)

}}}
 * Cse 3 : false/true
{{{
(snip)

  

  
(snip)















[root@rh64-heartbeat1 ~]# grep "Initiating action" /var/log/ha-log
Mar 22 21:38:51 rh64-heartbeat1 crmd: [2358]: info: te_rsc_command: Initiating 
action 2: probe_complete probe_complete on rh64-heartbeat1 (local) - no waiting
Mar 22 21:39:07 rh64-heartbeat1 crmd: [2358]: info: te_rsc

Re: [Pacemaker] [Question]About "sequential" designation of resource_set.

2013-03-21 Thread Andrew Beekhof
On Fri, Mar 22, 2013 at 1:02 PM,   wrote:
> Hi Andrew,
>
> Thank your for comment.
>
>> Sorry, I'm not sure I understand the question.
>
> Sorry
>
> When we use resource_set in substitution for ordered of a thing of group 
> resource, do we use "colocation set"?
> Or do we use "ordering set "?

If you want ordering, use an ordering set.
If you want colocation, use a colocation set.

In both cases you want sequential=true.
"sequential=false" is only useful when there are multiple sets.

Ie. you want A and B colocated with C _but_ you want A to continue
running if B is stopped (and B to keep running if A is stopped).

Then you'd have two sets, { A, B } colocated with { C } where the A,B
set has sequential=false.

>
> It does not seem to do work same as "group ordered=fase" if it is right to 
> use "ordering set".

Correct.  A group with ordered=false is only providing colocation - so
you want a colocation set.

 A group with colocate=false is only providing ordering - so you'd
want an order set.

> Best Regards,
> Hideo Yamauchi.
>
> --- On Fri, 2013/3/22, Andrew Beekhof  wrote:
>
>> On Fri, Mar 22, 2013 at 12:34 PM,   wrote:
>> > Hi Andrew,
>> >
>> > Thank your for comments.
>> >
>> >> > > You use the resource sets _instead_ of a group.
>> >> > > If you want group.ordered=false, then use a colocation set (with
>> >> > > sequential=true).
>> >>
>> >> In "colocation", I used "resource_set".
>> >> However, a result did not include the change.
>> >> Will this result be a mistake of my setting?
>> >>
>> >> Case 1) sequential=false
>> >> (snip)
>> >> 
>> >>   
>> >> > >> id="test-colocation-resource_set">
>> >>   
>> >>   
>> >> 
>> >>   
>> >> 
>> >>
>> >> What are trying to achieve with this?  It doesn't do anything because 
>> >> there is nothing to collocate master or rep with.
>> >> The only value here is to show that rep would not be stopped when master 
>> >> is.
>> >
>> > However, you made next reply.
>> > I used colocation_set in substitution for ordered=false.
>> >
>> >>>You use the resource sets _instead_ of a group.
>> >>>If you want group.ordered=false, then use a colocation set (with
>> >>>sequential=true).
>> >>>If you want group.colocated=false, then use an ordering set (with
>> >>>sequential=true).
>> >
>> > After all is it right that the substitute for ordered=false of group sets 
>> > order_set?
>>
>> Sorry, I'm not sure I understand the question.
>>
>> >
>> > Best Regards,
>> > Hideo Yamauchi.
>> >
>> >
>> >
>> > --- On Fri, 2013/3/22, Andrew Beekhof  wrote:
>> >
>> >>
>> >>
>> >> On Thursday, March 7, 2013,   wrote:
>> >> Hi Andrew,
>> >>
>> >> > > You use the resource sets _instead_ of a group.
>> >> > > If you want group.ordered=false, then use a colocation set (with
>> >> > > sequential=true).
>> >>
>> >> In "colocation", I used "resource_set".
>> >> However, a result did not include the change.
>> >> Will this result be a mistake of my setting?
>> >>
>> >> Case 1) sequential=false
>> >> (snip)
>> >> 
>> >>   
>> >> > >> id="test-colocation-resource_set">
>> >>   
>> >>   
>> >> 
>> >>   
>> >> 
>> >>
>> >> What are trying to achieve with this?  It doesn't do anything because 
>> >> there is nothing to collocate master or rep with.
>> >> The only value here is to show that rep would not be stopped when master 
>> >> is.
>> >>  (sip)
>> >> [root@rh63-heartbeat2 ~]# grep "Initiating action" /var/log/ha-log
>> >> Mar  8 00:20:52 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> >> Initiating action 2: probe_complete probe_complete on rh63-heartbeat1 - 
>> >> no waiting
>> >> Mar  8 00:20:52 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> >> Initiating action 3: probe_complete probe_complete on rh63-heartbeat2 
>> >> (local) - no waiting
>> >> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> >> Initiating action 4: monitor vip-master_monitor_0 on rh63-heartbeat1
>> >> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> >> Initiating action 7: monitor vip-master_monitor_0 on rh63-heartbeat2 
>> >> (local)
>> >> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> >> Initiating action 5: monitor vip-rep_monitor_0 on rh63-heartbeat1
>> >> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> >> Initiating action 8: monitor vip-rep_monitor_0 on rh63-heartbeat2 (local)
>> >> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> >> Initiating action 6: probe_complete probe_complete on rh63-heartbeat2 
>> >> (local) - no waiting
>> >> Mar  8 00:20:56 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> >> Initiating action 3: probe_complete probe_complete on rh63-heartbeat1 - 
>> >> no waiting
>> >> Mar  8 00:20:56 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> >> Initiating action 5: start vip-master_start_0 on rh63-heartbeat1
>> >> Mar  8 00:20:58 rh63-heartbeat2 crmd: 

Re: [Pacemaker] pacemaker node stuck offline

2013-03-21 Thread pacemaker

-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

On 03/21/2013 11:15 AM, Andreas Kurz wrote:
> On 2013-03-21 14:31, Patrick Hemmer wrote:
>> I've got a 2-node cluster where it seems last night one of the nodes
>> went offline, and I can't see any reason why.
>>
>> Attached are the logs from the 2 nodes (the relevant timeframe seems to
>> be 2013-03-21 between 06:05 and 06:10).
>> This is on ubuntu 12.04
>
> Looks like your non-redundant cluster-communication was interrupted at
> around that time for whatever reason and your cluster split-brained.
>
> Does the drbd-replication use a different network-connection? If yes,
> why not using it for a redundant ring setup ... and you should use
STONITH.
>
> I also wonder why you have defined "expected_votes='1'" in your
> cluster.conf.
>
> Regards,
> Andreas
But shouldn't it have recovered? The node shows as "OFFLINE", even
though it's clearly communicating with the rest of the cluster. What is
the procedure for getting the node back online. Anything other than
bouncing pacemaker?

Unfortunately no to the different network connection for drbd. These are
2 EC2 instances, so redundant connections aren't available. Though since
it is EC2, I could set up a STONITH to whack the other instance. The
only problem here would be a race condition. The EC2 api for shutting
down or rebooting an instance isn't instantaneous. Both nodes could end
up sending the signal to reboot the other node.

As for expected_votes=1, it's because it's a two-node cluster. Though I
apparently forgot to set the `two_node` attribute :-(

- -Patrick
-BEGIN PGP SIGNATURE-
Version: GnuPG v2.0.19 (GNU/Linux)
Comment: Using GnuPG with Thunderbird - http://www.enigmail.net/

iQEcBAEBAgAGBQJRS8RSAAoJED0CF0ckHb4J5/4IAIBTh92ySD9NatBjanOtvwIZ
G7ldoPD/o//pOD8A76ZzJnbN+m5PQ1cykpwuC6j+l+fHbkYlDHYEnjbrdRS2dJFY
i1PibEIIOjeEAiK9PmCphKQ2qbkrKJXB0QdFD0EZjFFeatNfx/MBHInTBVdFa5MI
wZ19qcNELxHZHsrAfgFxYGzKvA1mCVZuRhFXpMoZJ9vo3RUFT1GaLbLA/k8+NHgQ
qPbmiYR0RI1cB+HqWl/Hn+PpWnV9zrF/vcZXISHp+cWpZ+IxzmDowR6iIHP+tC7N
AslkXAfz4BlH0cuM2kjA9ZdkApzGttH7GkMyOrOQ4Rv8rV4teQjMtPogMcqdFuc=
=lYXu
-END PGP SIGNATURE-


___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] [Question]About "sequential" designation of resource_set.

2013-03-21 Thread renayama19661014
Hi Andrew,

Thank your for comment.

> Sorry, I'm not sure I understand the question.

Sorry

When we use resource_set in substitution for ordered of a thing of group 
resource, do we use "colocation set"?
Or do we use "ordering set "?

It does not seem to do work same as "group ordered=fase" if it is right to use 
"ordering set".

Best Regards,
Hideo Yamauchi.

--- On Fri, 2013/3/22, Andrew Beekhof  wrote:

> On Fri, Mar 22, 2013 at 12:34 PM,   wrote:
> > Hi Andrew,
> >
> > Thank your for comments.
> >
> >> > > You use the resource sets _instead_ of a group.
> >> > > If you want group.ordered=false, then use a colocation set (with
> >> > > sequential=true).
> >>
> >> In "colocation", I used "resource_set".
> >> However, a result did not include the change.
> >> Will this result be a mistake of my setting?
> >>
> >> Case 1) sequential=false
> >> (snip)
> >>     
> >>       
> >>         
> >>           
> >>           
> >>         
> >>       
> >>     
> >>
> >> What are trying to achieve with this?  It doesn't do anything because 
> >> there is nothing to collocate master or rep with.
> >> The only value here is to show that rep would not be stopped when master 
> >> is.
> >
> > However, you made next reply.
> > I used colocation_set in substitution for ordered=false.
> >
> >>>You use the resource sets _instead_ of a group.
> >>>If you want group.ordered=false, then use a colocation set (with
> >>>sequential=true).
> >>>If you want group.colocated=false, then use an ordering set (with
> >>>sequential=true).
> >
> > After all is it right that the substitute for ordered=false of group sets 
> > order_set?
> 
> Sorry, I'm not sure I understand the question.
> 
> >
> > Best Regards,
> > Hideo Yamauchi.
> >
> >
> >
> > --- On Fri, 2013/3/22, Andrew Beekhof  wrote:
> >
> >>
> >>
> >> On Thursday, March 7, 2013,   wrote:
> >> Hi Andrew,
> >>
> >> > > You use the resource sets _instead_ of a group.
> >> > > If you want group.ordered=false, then use a colocation set (with
> >> > > sequential=true).
> >>
> >> In "colocation", I used "resource_set".
> >> However, a result did not include the change.
> >> Will this result be a mistake of my setting?
> >>
> >> Case 1) sequential=false
> >> (snip)
> >>     
> >>       
> >>         
> >>           
> >>           
> >>         
> >>       
> >>     
> >>
> >> What are trying to achieve with this?  It doesn't do anything because 
> >> there is nothing to collocate master or rep with.
> >> The only value here is to show that rep would not be stopped when master 
> >> is.
> >>  (sip)
> >> [root@rh63-heartbeat2 ~]# grep "Initiating action" /var/log/ha-log
> >> Mar  8 00:20:52 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> >> Initiating action 2: probe_complete probe_complete on rh63-heartbeat1 - no 
> >> waiting
> >> Mar  8 00:20:52 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> >> Initiating action 3: probe_complete probe_complete on rh63-heartbeat2 
> >> (local) - no waiting
> >> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> >> Initiating action 4: monitor vip-master_monitor_0 on rh63-heartbeat1
> >> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> >> Initiating action 7: monitor vip-master_monitor_0 on rh63-heartbeat2 
> >> (local)
> >> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> >> Initiating action 5: monitor vip-rep_monitor_0 on rh63-heartbeat1
> >> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> >> Initiating action 8: monitor vip-rep_monitor_0 on rh63-heartbeat2 (local)
> >> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> >> Initiating action 6: probe_complete probe_complete on rh63-heartbeat2 
> >> (local) - no waiting
> >> Mar  8 00:20:56 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> >> Initiating action 3: probe_complete probe_complete on rh63-heartbeat1 - no 
> >> waiting
> >> Mar  8 00:20:56 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> >> Initiating action 5: start vip-master_start_0 on rh63-heartbeat1
> >> Mar  8 00:20:58 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> >> Initiating action 1: stop vip-master_stop_0 on rh63-heartbeat1
> >>
> >>
> >> Case 2) sequential=true
> >> (snip)
> >>     
> >>       
> >>         
> >>           
> >>           
> >>         
> >>       
> >>     
> >> (snip)
> >> [root@rh63-heartbeat2 ~]# grep "Initiating action" /var/log/ha-log
> >> Mar  7 23:54:44 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> >> Initiating action 2: probe_complete probe_complete on rh63-heartbeat1 - no 
> >> waiting
> >> Mar  7 23:54:44 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> >> Initiating action 3: probe_complete probe_complete on rh63-heartbeat2 
> >> (local) - no waiting
> >> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> >> Initiating action 4: monitor vip-master_monitor_0 on rh63-heartbeat1
> >> Mar  7 23:54:48 rh63-heart

Re: [Pacemaker] [Question]About "sequential" designation of resource_set.

2013-03-21 Thread Andrew Beekhof
On Fri, Mar 22, 2013 at 12:34 PM,   wrote:
> Hi Andrew,
>
> Thank your for comments.
>
>> > > You use the resource sets _instead_ of a group.
>> > > If you want group.ordered=false, then use a colocation set (with
>> > > sequential=true).
>>
>> In "colocation", I used "resource_set".
>> However, a result did not include the change.
>> Will this result be a mistake of my setting?
>>
>> Case 1) sequential=false
>> (snip)
>> 
>>   
>> 
>>   
>>   
>> 
>>   
>> 
>>
>> What are trying to achieve with this?  It doesn't do anything because there 
>> is nothing to collocate master or rep with.
>> The only value here is to show that rep would not be stopped when master is.
>
> However, you made next reply.
> I used colocation_set in substitution for ordered=false.
>
>>>You use the resource sets _instead_ of a group.
>>>If you want group.ordered=false, then use a colocation set (with
>>>sequential=true).
>>>If you want group.colocated=false, then use an ordering set (with
>>>sequential=true).
>
> After all is it right that the substitute for ordered=false of group sets 
> order_set?

Sorry, I'm not sure I understand the question.

>
> Best Regards,
> Hideo Yamauchi.
>
>
>
> --- On Fri, 2013/3/22, Andrew Beekhof  wrote:
>
>>
>>
>> On Thursday, March 7, 2013,   wrote:
>> Hi Andrew,
>>
>> > > You use the resource sets _instead_ of a group.
>> > > If you want group.ordered=false, then use a colocation set (with
>> > > sequential=true).
>>
>> In "colocation", I used "resource_set".
>> However, a result did not include the change.
>> Will this result be a mistake of my setting?
>>
>> Case 1) sequential=false
>> (snip)
>> 
>>   
>> 
>>   
>>   
>> 
>>   
>> 
>>
>> What are trying to achieve with this?  It doesn't do anything because there 
>> is nothing to collocate master or rep with.
>> The only value here is to show that rep would not be stopped when master is.
>>  (sip)
>> [root@rh63-heartbeat2 ~]# grep "Initiating action" /var/log/ha-log
>> Mar  8 00:20:52 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> Initiating action 2: probe_complete probe_complete on rh63-heartbeat1 - no 
>> waiting
>> Mar  8 00:20:52 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> Initiating action 3: probe_complete probe_complete on rh63-heartbeat2 
>> (local) - no waiting
>> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> Initiating action 4: monitor vip-master_monitor_0 on rh63-heartbeat1
>> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> Initiating action 7: monitor vip-master_monitor_0 on rh63-heartbeat2 (local)
>> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> Initiating action 5: monitor vip-rep_monitor_0 on rh63-heartbeat1
>> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> Initiating action 8: monitor vip-rep_monitor_0 on rh63-heartbeat2 (local)
>> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> Initiating action 6: probe_complete probe_complete on rh63-heartbeat2 
>> (local) - no waiting
>> Mar  8 00:20:56 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> Initiating action 3: probe_complete probe_complete on rh63-heartbeat1 - no 
>> waiting
>> Mar  8 00:20:56 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> Initiating action 5: start vip-master_start_0 on rh63-heartbeat1
>> Mar  8 00:20:58 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
>> Initiating action 1: stop vip-master_stop_0 on rh63-heartbeat1
>>
>>
>> Case 2) sequential=true
>> (snip)
>> 
>>   
>> 
>>   
>>   
>> 
>>   
>> 
>> (snip)
>> [root@rh63-heartbeat2 ~]# grep "Initiating action" /var/log/ha-log
>> Mar  7 23:54:44 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
>> Initiating action 2: probe_complete probe_complete on rh63-heartbeat1 - no 
>> waiting
>> Mar  7 23:54:44 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
>> Initiating action 3: probe_complete probe_complete on rh63-heartbeat2 
>> (local) - no waiting
>> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
>> Initiating action 4: monitor vip-master_monitor_0 on rh63-heartbeat1
>> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
>> Initiating action 7: monitor vip-master_monitor_0 on rh63-heartbeat2 (local)
>> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
>> Initiating action 5: monitor vip-rep_monitor_0 on rh63-heartbeat1
>> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
>> Initiating action 8: monitor vip-rep_monitor_0 on rh63-heartbeat2 (local)
>> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
>> Initiating action 6: probe_complete probe_complete on rh63-heartbeat2 
>> (local) - no waiting
>> Mar  7 23:54:49 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
>> Initiating action 3: pr

Re: [Pacemaker] [Question]About "sequential" designation of resource_set.

2013-03-21 Thread renayama19661014
Hi Andrew,

Thank your for comments.
 
> > > You use the resource sets _instead_ of a group.
> > > If you want group.ordered=false, then use a colocation set (with
> > > sequential=true).
> 
> In "colocation", I used "resource_set".
> However, a result did not include the change.
> Will this result be a mistake of my setting?
> 
> Case 1) sequential=false
> (snip)
> 
>   
> 
>   
>   
> 
>   
> 
> 
> What are trying to achieve with this?  It doesn't do anything because there 
> is nothing to collocate master or rep with.
> The only value here is to show that rep would not be stopped when master is. 

However, you made next reply.
I used colocation_set in substitution for ordered=false.

>>You use the resource sets _instead_ of a group. 
>>If you want group.ordered=false, then use a colocation set (with 
>>sequential=true). 
>>If you want group.colocated=false, then use an ordering set (with 
>>sequential=true). 

After all is it right that the substitute for ordered=false of group sets 
order_set?

Best Regards,
Hideo Yamauchi.



--- On Fri, 2013/3/22, Andrew Beekhof  wrote:

> 
> 
> On Thursday, March 7, 2013,   wrote:
> Hi Andrew,
> 
> > > You use the resource sets _instead_ of a group.
> > > If you want group.ordered=false, then use a colocation set (with
> > > sequential=true).
> 
> In "colocation", I used "resource_set".
> However, a result did not include the change.
> Will this result be a mistake of my setting?
> 
> Case 1) sequential=false
> (snip)
>     
>       
>         
>           
>           
>         
>       
>     
> 
> What are trying to achieve with this?  It doesn't do anything because there 
> is nothing to collocate master or rep with.
> The only value here is to show that rep would not be stopped when master is. 
>  (sip)
> [root@rh63-heartbeat2 ~]# grep "Initiating action" /var/log/ha-log
> Mar  8 00:20:52 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> Initiating action 2: probe_complete probe_complete on rh63-heartbeat1 - no 
> waiting
> Mar  8 00:20:52 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> Initiating action 3: probe_complete probe_complete on rh63-heartbeat2 (local) 
> - no waiting
> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> Initiating action 4: monitor vip-master_monitor_0 on rh63-heartbeat1
> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> Initiating action 7: monitor vip-master_monitor_0 on rh63-heartbeat2 (local)
> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> Initiating action 5: monitor vip-rep_monitor_0 on rh63-heartbeat1
> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> Initiating action 8: monitor vip-rep_monitor_0 on rh63-heartbeat2 (local)
> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> Initiating action 6: probe_complete probe_complete on rh63-heartbeat2 (local) 
> - no waiting
> Mar  8 00:20:56 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> Initiating action 3: probe_complete probe_complete on rh63-heartbeat1 - no 
> waiting
> Mar  8 00:20:56 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> Initiating action 5: start vip-master_start_0 on rh63-heartbeat1
> Mar  8 00:20:58 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> Initiating action 1: stop vip-master_stop_0 on rh63-heartbeat1
> 
> 
> Case 2) sequential=true
> (snip)
>     
>       
>         
>           
>           
>         
>       
>     
> (snip)
> [root@rh63-heartbeat2 ~]# grep "Initiating action" /var/log/ha-log
> Mar  7 23:54:44 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> Initiating action 2: probe_complete probe_complete on rh63-heartbeat1 - no 
> waiting
> Mar  7 23:54:44 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> Initiating action 3: probe_complete probe_complete on rh63-heartbeat2 (local) 
> - no waiting
> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> Initiating action 4: monitor vip-master_monitor_0 on rh63-heartbeat1
> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> Initiating action 7: monitor vip-master_monitor_0 on rh63-heartbeat2 (local)
> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> Initiating action 5: monitor vip-rep_monitor_0 on rh63-heartbeat1
> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> Initiating action 8: monitor vip-rep_monitor_0 on rh63-heartbeat2 (local)
> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> Initiating action 6: probe_complete probe_complete on rh63-heartbeat2 (local) 
> - no waiting
> Mar  7 23:54:49 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> Initiating action 3: probe_complete probe_complete on rh63-heartbeat1 - no 
> waiting
> Mar  7 23:54:49 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> Initiating action 5: start vip-master_start_0 on rh63-heartbeat1
> Mar  7 23:

Re: [Pacemaker] [Question]About "sequential" designation of resource_set.

2013-03-21 Thread Andrew Beekhof
On Thursday, March 7, 2013, wrote:

> Hi Andrew,
>
> > > You use the resource sets _instead_ of a group.
> > > If you want group.ordered=false, then use a colocation set (with
> > > sequential=true).
>
> In "colocation", I used "resource_set".
> However, a result did not include the change.


> Will this result be a mistake of my setting?
>
> Case 1) sequential=false
> (snip)
> 
>   
> 
>   
>   
> 
>   
> 


What are trying to achieve with this?  It doesn't do anything because there
is nothing to collocate master or rep with.
The only value here is to show that rep would not be stopped when master
is.


> (sip)
> [root@rh63-heartbeat2 ~]# grep "Initiating action" /var/log/ha-log
> Mar  8 00:20:52 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command:
> Initiating action 2: probe_complete probe_complete on rh63-heartbeat1 - no
> waiting
> Mar  8 00:20:52 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command:
> Initiating action 3: probe_complete probe_complete on rh63-heartbeat2
> (local) - no waiting
> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command:
> Initiating action 4: monitor vip-master_monitor_0 on rh63-heartbeat1
> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command:
> Initiating action 7: monitor vip-master_monitor_0 on rh63-heartbeat2 (local)
> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command:
> Initiating action 5: monitor vip-rep_monitor_0 on rh63-heartbeat1
> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command:
> Initiating action 8: monitor vip-rep_monitor_0 on rh63-heartbeat2 (local)
> Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command:
> Initiating action 6: probe_complete probe_complete on rh63-heartbeat2
> (local) - no waiting
> Mar  8 00:20:56 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command:
> Initiating action 3: probe_complete probe_complete on rh63-heartbeat1 - no
> waiting
> Mar  8 00:20:56 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command:
> Initiating action 5: start vip-master_start_0 on rh63-heartbeat1
> Mar  8 00:20:58 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command:
> Initiating action 1: stop vip-master_stop_0 on rh63-heartbeat1
>
>
> Case 2) sequential=true
> (snip)
> 
>   
> 
>   
>   
> 
>   
> 
> (snip)
> [root@rh63-heartbeat2 ~]# grep "Initiating action" /var/log/ha-log
> Mar  7 23:54:44 rh63-heartbeat2 crmd: [4]: info: te_rsc_command:
> Initiating action 2: probe_complete probe_complete on rh63-heartbeat1 - no
> waiting
> Mar  7 23:54:44 rh63-heartbeat2 crmd: [4]: info: te_rsc_command:
> Initiating action 3: probe_complete probe_complete on rh63-heartbeat2
> (local) - no waiting
> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command:
> Initiating action 4: monitor vip-master_monitor_0 on rh63-heartbeat1
> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command:
> Initiating action 7: monitor vip-master_monitor_0 on rh63-heartbeat2 (local)
> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command:
> Initiating action 5: monitor vip-rep_monitor_0 on rh63-heartbeat1
> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command:
> Initiating action 8: monitor vip-rep_monitor_0 on rh63-heartbeat2 (local)
> Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command:
> Initiating action 6: probe_complete probe_complete on rh63-heartbeat2
> (local) - no waiting
> Mar  7 23:54:49 rh63-heartbeat2 crmd: [4]: info: te_rsc_command:
> Initiating action 3: probe_complete probe_complete on rh63-heartbeat1 - no
> waiting
> Mar  7 23:54:49 rh63-heartbeat2 crmd: [4]: info: te_rsc_command:
> Initiating action 5: start vip-master_start_0 on rh63-heartbeat1
> Mar  7 23:54:51 rh63-heartbeat2 crmd: [4]: info: te_rsc_command:
> Initiating action 1: stop vip-master_stop_0 on rh63-heartbeat1
>
>
You set target-role=stopped for master but rep did not stop?
If so that is certainly a bug.


>
> Best Regards,
> Hideo Yamauchi.
>
>
> ___
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org 
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] [Question]About "sequential" designation of resource_set.

2013-03-21 Thread renayama19661014
Hi Andrew,

I registered this question with Bugzilla.

 * http://bugs.clusterlabs.org/show_bug.cgi?id=5147

Best Regards,
Hideo Yamauchi.

--- On Thu, 2013/3/14, renayama19661...@ybb.ne.jp  
wrote:

> Hi Andrew,
> 
> > In "colocation", I used "resource_set".
> > However, a result did not include the change.
> 
> Please, about the result that I tried, give me comment.
> 
> Best Regards,
> Hideo Yamauchi.
> 
> --- On Thu, 2013/3/7, renayama19661...@ybb.ne.jp  
> wrote:
> 
> > Hi Andrew,
> > 
> > > > You use the resource sets _instead_ of a group.
> > > > If you want group.ordered=false, then use a colocation set (with
> > > > sequential=true).
> > 
> > In "colocation", I used "resource_set".
> > However, a result did not include the change.
> > 
> > Will this result be a mistake of my setting?
> > 
> > Case 1) sequential=false
> > (snip)
> >     
> >       
> >         
> >           
> >           
> >         
> >       
> >     
> > (sip)
> > [root@rh63-heartbeat2 ~]# grep "Initiating action" /var/log/ha-log
> > Mar  8 00:20:52 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> > Initiating action 2: probe_complete probe_complete on rh63-heartbeat1 - no 
> > waiting
> > Mar  8 00:20:52 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> > Initiating action 3: probe_complete probe_complete on rh63-heartbeat2 
> > (local) - no waiting
> > Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> > Initiating action 4: monitor vip-master_monitor_0 on rh63-heartbeat1
> > Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> > Initiating action 7: monitor vip-master_monitor_0 on rh63-heartbeat2 (local)
> > Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> > Initiating action 5: monitor vip-rep_monitor_0 on rh63-heartbeat1
> > Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> > Initiating action 8: monitor vip-rep_monitor_0 on rh63-heartbeat2 (local)
> > Mar  8 00:20:55 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> > Initiating action 6: probe_complete probe_complete on rh63-heartbeat2 
> > (local) - no waiting
> > Mar  8 00:20:56 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> > Initiating action 3: probe_complete probe_complete on rh63-heartbeat1 - no 
> > waiting
> > Mar  8 00:20:56 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> > Initiating action 5: start vip-master_start_0 on rh63-heartbeat1
> > Mar  8 00:20:58 rh63-heartbeat2 crmd: [22372]: info: te_rsc_command: 
> > Initiating action 1: stop vip-master_stop_0 on rh63-heartbeat1
> > 
> > 
> > Case 2) sequential=true
> > (snip)
> >     
> >       
> >         
> >           
> >           
> >         
> >       
> >     
> > (snip)
> > [root@rh63-heartbeat2 ~]# grep "Initiating action" /var/log/ha-log
> > Mar  7 23:54:44 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> > Initiating action 2: probe_complete probe_complete on rh63-heartbeat1 - no 
> > waiting
> > Mar  7 23:54:44 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> > Initiating action 3: probe_complete probe_complete on rh63-heartbeat2 
> > (local) - no waiting
> > Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> > Initiating action 4: monitor vip-master_monitor_0 on rh63-heartbeat1
> > Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> > Initiating action 7: monitor vip-master_monitor_0 on rh63-heartbeat2 (local)
> > Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> > Initiating action 5: monitor vip-rep_monitor_0 on rh63-heartbeat1
> > Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> > Initiating action 8: monitor vip-rep_monitor_0 on rh63-heartbeat2 (local)
> > Mar  7 23:54:48 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> > Initiating action 6: probe_complete probe_complete on rh63-heartbeat2 
> > (local) - no waiting
> > Mar  7 23:54:49 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> > Initiating action 3: probe_complete probe_complete on rh63-heartbeat1 - no 
> > waiting
> > Mar  7 23:54:49 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> > Initiating action 5: start vip-master_start_0 on rh63-heartbeat1
> > Mar  7 23:54:51 rh63-heartbeat2 crmd: [4]: info: te_rsc_command: 
> > Initiating action 1: stop vip-master_stop_0 on rh63-heartbeat1
> > 
> > 
> > Best Regards,
> > Hideo Yamauchi.
> > 
> > 
> > ___
> > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> > 
> > Project Home: http://www.clusterlabs.org
> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
> > 
> 
> ___
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting starte

Re: [Pacemaker] CMAN, corosync & pacemaker

2013-03-21 Thread Andrew Beekhof
On Fri, Mar 22, 2013 at 1:28 AM, Leon Fauster
 wrote:
> Am 21.03.2013 um 13:32 schrieb Ron Kerry :
>> I believe I understand how to configure CMAN to work with corosync and 
>> pacemaker, but I do not really understand what roles these three packages 
>> all play. Perhaps this information is exposed in the written documentation 
>> that is available in various places, but I have not been able to discern it.
>
> It seems you are in the same situation like me (last week) :-) So far i will
> provide my big picture. Deeper insights must be provided from the experts 
> here.
>
>
>
>> You can run the corosync/pacemaker (either ver=0 or ver=1) combination and 
>> get fully functioning HA clusters. What does adding CMAN to the mix add to 
>> the functionality? Or perhaps a better question, what functions does CMAN 
>> take over from either corosync or pacemaker and why is this considered a 
>> better combination? What are the deficiencies in corosync that running CMAN 
>> on top of it resolves?
>
>
> AFAIK - the architecture should be seen mainly as a stack.

Yep.

> On top of all is the so called
> "cluster resource manager (crm)" like pacemaker or rgmanager. Lets call it 
> crm-layer and
> we use pacemaker for it (check "pacemaker internals" in the documentation).
>
> Under the crm-layer is the communication layer (corosync) that checks e.g. 
> cluster membership.
>
> Membership is the keyword here when we take a look at cman - the 
> quorum-concept handles
> the active cluster/members and this can be also done by pacemaker/corosync. 
> cman can do
> this (quorum handling) very well (mature code) and is doing this in the lower 
> layers
> (extending corosync) - so thats the difference. When the lower level has some 
> information
> (e.g. about the quorum status) then all upper levels have the same 
> information. If pacemaker
> is doing this then they might be situations where different informations 
> exists in the different
> levels/layers.
>
>
>> I believe the preferred pacemaker based HA configuration in RHEL 6.4 uses 
>> all three packages and the preferred configuration in SLES11 SP2 is just 
>> corosync/pacemaker (I do not believe CMAN is even available in SLE-HAE).
>> Why the different approaches and what is the advantage of each configuration?
>
> cman is part of the RH cluster suite - the motivations of the different ha 
> approaches for
> both vendors are for sure more based in strategic/marketing needs.

Basically convergence takes a long time.

In both stacks there is corosync, pacemaker and a plugin that bridges
the gap (they both expose an API that provides membership and quorum).
On SLES that plugin comes from pacemaker, on RHEL it is CMAN.

-
resource management (pacemaker)
-
quorum/membership plugin (cman's or pacemaker's)
-
cluster messaging (corosync)
-

The Pacemaker plugin should never have existed, but back then we
didn't understand "CMAN" as well as we thought we did.
There may have also been some changes to the way it worked since then
that allow us to use it more easily today.

We all make mistakes, but that was one of my bigger ones.
Thankfully it will no longer matter eventually, as corosync 2.x
provides the necessary APIs natively so both plugins go away.

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] pacemaker node stuck offline

2013-03-21 Thread Andreas Kurz
On 2013-03-21 14:31, Patrick Hemmer wrote:
> I've got a 2-node cluster where it seems last night one of the nodes
> went offline, and I can't see any reason why.
> 
> Attached are the logs from the 2 nodes (the relevant timeframe seems to
> be 2013-03-21 between 06:05 and 06:10).
> This is on ubuntu 12.04

Looks like your non-redundant cluster-communication was interrupted at
around that time for whatever reason and your cluster split-brained.

Does the drbd-replication use a different network-connection? If yes,
why not using it for a redundant ring setup ... and you should use STONITH.

I also wonder why you have defined "expected_votes='1'" in your
cluster.conf.

Regards,
Andreas

-- 
Need help with Pacemaker?
http://www.hastexo.com/now


> 
> # crm status
> 
> Last updated: Thu Mar 21 13:17:21 2013
> Last change: Thu Mar 14 14:42:18 2013 via crm_shadow on i-a706d8ff
> Stack: cman
> Current DC: i-a706d8ff - partition WITHOUT quorum
> Version: 1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c
> 2 Nodes configured, unknown expected votes
> 5 Resources configured.
> 
> 
> Online: [ i-a706d8ff ]
> OFFLINE: [ i-3307d96b ]
> 
>  dns-postgresql(ocf::cloud:route53):Started i-a706d8ff
>  Master/Slave Set: ms-drbd-postgresql [drbd-postgresql]
>  Masters: [ i-a706d8ff ]
>  Stopped: [ drbd-postgresql:0 ]
>  fs-drbd-postgresql(ocf::heartbeat:Filesystem):Started i-a706d8ff
>  postgresql(ocf::heartbeat:pgsql):Started i-a706d8ff
> 
> 
> # cman_tool nodes
> Node  Sts   Inc   Joined   Name
> 181480898   M  4   2013-03-14 14:25:27  i-3307d96b
> 181481642   M   5132   2013-03-21 06:07:40  i-a706d8ff
> 
> 
> # cman_tool status
> Version: 6.2.0
> Config Version: 1
> Cluster Name: cloudapp-servic
> Cluster Id: 63629
> Cluster Member: Yes
> Cluster Generation: 5132
> Membership state: Cluster-Member
> Nodes: 2
> Expected votes: 1
> Total votes: 2
> Node votes: 1
> Quorum: 2 
> Active subsystems: 4
> Flags:
> Ports Bound: 0 
> Node name: i-3307d96b
> Node ID: 181480898
> Multicast addresses: 255.255.255.255
> Node addresses: 10.209.45.194
> 
> 
> 
> # cat /etc/cluster/cluster.conf
> 
> 
>  syslog_priority='debug' />
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> ___
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
> 





signature.asc
Description: OpenPGP digital signature
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] CMAN, corosync & pacemaker

2013-03-21 Thread Leon Fauster
Am 21.03.2013 um 13:32 schrieb Ron Kerry :
> I believe I understand how to configure CMAN to work with corosync and 
> pacemaker, but I do not really understand what roles these three packages all 
> play. Perhaps this information is exposed in the written documentation that 
> is available in various places, but I have not been able to discern it.

It seems you are in the same situation like me (last week) :-) So far i will 
provide my big picture. Deeper insights must be provided from the experts here.



> You can run the corosync/pacemaker (either ver=0 or ver=1) combination and 
> get fully functioning HA clusters. What does adding CMAN to the mix add to 
> the functionality? Or perhaps a better question, what functions does CMAN 
> take over from either corosync or pacemaker and why is this considered a 
> better combination? What are the deficiencies in corosync that running CMAN 
> on top of it resolves?


AFAIK - the architecture should be seen mainly as a stack. On top of all is the 
so called 
"cluster resource manager (crm)" like pacemaker or rgmanager. Lets call it 
crm-layer and
we use pacemaker for it (check "pacemaker internals" in the documentation). 

Under the crm-layer is the communication layer (corosync) that checks e.g. 
cluster membership.

Membership is the keyword here when we take a look at cman - the quorum-concept 
handles 
the active cluster/members and this can be also done by pacemaker/corosync. 
cman can do
this (quorum handling) very well (mature code) and is doing this in the lower 
layers 
(extending corosync) - so thats the difference. When the lower level has some 
information 
(e.g. about the quorum status) then all upper levels have the same information. 
If pacemaker 
is doing this then they might be situations where different informations exists 
in the different 
levels/layers. 


> I believe the preferred pacemaker based HA configuration in RHEL 6.4 uses all 
> three packages and the preferred configuration in SLES11 SP2 is just 
> corosync/pacemaker (I do not believe CMAN is even available in SLE-HAE).
> Why the different approaches and what is the advantage of each configuration?

cman is part of the RH cluster suite - the motivations of the different ha 
approaches for 
both vendors are for sure more based in strategic/marketing needs.

--
LF





___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[Pacemaker] CMAN, corosync & pacemaker

2013-03-21 Thread Ron Kerry
I believe I understand how to configure CMAN to work with corosync and pacemaker, but I do not 
really understand what roles these three packages all play. Perhaps this information is exposed in 
the written documentation that is available in various places, but I have not been able to discern it.


You can run the corosync/pacemaker (either ver=0 or ver=1) combination and get fully functioning HA 
clusters. What does adding CMAN to the mix add to the functionality? Or perhaps a better question, 
what functions does CMAN take over from either corosync or pacemaker and why is this considered a 
better combination? What are the deficiencies in corosync that running CMAN on top of it resolves?


I believe the preferred pacemaker based HA configuration in RHEL 6.4 uses all three packages and the 
preferred configuration in SLES11 SP2 is just corosync/pacemaker (I do not believe CMAN is even 
available in SLE-HAE). Why the different approaches and what is the advantage of each configuration?


--

Ron Kerry



___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Exchanging data between resource agent instances

2013-03-21 Thread Riccardo Bicelli
Sorry for spamming :)

Il 19/03/2013 10:43, Lars Ellenberg ha scritto:
>
> Is that so.  What for?
> Can you explain in more detail?

I'm writing a resource agent for handling SCST in master/slave with ALUA
multipath support.
On Master node it has to create a SCSI LUN mapped to a real device, and map
the path as active.
On Slave node (and even on Master when the resource starts) ot has to
create a dummy lun (with specific null handler) in the same size of real
device, and map path as standby.

When switching roles, it simply puts path both in standby, replaces on
Master node the lun with null lun, replaces on the slave node the null lun
with real device, put in active state the Slave node.
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org