Re: [Pacemaker] why sometimes pengine seems lazy

2015-03-29 Thread Andrew Beekhof

> On 10 Feb 2015, at 6:50 pm, d tbsky  wrote:
> 
> hi:
>   I was using pacemaker and drbd with sl linux 6.5/6.6. all are fine.
> 
>   now I am tesing sl linux 7.0 and I notice when I want to promote
> the drbd resource with "pcs resource meta  my-ms-drbd master-max=2".
> 
>sometimes pengine find the change immediately, but sometimes it
> find the change after about a minute. I don't know if the delay is
> normal? I didn't notice the delay when I using sl linux 6.5/6.6.
> 
>   the "good" result. kvm-3-ms-drbd set master-max = 2 at 13:00:07 and
> pengine find it at 13:00:07

[...]

> 
> 
>   the "lazy" result: kvm-2-ms-drbd set master-max = 2 at  12:47:09.
> pengine find it at 12:47:52
> pacemaker seems quite busy.

In a nutshell, this is your answer.
The cluster was busy handling nodes joining and leaving the cluster and wasn't 
yet ready to act on your change.


___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[Pacemaker] why sometimes pengine seems lazy

2015-02-10 Thread d tbsky
hi:
   I was using pacemaker and drbd with sl linux 6.5/6.6. all are fine.

   now I am tesing sl linux 7.0 and I notice when I want to promote
the drbd resource with "pcs resource meta  my-ms-drbd master-max=2".

sometimes pengine find the change immediately, but sometimes it
find the change after about a minute. I don't know if the delay is
normal? I didn't notice the delay when I using sl linux 6.5/6.6.

   the "good" result. kvm-3-ms-drbd set master-max = 2 at 13:00:07 and
 pengine find it at 13:00:07

Feb 10 13:00:06 [2893] love1-test.lhy.com.twcib: info:
cib_process_request: Completed cib_query operation for section
//constraints: OK (rc=0, origin=love2-test.lhy.com.tw/cibadmin/2,
version=0.2084.3)
Feb 10 13:00:06 [2893] love1-test.lhy.com.twcib: info:
cib_process_request: Completed cib_query operation for section
//constraints: OK (rc=0, origin=love2-test.lhy.com.tw/cibadmin/2,
version=0.2084.3)
Feb 10 13:00:06 [2893] love1-test.lhy.com.twcib: info:
cib_process_request: Completed cib_query operation for section
//constraints: OK (rc=0, origin=love2-test.lhy.com.tw/cibadmin/2,
version=0.2084.3)
Feb 10 13:00:07 [2893] love1-test.lhy.com.twcib:   notice:
cib:diff:Diff: --- 0.2084.3
Feb 10 13:00:07 [2893] love1-test.lhy.com.twcib:   notice:
cib:diff:Diff: +++ 0.2085.1 206a58e68f4a9cd8e72c7ebb40bef026
Feb 10 13:00:07 [2893] love1-test.lhy.com.twcib:   notice:
cib:diff:--   
Feb 10 13:00:07 [2893] love1-test.lhy.com.twcib:   notice:
cib:diff:++   
Feb 10 13:00:07 [2893] love1-test.lhy.com.twcib: info:
cib_process_request: Completed cib_replace operation for
section configuration: OK (rc=0,
origin=love2-test.lhy.com.tw/cibadmin/2, version=0.2085.1)
Feb 10 13:00:07 [2898] love1-test.lhy.com.tw   crmd: info:
abort_transition_graph:  te_update_diff:126 - Triggered transition
abort (complete=1, node=, tag=diff, id=(null), magic=NA, cib=0.2085.1)
: Non-status change
Feb 10 13:00:07 [2898] love1-test.lhy.com.tw   crmd:   notice:
do_state_transition: State transition S_IDLE ->
S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL
origin=abort_transition_graph ]
Feb 10 13:00:07 [2893] love1-test.lhy.com.twcib: info:
cib_process_request: Completed cib_query operation for section
'all': OK (rc=0, origin=local/crmd/839, version=0.2085.1)
Feb 10 13:00:07 [2893] love1-test.lhy.com.twcib: info:
write_cib_contents:  Archived previous version as
/var/lib/pacemaker/cib/cib-17.raw
Feb 10 13:00:07 [2897] love1-test.lhy.com.twpengine:   notice:
unpack_config:   On loss of CCM Quorum: Ignore
Feb 10 13:00:07 [2897] love1-test.lhy.com.twpengine: info:
determine_online_status: Node love2-test.lhy.com.tw is online
Feb 10 13:00:07 [2897] love1-test.lhy.com.twpengine: info:
determine_online_status: Node love1-test.lhy.com.tw is online
Feb 10 13:00:07 [2893] love1-test.lhy.com.twcib: info:
write_cib_contents:  Wrote version 0.2085.0 of the CIB to disk
(digest: bfdd9b0a25cde05a4b2777b6fc670519)
Feb 10 13:00:07 [2897] love1-test.lhy.com.twpengine:   notice:
unpack_rsc_op:   Operation monitor found resource kvm-6-drbd:0
active in master mode on love2-test.lhy.com.tw
Feb 10 13:00:07 [2897] love1-test.lhy.com.twpengine: info:
unpack_rsc_op:   Operation monitor found resource kvm-6 active on
love2-test.lhy.com.tw
Feb 10 13:00:07 [2897] love1-test.lhy.com.twpengine: info:
unpack_rsc_op:   Operation monitor found resource kvm-1-drbd:1
active on love1-test.lhy.com.tw
Feb 10 13:00:07 [2893] love1-test.lhy.com.twcib: info:
retrieveCib: Reading cluster configuration from:
/var/lib/pacemaker/cib/cib.2Mn5wa (digest:
/var/lib/pacemaker/cib/cib.0nfve5)
Feb 10 13:00:07 [2897] love1-test.lhy.com.twpengine: info:
unpack_rsc_op:   Operation monitor found resource kvm-3-drbd:1
active on love1-test.lhy.com.tw
Feb 10 13:00:07 [2897] love1-test.lhy.com.twpengine:   notice:
unpack_rsc_op:   Re-initiated expired calculated failure
kvm-4_last_failure_0 (rc=7,
magic=0:7;144:22:0:87034531-de2d-4395-b3c0-9bc0cecfc50e) on
love1-test.lhy.com.tw
Feb 10 13:00:07 [2897] love1-test.lhy.com.twpengine: info:
unpack_rsc_op:   Operation monitor found resource kvm-2-drbd:1
active on love1-test.lhy.com.tw
Feb 10 13:00:07 [2897] love1-test.lhy.com.twpengine: info:
unpack_rsc_op:   Operation monitor found resource kvm-5 active on
love1-test.lhy.com.tw
Feb 10 13:00:07 [2897] love1-test.lhy.com.twpengine:   notice:
unpack_rsc_op:   Operation monitor found resource kvm-5-drbd:1
active in master mode on love1-test.lhy.com.tw
Feb 10 13:00:07 [2897] love1-test.lhy.com.twpengine: info:
unpack_rsc_op:   Operation monitor found resource kvm-6-drbd:1
active on love1-test.lhy.com.tw
Feb 10 13:00:07 [2897] love1-test.lhy.com.tw