The problem with this approach is that ordering of the constraints in
the cib is not preserved between the nodes.
I will follow up further on the bugzilla.

On Wed, Aug 1, 2012 at 11:35 AM,  <renayama19661...@ybb.ne.jp> wrote:
> Hi All,
>
> When the ID of the resource changed, we confirmed that control of different 
> colocation was carried out.
> The control of the resource may vary according to this problem only in the 
> difference in ID name of the resource.
>
>
> The first pattern) The state transition that we expect is made. (pe-input-423)
>
> [root@drbd2 trac2114]# ptest -x pe-input-423 -VVV
> ptest[13220]: 2012/08/01_10:29:16 notice: unpack_config: On loss of CCM 
> Quorum: Ignore
> ptest[13220]: 2012/08/01_10:29:16 WARN: unpack_nodes: Blind faith: not 
> fencing unseen nodes
> ptest[13220]: 2012/08/01_10:29:16 WARN: unpack_rsc_op: Processing failed op 
> postgresql:0_monitor_9000 on 02-sl6: not running (7)
> ptest[13220]: 2012/08/01_10:29:16 notice: native_print: vipCheck        
> (ocf::pacemaker:Dummy): Started 02-sl6
> ptest[13220]: 2012/08/01_10:29:16 notice: native_print: vipCheckSupport 
> (ocf::pacemaker:Dummy): Started 02-sl6
> ptest[13220]: 2012/08/01_10:29:16 notice: group_print:  Resource Group: 
> master-group
> ptest[13220]: 2012/08/01_10:29:16 notice: native_print:      vip-master 
> (ocf::heartbeat:IPaddr2):       Started 02-sl6
> ptest[13220]: 2012/08/01_10:29:16 notice: native_print:      vip-rep    
> (ocf::heartbeat:IPaddr2):       Started 02-sl6
> ptest[13220]: 2012/08/01_10:29:16 notice: clone_print:  Master/Slave Set: 
> msPostgresql
> ptest[13220]: 2012/08/01_10:29:16 notice: native_print:      postgresql:0     
>   (ocf::heartbeat:pgsql): Slave 02-sl6 FAILED
> ptest[13220]: 2012/08/01_10:29:16 notice: short_print:      Slaves: [ 03-sl6 ]
> ptest[13220]: 2012/08/01_10:29:16 notice: clone_print:  Clone Set: clnDiskd1
> ptest[13220]: 2012/08/01_10:29:16 notice: short_print:      Started: [ 02-sl6 
> 03-sl6 ]
> ptest[13220]: 2012/08/01_10:29:16 notice: clone_print:  Clone Set: clnDiskd2
> ptest[13220]: 2012/08/01_10:29:16 notice: short_print:      Started: [ 02-sl6 
> 03-sl6 ]
> ptest[13220]: 2012/08/01_10:29:16 notice: clone_print:  Clone Set: clnPingd
> ptest[13220]: 2012/08/01_10:29:16 notice: short_print:      Started: [ 02-sl6 
> 03-sl6 ]
> ptest[13220]: 2012/08/01_10:29:16 WARN: common_apply_stickiness: Forcing 
> msPostgresql away from 02-sl6 after 1 failures (max=1)
> ptest[13220]: 2012/08/01_10:29:16 WARN: common_apply_stickiness: Forcing 
> msPostgresql away from 02-sl6 after 1 failures (max=1)
> ptest[13220]: 2012/08/01_10:29:16 notice: RecurringOp:  Start recurring 
> monitor (10s) for vip-master on 03-sl6
> ptest[13220]: 2012/08/01_10:29:16 notice: RecurringOp:  Start recurring 
> monitor (10s) for vip-rep on 03-sl6
> ptest[13220]: 2012/08/01_10:29:16 notice: RecurringOp:  Start recurring 
> monitor (9s) for postgresql:1 on 03-sl6
> ptest[13220]: 2012/08/01_10:29:16 notice: RecurringOp:  Start recurring 
> monitor (9s) for postgresql:1 on 03-sl6
> ptest[13220]: 2012/08/01_10:29:16 notice: LogActions: Move    resource 
> vipCheck (Started 02-sl6 -> 03-sl6)
> ptest[13220]: 2012/08/01_10:29:16 notice: LogActions: Move    resource 
> vipCheckSupport  (Started 02-sl6 -> 03-sl6)
> ptest[13220]: 2012/08/01_10:29:16 notice: LogActions: Move    resource 
> vip-master       (Started 02-sl6 -> 03-sl6)
> ptest[13220]: 2012/08/01_10:29:16 notice: LogActions: Move    resource 
> vip-rep  (Started 02-sl6 -> 03-sl6)
> ptest[13220]: 2012/08/01_10:29:16 notice: LogActions: Stop    resource 
> postgresql:0     (02-sl6)
> ptest[13220]: 2012/08/01_10:29:16 notice: LogActions: Promote postgresql:1    
>   (Slave -> Master 03-sl6)
> ptest[13220]: 2012/08/01_10:29:16 notice: LogActions: Leave   resource 
> prmDiskd1:0      (Started 02-sl6)
> ptest[13220]: 2012/08/01_10:29:16 notice: LogActions: Leave   resource 
> prmDiskd1:1      (Started 03-sl6)
> ptest[13220]: 2012/08/01_10:29:16 notice: LogActions: Leave   resource 
> prmDiskd2:0      (Started 02-sl6)
> ptest[13220]: 2012/08/01_10:29:16 notice: LogActions: Leave   resource 
> prmDiskd2:1      (Started 03-sl6)
> ptest[13220]: 2012/08/01_10:29:16 notice: LogActions: Leave   resource 
> pingCheck:0      (Started 02-sl6)
> ptest[13220]: 2012/08/01_10:29:16 notice: LogActions: Leave   resource 
> pingCheck:1      (Started 03-sl6)
>
>
> The second pattern) Different state transition is made only by resource ID 
> being different.(pe-input-396)
>  * I changed a resource name into gtmproxy1 from vipCheck.
>  * I changed a resource name into gtmproxy1Support from vipCheckSupport.
>
> [root@drbd2 trac2114]# ptest -x pe-input-396 -VVV
> ptest[13221]: 2012/08/01_10:29:36 notice: unpack_config: On loss of CCM 
> Quorum: Ignore
> ptest[13221]: 2012/08/01_10:29:36 WARN: unpack_nodes: Blind faith: not 
> fencing unseen nodes
> ptest[13221]: 2012/08/01_10:29:36 WARN: unpack_rsc_op: Processing failed op 
> datanode1:0_monitor_9000 on 02-sl6: not running (7)
> ptest[13221]: 2012/08/01_10:29:36 notice: native_print: gtmproxy1       
> (ocf::pacemaker:Dummy): Started 02-sl6
> ptest[13221]: 2012/08/01_10:29:36 notice: native_print: gtmproxy1Support      
>   (ocf::pacemaker:Dummy): Started 02-sl6
> ptest[13221]: 2012/08/01_10:29:36 notice: group_print:  Resource Group: 
> master-group1
> ptest[13221]: 2012/08/01_10:29:36 notice: native_print:      vip-master1      
>   (ocf::heartbeat:IPaddr2):       Started 02-sl6
> ptest[13221]: 2012/08/01_10:29:36 notice: native_print:      vip-rep1   
> (ocf::heartbeat:IPaddr2):       Started 02-sl6
> ptest[13221]: 2012/08/01_10:29:36 notice: clone_print:  Master/Slave Set: 
> msDatanode1
> ptest[13221]: 2012/08/01_10:29:36 notice: native_print:      datanode1:0      
>   (ocf::heartbeat:pgsql): Slave 02-sl6 FAILED
> ptest[13221]: 2012/08/01_10:29:36 notice: short_print:      Slaves: [ 03-sl6 ]
> ptest[13221]: 2012/08/01_10:29:36 notice: clone_print:  Clone Set: clnDiskd1
> ptest[13221]: 2012/08/01_10:29:36 notice: short_print:      Started: [ 02-sl6 
> 03-sl6 ]
> ptest[13221]: 2012/08/01_10:29:36 notice: clone_print:  Clone Set: clnDiskd2
> ptest[13221]: 2012/08/01_10:29:36 notice: short_print:      Started: [ 02-sl6 
> 03-sl6 ]
> ptest[13221]: 2012/08/01_10:29:36 notice: clone_print:  Clone Set: clnPingd
> ptest[13221]: 2012/08/01_10:29:36 notice: short_print:      Started: [ 02-sl6 
> 03-sl6 ]
> ptest[13221]: 2012/08/01_10:29:36 WARN: common_apply_stickiness: Forcing 
> msDatanode1 away from 02-sl6 after 1 failures (max=1)
> ptest[13221]: 2012/08/01_10:29:36 WARN: common_apply_stickiness: Forcing 
> msDatanode1 away from 02-sl6 after 1 failures (max=1)
> ptest[13221]: 2012/08/01_10:29:36 notice: RecurringOp:  Start recurring 
> monitor (9s) for datanode1:1 on 03-sl6
> ptest[13221]: 2012/08/01_10:29:36 notice: RecurringOp:  Start recurring 
> monitor (9s) for datanode1:1 on 03-sl6
> ptest[13221]: 2012/08/01_10:29:36 notice: LogActions: Stop    resource 
> gtmproxy1        (02-sl6)
> ptest[13221]: 2012/08/01_10:29:36 notice: LogActions: Stop    resource 
> gtmproxy1Support (Started 02-sl6)
> ptest[13221]: 2012/08/01_10:29:36 notice: LogActions: Stop    resource 
> vip-master1      (02-sl6)
> ptest[13221]: 2012/08/01_10:29:36 notice: LogActions: Stop    resource 
> vip-rep1 (02-sl6)
> ptest[13221]: 2012/08/01_10:29:36 notice: LogActions: Stop    resource 
> datanode1:0      (02-sl6)
> ptest[13221]: 2012/08/01_10:29:36 notice: LogActions: Promote datanode1:1     
>   (Slave -> Master 03-sl6)
> ptest[13221]: 2012/08/01_10:29:36 notice: LogActions: Leave   resource 
> prmDiskd1:0      (Started 02-sl6)
> ptest[13221]: 2012/08/01_10:29:36 notice: LogActions: Leave   resource 
> prmDiskd1:1      (Started 03-sl6)
> ptest[13221]: 2012/08/01_10:29:36 notice: LogActions: Leave   resource 
> prmDiskd2:0      (Started 02-sl6)
> ptest[13221]: 2012/08/01_10:29:36 notice: LogActions: Leave   resource 
> prmDiskd2:1      (Started 03-sl6)
> ptest[13221]: 2012/08/01_10:29:36 notice: LogActions: Leave   resource 
> pingCheck:0      (Started 02-sl6)
> ptest[13221]: 2012/08/01_10:29:36 notice: LogActions: Leave   resource 
> pingCheck:1      (Started 03-sl6)
>
>
> I made the patch(trac2114.patch) which solved this problem.(for 
> ClusterLabs-pacemaker-1.0-Pacemaker-1.0.12-19-g489cf4e)
>
> Please confirm the contents of my patch.
> And please apply to a repository.
> Or please solve a problem by a better correction
>
>  * I do not confirm movement in Pacemaker1.1.
>  * The correction may be necessary in Pacemaker1.1.
>
> Best Regards,
> Hideo Yamauchi.
> _______________________________________________
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>

_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to