Hi All,

We confirmed movement of the trouble of the clone resource that we combined 
with Master/Slave resource.

The master / slave resources are replaced under the influence of the trouble of 
the clonal resource.

We confirmed it in the next procedure.


Step1) We start a cluster and send cib.

============
Last updated: Mon Sep 10 15:26:25 2012
Stack: Heartbeat
Current DC: drbd2 (08607c71-da7b-4abf-b6d5-39ee39552e89) - partition with quorum
Version: 1.0.12-c6770b8
2 Nodes configured, unknown expected votes
6 Resources configured.
============

Online: [ drbd1 drbd2 ]

 Resource Group: grpPostgreSQLDB
     prmApPostgreSQLDB  (ocf::pacemaker:Dummy): Started drbd1
 Resource Group: grpStonith1
     prmStonith1-2      (stonith:external/ssh): Started drbd2
     prmStonith1-3      (stonith:meatware):     Started drbd2
 Resource Group: grpStonith2
     prmStonith2-2      (stonith:external/ssh): Started drbd1
     prmStonith2-3      (stonith:meatware):     Started drbd1
 Master/Slave Set: msDrPostgreSQLDB
     Masters: [ drbd1 ]
     Slaves: [ drbd2 ]
 Clone Set: clnDiskd1
     Started: [ drbd1 drbd2 ]
 Clone Set: clnPingd
     Started: [ drbd1 drbd2 ]

Step2) We cause a monitor error in pingd.

[root@drbd1 ~]# rm -rf /var/run/pingd-default_ping_set 

Step3) FailOver is finished.

============
Last updated: Mon Sep 10 15:27:08 2012
Stack: Heartbeat
Current DC: drbd2 (08607c71-da7b-4abf-b6d5-39ee39552e89) - partition with quorum
Version: 1.0.12-c6770b8
2 Nodes configured, unknown expected votes
6 Resources configured.
============

Online: [ drbd1 drbd2 ]

 Resource Group: grpPostgreSQLDB
     prmApPostgreSQLDB  (ocf::pacemaker:Dummy): Started drbd2
 Resource Group: grpStonith1
     prmStonith1-2      (stonith:external/ssh): Started drbd2
     prmStonith1-3      (stonith:meatware):     Started drbd2
 Resource Group: grpStonith2
     prmStonith2-2      (stonith:external/ssh): Started drbd1
     prmStonith2-3      (stonith:meatware):     Started drbd1
 Master/Slave Set: msDrPostgreSQLDB
     Masters: [ drbd2 ]
     Stopped: [ prmDrPostgreSQLDB:1 ]
 Clone Set: clnDiskd1
     Started: [ drbd1 drbd2 ]
 Clone Set: clnPingd
     Started: [ drbd2 ]
     Stopped: [ prmPingd:0 ]

Failed actions:
    prmPingd:0_monitor_10000 (node=drbd1, call=14, rc=7, status=complete): not 
running



However, Master/Slave resources seemed to be replaced when we watched log.

Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Move    resource 
prmApPostgreSQLDB#011(Started drbd1 -> drbd2)
Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Leave   resource 
prmStonith1-2#011(Started drbd2)
Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Leave   resource 
prmStonith1-3#011(Started drbd2)
Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Leave   resource 
prmStonith2-2#011(Started drbd1)
Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Leave   resource 
prmStonith2-3#011(Started drbd1)
Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Move    resource 
prmDrPostgreSQLDB:0#011(Master drbd1 -> drbd2)
Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Stop    resource 
prmDrPostgreSQLDB:1#011(drbd2)
Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Leave   resource 
prmDiskd1:0#011(Started drbd1)
Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Leave   resource 
prmDiskd1:1#011(Started drbd2)
Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Stop    resource 
prmPingd:0#011(drbd1)
Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Leave   resource 
prmPingd:1#011(Started drbd2)

The replacement is unnecessary, and Slave becomes Master, and inoperative 
Master should have only to originally stop.

However, this problem seems to be solved in Pacemaker1.1.

Will the correction be possible for Pacemaker1.0?
Because I have a big difference in placement processing with Pacemaker1.1, I 
think that the correction to Pacemaker1.0 is difficult.

 * This problem may have been reported as a known problem.
 * I registered this problem with Bugzilla.
  * http://bugs.clusterlabs.org/show_bug.cgi?id=5103

Best Regards,
Hideo Yamauchi.


_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to