I have configured some network resources to automatically standby their node if 
the system detects a failure on them. However, the DRBD slave that I have 
configured does not automatically restart after the node is "unstandby-ed" 
after the failure-timeout expires.
Is there any way to make the "stopped" DRBDSlave resource automatically start 
again once the node is recovered?

See the  progression of events below:

Running cluster:
Wed Mar 30 16:04:20 UTC 2016
Cluster name:
Last updated: Wed Mar 30 16:04:20 2016
Last change: Wed Mar 30 16:03:24 2016
Stack: classic openais (with plugin)
Current DC: ha-d1.tw.com - partition with quorum
Version: 1.1.12-561c4cf
2 Nodes configured, 2 expected votes
7 Resources configured


Online: [ ha-d1.tw.com ha-d2.tw.com ]

Full list of resources:

 Resource Group: network
     inif       (ocf::custom:ip.sh):       Started ha-d1.tw.com
     outif      (ocf::custom:ip.sh):       Started ha-d1.tw.com
     dmz1       (ocf::custom:ip.sh):       Started ha-d1.tw.com
 Master/Slave Set: DRBDMaster [DRBDSlave]
     Masters: [ ha-d1.tw.com ]
     Slaves: [ ha-d2.tw.com ]
 Resource Group: filesystem
     DRBDFS     (ocf::heartbeat:Filesystem):    Started ha-d1.tw.com
 Resource Group: application
     service_failover   (ocf::custom:service_failover):    Started ha-d1.tw.com


version: 8.4.5 (api:1/proto:86-101)
srcversion: 315FB2BBD4B521D13C20BF4

 1: cs:Connected ro:Primary/Secondary ds:UpToDate/UpToDate C r-----
    ns:4 nr:0 dw:4 dr:757 al:1 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:f oos:0
[153766.565352] block drbd1: send bitmap stats [Bytes(packets)]: plain 0(0), 
RLE 21(1), total 21; compression: 100.0%
[153766.568303] block drbd1: receive bitmap stats [Bytes(packets)]: plain 0(0), 
RLE 21(1), total 21; compression: 100.0%
[153766.568316] block drbd1: helper command: /sbin/drbdadm before-resync-source 
minor-1
[153766.568356] block drbd1: helper command: /sbin/drbdadm before-resync-source 
minor-1 exit code 255 (0xfffffffe)
[153766.568363] block drbd1: conn( WFBitMapS -> SyncSource ) pdsk( Consistent 
-> Inconsistent )
[153766.568374] block drbd1: Began resync as SyncSource (will sync 4 KB [1 bits 
set]).
[153766.568444] block drbd1: updated sync UUID 
B0DA745C79C56591:36E0631B6F022952:36DF631B6F022952:133127197CF097C6
[153766.577695] block drbd1: Resync done (total 1 sec; paused 0 sec; 4 K/sec)
[153766.577700] block drbd1: updated UUIDs 
B0DA745C79C56591:0000000000000000:36E0631B6F022952:36DF631B6F022952
[153766.577705] block drbd1: conn( SyncSource -> Connected ) pdsk( Inconsistent 
-> UpToDate )¯

Failure detected:
Wed Mar 30 16:08:22 UTC 2016
Cluster name:
Last updated: Wed Mar 30 16:08:22 2016
Last change: Wed Mar 30 16:03:24 2016
Stack: classic openais (with plugin)
Current DC: ha-d1.tw.com - partition with quorum
Version: 1.1.12-561c4cf
2 Nodes configured, 2 expected votes
7 Resources configured


Node ha-d1.tw.com: standby (on-fail)
Online: [ ha-d2.tw.com ]

Full list of resources:

 Resource Group: network
     inif       (ocf::custom:ip.sh):       Started ha-d1.tw.com
     outif      (ocf::custom:ip.sh):       Started ha-d1.tw.com
     dmz1       (ocf::custom:ip.sh):       FAILED ha-d1.tw.com
 Master/Slave Set: DRBDMaster [DRBDSlave]
     Masters: [ ha-d1.tw.com ]
     Slaves: [ ha-d2.tw.com ]
 Resource Group: filesystem
     DRBDFS     (ocf::heartbeat:Filesystem):    Started ha-d1.tw.com
 Resource Group: application
     service_failover   (ocf::custom:service_failover):    Started ha-d1.tw.com

Failed actions:
    dmz1_monitor_7000 on ha-d1.tw.com 'not running' (7): call=156, 
status=complete, last-rc-change='Wed Mar 30 16:08:19 2016', queued=0ms, exec=0ms



version: 8.4.5 (api:1/proto:86-101)
srcversion: 315FB2BBD4B521D13C20BF4

 1: cs:Connected ro:Primary/Secondary ds:UpToDate/UpToDate C r-----
    ns:4 nr:0 dw:4 dr:765 al:1 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:f oos:0
[153766.568356] block drbd1: helper command: /sbin/drbdadm before-resync-source 
minor-1 exit code 255 (0xfffffffe)
[153766.568363] block drbd1: conn( WFBitMapS -> SyncSource ) pdsk( Consistent 
-> Inconsistent )
[153766.568374] block drbd1: Began resync as SyncSource (will sync 4 KB [1 bits 
set]).
[153766.568444] block drbd1: updated sync UUID 
B0DA745C79C56591:36E0631B6F022952:36DF631B6F022952:133127197CF097C6
[153766.577695] block drbd1: Resync done (total 1 sec; paused 0 sec; 4 K/sec)
[153766.577700] block drbd1: updated UUIDs 
B0DA745C79C56591:0000000000000000:36E0631B6F022952:36DF631B6F022952
[153766.577705] block drbd1: conn( SyncSource -> Connected ) pdsk( Inconsistent 
-> UpToDate )
[154057.455270] e1000: eth2 NIC Link is Down
[154057.455451] e1000 0000:02:02.0 eth2: Reset adapter

Failover complete:
Wed Mar 30 16:09:02 UTC 2016
Cluster name:
Last updated: Wed Mar 30 16:09:02 2016
Last change: Wed Mar 30 16:03:24 2016
Stack: classic openais (with plugin)
Current DC: ha-d1.tw.com - partition with quorum
Version: 1.1.12-561c4cf
2 Nodes configured, 2 expected votes
7 Resources configured


Node ha-d1.tw.com: standby (on-fail)
Online: [ ha-d2.tw.com ]

Full list of resources:

 Resource Group: network
     inif       (ocf::custom:ip.sh):       Started ha-d2.tw.com
     outif      (ocf::custom:ip.sh):       Started ha-d2.tw.com
     dmz1       (ocf::custom:ip.sh):       Started ha-d2.tw.com
 Master/Slave Set: DRBDMaster [DRBDSlave]
     Masters: [ ha-d2.tw.com ]
     Stopped: [ ha-d1.tw.com ]
 Resource Group: filesystem
     DRBDFS     (ocf::heartbeat:Filesystem):    Started ha-d2.tw.com
 Resource Group: application
     service_failover   (ocf::custom:service_failover):    Started ha-d2.tw.com

Failed actions:
    dmz1_monitor_7000 on ha-d1.tw.com 'not running' (7): call=156, 
status=complete, last-rc-change='Wed Mar 30 16:08:19 2016', queued=0ms, exec=0ms



version: 8.4.5 (api:1/proto:86-101)
srcversion: 315FB2BBD4B521D13C20BF4
[154094.894524] drbd wwwdata: conn( Disconnecting -> StandAlone )
[154094.894525] drbd wwwdata: receiver terminated
[154094.894527] drbd wwwdata: Terminating drbd_r_wwwdata
[154094.894559] block drbd1: disk( UpToDate -> Failed )
[154094.894569] block drbd1: bitmap WRITE of 0 pages took 0 jiffies
[154094.894571] block drbd1: 4 KB (1 bits) marked out-of-sync by on disk 
bit-map.
[154094.894574] block drbd1: disk( Failed -> Diskless )
[154094.894647] block drbd1: drbd_bm_resize called with capacity == 0
[154094.894652] drbd wwwdata: Terminating drbd_w_wwwdata

Standby node recovered, with DRBDSlave stopped (I want DRBDSlave started here):
Wed Mar 30 16:13:01 UTC 2016
Cluster name:
Last updated: Wed Mar 30 16:13:01 2016
Last change: Wed Mar 30 16:03:24 2016
Stack: classic openais (with plugin)
Current DC: ha-d1.tw.com - partition with quorum
Version: 1.1.12-561c4cf
2 Nodes configured, 2 expected votes
7 Resources configured


Online: [ ha-d1.tw.com ha-d2.tw.com ]

Full list of resources:

 Resource Group: network
     inif       (ocf::custom:ip.sh):       Started ha-d2.tw.com
     outif      (ocf::custom:ip.sh):       Started ha-d2.tw.com
     dmz1       (ocf::custom:ip.sh):       Started ha-d2.tw.com
 Master/Slave Set: DRBDMaster [DRBDSlave]
     Masters: [ ha-d2.tw.com ]
     Stopped: [ ha-d1.tw.com ]
 Resource Group: filesystem
     DRBDFS     (ocf::heartbeat:Filesystem):    Started ha-d2.tw.com
 Resource Group: application
     service_failover   (ocf::custom:service_failover):    Started ha-d2.tw.com


version: 8.4.5 (api:1/proto:86-101)
srcversion: 315FB2BBD4B521D13C20BF4
[154094.894574] block drbd1: disk( Failed -> Diskless )
[154094.894647] block drbd1: drbd_bm_resize called with capacity == 0
[154094.894652] drbd wwwdata: Terminating drbd_w_wwwdata

--
Sam Gardner
Trustwave | SMART SECURITY ON DEMAND

________________________________

This transmission may contain information that is privileged, confidential, 
and/or exempt from disclosure under applicable law. If you are not the intended 
recipient, you are hereby notified that any disclosure, copying, distribution, 
or use of the information contained herein (including any reliance thereon) is 
strictly prohibited. If you received this transmission in error, please 
immediately contact the sender and destroy the material in its entirety, 
whether in electronic or hard copy format.
_______________________________________________
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to