On Thu, May 19, 2011 at 11:53:24PM -0700, Randy Katz wrote: > Lars, > > Thank you much for the answer on the "standby" issue. > It seems that that was the tip of my real issue. So now I have both nodes > coming online. And it seems ha1 starts fine with all the resources starting. > > With them both online if I issue the: crm mode standby ha1.iohost.com
Why. Learn about "crm resource move". (and unmove, for that matter). > Then I see IP Takeover on ha2 but the other resources do not start, > ever, it remains: > > Node ha1.iohost.com (b159178d-c19b-4473-aa8e-13e487b65e33): standby > Online: [ ha2.iohost.com ] > > Resource Group: WebServices > ip1 (ocf::heartbeat:IPaddr2): Started ha2.iohost.com > ip1arp (ocf::heartbeat:SendArp): Started ha2.iohost.com > fs_webfs (ocf::heartbeat:Filesystem): Stopped > fs_mysql (ocf::heartbeat:Filesystem): Stopped > apache2 (lsb:httpd): Stopped > mysql (ocf::heartbeat:mysql): Stopped > Master/Slave Set: ms_drbd_mysql > Slaves: [ ha2.iohost.com ] > Stopped: [ drbd_mysql:0 ] > Master/Slave Set: ms_drbd_webfs > Slaves: [ ha2.iohost.com ] > Stopped: [ drbd_webfs:0 ] > > In looking in the recent log I see this: May 20 12:46:42 ha2.iohost.com > pengine: [3117]: info: native_color: Resource fs_webfs cannot run anywhere > > I am not sure why it cannot promote the other resources on ha2, I > checked drbd before putting ha1 on standby and it was up to date. Double check the status of drbd: # cat /proc/drbd Check what the cluster would do, and why: # ptest -LVVV -s [add more Vs to see more detail, but brace yourself for maximum confusion ;-)] Check for constraints that get in the way: # crm configure show | grep -Ee 'location|order' check the "master scores" in the cib: # cibadmin -Ql -o status | grep master Look at the actions that have been performed on the resource, on both nodes: vvvvvvvvvv-- the ID of your primitive # grep "lrmd:.*drbd_mysql" /var/log/ha.log or wherever that ends up on your box > Here are the surrounding log entries, the only thing I changed in the > config is standby="off" on both nodes: > > May 20 12:47:06 ha2.iohost.com pengine: [3117]: notice: group_print: > Resource Group: WebServices > May 20 12:47:06 ha2.iohost.com pengine: [3117]: notice: native_print: > ip1 (ocf::heartbeat:IPaddr2): Started ha2.iohost.com > May 20 12:47:06 ha2.iohost.com pengine: [3117]: notice: native_print: > ip1arp (ocf::heartbeat:SendArp): Started ha2.iohost.com > May 20 12:47:06 ha2.iohost.com pengine: [3117]: notice: native_print: > fs_webfs (ocf::heartbeat:Filesystem): Stopped > May 20 12:47:06 ha2.iohost.com pengine: [3117]: notice: native_print: > fs_mysql (ocf::heartbeat:Filesystem): Stopped > May 20 12:47:06 ha2.iohost.com pengine: [3117]: notice: native_print: > apache2 (lsb:httpd): Stopped > May 20 12:47:06 ha2.iohost.com pengine: [3117]: notice: native_print: > mysql (ocf::heartbeat:mysql): Stopped > May 20 12:47:06 ha2.iohost.com pengine: [3117]: notice: clone_print: > Master/Slave Set: ms_drbd_mysql > May 20 12:47:06 ha2.iohost.com pengine: [3117]: notice: short_print: > Slaves: [ ha2.iohost.com ] > May 20 12:47:06 ha2.iohost.com pengine: [3117]: notice: short_print: > Stopped: [ drbd_mysql:0 ] > May 20 12:47:06 ha2.iohost.com pengine: [3117]: notice: clone_print: > Master/Slave Set: ms_drbd_webfs > May 20 12:47:06 ha2.iohost.com pengine: [3117]: notice: short_print: > Slaves: [ ha1.iohost.com ha2.iohost.com ] > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: rsc_merge_weights: > ip1arp: Breaking dependency loop at ip1 > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: rsc_merge_weights: ip1: > Breaking dependency loop at ip1arp You got a dependency loop? Maybe you should fix that? You put some things in a group in a specific order, then you specify the reverse order in an explicit order and colocation constraint. That is not particularly useful. Either use a group, or use explicit order/colocation constraints, don't try to use both for the same resources. But that's nothing that would affect DRBD at this point. And as long as your DRBD is not (or can not?) be promoted, nothing that depends on it will run, obviously. > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: native_color: Resource > drbd_webfs:0 cannot run anywhere > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: master_color: > ms_drbd_webfs: Promoted 0 instances of a possible 1 to master > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: rsc_merge_weights: > fs_webfs: Rolling back scores from fs_mysql > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: native_color: Resource > fs_webfs cannot run anywhere > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: native_color: Resource > drbd_mysql:0 cannot run anywhere > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: master_color: > ms_drbd_mysql: Promoted 0 instances of a possible 1 to master > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: master_color: > ms_drbd_mysql: Promoted 0 instances of a possible 1 to master > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: rsc_merge_weights: > fs_mysql: Rolling back scores from apache2 > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: native_color: Resource > fs_mysql cannot run anywhere > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: master_color: > ms_drbd_mysql: Promoted 0 instances of a possible 1 to master > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: master_color: > ms_drbd_webfs: Promoted 0 instances of a possible 1 to master > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: rsc_merge_weights: > apache2: Rolling back scores from mysql > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: native_color: Resource > apache2 cannot run anywhere > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: native_color: Resource > mysql cannot run anywhere > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: master_color: > ms_drbd_mysql: Promoted 0 instances of a possible 1 to master > May 20 12:47:06 ha2.iohost.com pengine: [3117]: info: master_color: > ms_drbd_webfs: Promoted 0 instances of a possible 1 to master > > Regards, > Randy -- : Lars Ellenberg : LINBIT | Your Way to High Availability : DRBD/HA support and consulting http://www.linbit.com DRBD® and LINBIT® are registered trademarks of LINBIT, Austria. _______________________________________________ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems