Re: [ClusterLabs] corosync doesn't start any resource
Hello Andrei, > Then you need to set symmetrical="false". yep, now it seems to work now, thank you very much! > I assume this would be "pcs constraint order set ... > symmetrical=false". yes almost: pcs constraint order set nfs-server vm_storage ha-ip action=start setoptions symmetrical=false Thank you very very much! best regards Stefan > Gesendet: Samstag, 23. Juni 2018 um 22:13 Uhr > Von: "Andrei Borzenkov" > An: users@clusterlabs.org > Betreff: Re: [ClusterLabs] corosync doesn't start any resource > > 22.06.2018 11:22, Stefan Krueger пишет: > > Hello Andrei, > > > > thanks for this hint, but I need this "special" order. In an other setup it > > works. > > > > Then you need to set symmetrical="false". Otherwise pacemaker implicitly > creates reverse order which leads to deadlock. I am not intimately > familiar with pcs, I assume this would be "pcs constraint order set ... > symmetrical=false". > > > best regards > > Stefan > > > >> Gesendet: Freitag, 22. Juni 2018 um 06:57 Uhr > >> Von: "Andrei Borzenkov" > >> An: users@clusterlabs.org > >> Betreff: Re: [ClusterLabs] corosync doesn't start any resource > >> > >> 21.06.2018 16:04, Stefan Krueger пишет: > >>> Hi Ken, > >>> > >>>> Can you attach the pe-input file listed just above here? > >>> done ;) > >>> > >>> And thank you for your patience! > >>> > >> > >> You delete all context which makes it hard to answer. This is not web > >> forum where users can simply scroll up to see previous reply. > >> > >> Both your logs and pe-input show that nfs-server and vm-storage wait for > >> each other. > >> > >> My best guess is that you have incorrect ordering for start and stop > >> which causes loop in pacemaker decision. Your start order is "nfs-server > >> vm-storage" and your stop order is "nfs-server vm-storage", while it > >> should normally be symmetrical. Reversing order in one of sets makes it > >> work as intended (verified). > >> > >> I would actually expect that asymmetrical configuration still should > >> work, so I leave it to pacemaker developers to comment whether this is a > >> bug or feature :) > >> > >> ___ > >> Users mailing list: Users@clusterlabs.org > >> https://lists.clusterlabs.org/mailman/listinfo/users > >> > >> Project Home: http://www.clusterlabs.org > >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > >> Bugs: http://bugs.clusterlabs.org > >> > > ___ > > Users mailing list: Users@clusterlabs.org > > https://lists.clusterlabs.org/mailman/listinfo/users > > > > Project Home: http://www.clusterlabs.org > > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > > Bugs: http://bugs.clusterlabs.org > > > > ___ > Users mailing list: Users@clusterlabs.org > https://lists.clusterlabs.org/mailman/listinfo/users > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > ___ Users mailing list: Users@clusterlabs.org https://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
Re: [ClusterLabs] corosync doesn't start any resource
Hello Andrei, thanks for this hint, but I need this "special" order. In an other setup it works. best regards Stefan > Gesendet: Freitag, 22. Juni 2018 um 06:57 Uhr > Von: "Andrei Borzenkov" > An: users@clusterlabs.org > Betreff: Re: [ClusterLabs] corosync doesn't start any resource > > 21.06.2018 16:04, Stefan Krueger пишет: > > Hi Ken, > > > >> Can you attach the pe-input file listed just above here? > > done ;) > > > > And thank you for your patience! > > > > You delete all context which makes it hard to answer. This is not web > forum where users can simply scroll up to see previous reply. > > Both your logs and pe-input show that nfs-server and vm-storage wait for > each other. > > My best guess is that you have incorrect ordering for start and stop > which causes loop in pacemaker decision. Your start order is "nfs-server > vm-storage" and your stop order is "nfs-server vm-storage", while it > should normally be symmetrical. Reversing order in one of sets makes it > work as intended (verified). > > I would actually expect that asymmetrical configuration still should > work, so I leave it to pacemaker developers to comment whether this is a > bug or feature :) > > ___ > Users mailing list: Users@clusterlabs.org > https://lists.clusterlabs.org/mailman/listinfo/users > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > ___ Users mailing list: Users@clusterlabs.org https://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
Re: [ClusterLabs] corosync doesn't start any resource
Hi Ken, > Can you attach the pe-input file listed just above here? done ;) And thank you for your patience! best regards Stefan pre-input-228.bz2 Description: application/bzip ___ Users mailing list: Users@clusterlabs.org https://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
Re: [ClusterLabs] corosync doesn't start any resource
Hi Ken, I don't see any issues in the logs, periodically this is in the logs: Jun 20 11:52:19 [5613] zfs-serv3 crmd: info: crm_timer_popped: PEngine Recheck Timer (I_PE_CALC) just popped (90ms) Jun 20 11:52:19 [5613] zfs-serv3 crmd: notice: do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE | input=I_PE_CALC cause=C_TIMER_POPPED origin=crm_timer_popped Jun 20 11:52:19 [5613] zfs-serv3 crmd: info: do_state_transition: Progressed to state S_POLICY_ENGINE after C_TIMER_POPPED Jun 20 11:52:19 [5612] zfs-serv3pengine: info: process_pe_message: Input has not changed since last time, not saving to disk Jun 20 11:52:19 [5612] zfs-serv3pengine: info: determine_online_status: Node zfs-serv3 is online Jun 20 11:52:19 [5612] zfs-serv3pengine: info: determine_online_status: Node zfs-serv4 is online Jun 20 11:52:19 [5612] zfs-serv3pengine: info: native_print: vm_storage (ocf::heartbeat:ZFS): Stopped Jun 20 11:52:19 [5612] zfs-serv3pengine: info: native_print:ha-ip (ocf::heartbeat:IPaddr2): Stopped Jun 20 11:52:19 [5612] zfs-serv3pengine: info: native_print: resIPMI-zfs4(stonith:external/ipmi):Started zfs-serv3 Jun 20 11:52:19 [5612] zfs-serv3pengine: info: native_print: resIPMI-zfs3(stonith:external/ipmi):Started zfs-serv4 Jun 20 11:52:19 [5612] zfs-serv3pengine: info: native_print: nfs-server (systemd:nfs-server): Stopped Jun 20 11:52:19 [5612] zfs-serv3pengine: info: RecurringOp: Start recurring monitor (5s) for vm_storage on zfs-serv3 Jun 20 11:52:19 [5612] zfs-serv3pengine: info: RecurringOp: Start recurring monitor (10s) for ha-ip on zfs-serv3 Jun 20 11:52:19 [5612] zfs-serv3pengine: info: RecurringOp: Start recurring monitor (60s) for nfs-server on zfs-serv3 Jun 20 11:52:19 [5612] zfs-serv3pengine: notice: LogActions: Start vm_storage (zfs-serv3) Jun 20 11:52:19 [5612] zfs-serv3pengine: notice: LogActions: Start ha-ip (zfs-serv3) Jun 20 11:52:19 [5612] zfs-serv3pengine: info: LogActions: Leave resIPMI-zfs4(Started zfs-serv3) Jun 20 11:52:19 [5612] zfs-serv3pengine: info: LogActions: Leave resIPMI-zfs3(Started zfs-serv4) Jun 20 11:52:19 [5612] zfs-serv3pengine: notice: LogActions: Start nfs-server (zfs-serv3) Jun 20 11:52:19 [5612] zfs-serv3pengine: notice: process_pe_message: Calculated transition 80, saving inputs in /var/lib/pacemaker/pengine/pe-input-228.bz2 Jun 20 11:52:19 [5613] zfs-serv3 crmd: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE | input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response Jun 20 11:52:19 [5613] zfs-serv3 crmd: info: do_te_invoke: Processing graph 80 (ref=pe_calc-dc-1529488339-113) derived from /var/lib/pacemaker/pengine/pe-input-228.bz2 Jun 20 11:52:19 [5613] zfs-serv3 crmd: warning: run_graph: Transition 80 (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=6, Source=/var/lib/pacemaker/pengine/pe-input-228.bz2): Terminated Jun 20 11:52:19 [5613] zfs-serv3 crmd: warning: te_graph_trigger: Transition failed: terminated Jun 20 11:52:19 [5613] zfs-serv3 crmd: notice: print_graph: Graph 80 with 6 actions: batch-limit=0 jobs, network-delay=6ms Jun 20 11:52:19 [5613] zfs-serv3 crmd: notice: print_synapse: [Action 5]: Pending rsc op vm_storage_monitor_5000 on zfs-serv3 (priority: 0, waiting: 4) Jun 20 11:52:19 [5613] zfs-serv3 crmd: notice: print_synapse: [Action 4]: Pending rsc op vm_storage_start_0 on zfs-serv3 (priority: 0, waiting: 12) Jun 20 11:52:19 [5613] zfs-serv3 crmd: notice: print_synapse: [Action 7]: Pending rsc op ha-ip_monitor_1 on zfs-serv3 (priority: 0, waiting: 6) Jun 20 11:52:19 [5613] zfs-serv3 crmd: notice: print_synapse: [Action 6]: Pending rsc op ha-ip_start_0 on zfs-serv3 (priority: 0, waiting: 4 12) Jun 20 11:52:19 [5613] zfs-serv3 crmd: notice: print_synapse: [Action 13]: Pending rsc op nfs-server_monitor_6on zfs-serv3 (priority: 0, waiting: 12) Jun 20 11:52:19 [5613] zfs-serv3 crmd: notice: print_synapse: [Action 12]: Pending rsc op nfs-server_start_0 on zfs-serv3 (priority: 0, waiting: 4) Jun 20 11:52:19 [5613] zfs-serv3 crmd: info: do_log: Input I_TE_SUCCESS received in state S_TRANSITION_ENGINE from notify_crmd Jun 20 11:52:19 [5613] zfs-serv3 crmd: notice: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE | input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd Jun 20 12:07:19 [5613] zfs-serv3 crmd: info: crm_timer_popped: PEngine
Re: [ClusterLabs] corosync doesn't start any resource
Hi Ken, thanks for help! I create a stonith-device and delete the no-quorum-policy. It doesn't change anything, so I delete the orders, (co)locations and one ressource (nfs-server). at first it works fine but when I stop a cluster via 'pcs cluster stop' it takes infinity time, it looks like it has an problem with the nfs server so I tried to stop them manuelly via systemctl stop nfs-server, but it didn't change anything - the nfs-server won't stop. So I did a reset the server, now everything should move to the other node but it also didn't happen :( Manually I can start/stop the nfs-server without any problems (nobody mount the nfs-share yet): systemctl start nfs-server.service ; sleep 5; systemctl status nfs-server.service ; sleep 5; systemctl stop nfs-server so, again my ressources won't start pcs status Cluster name: zfs-vmstorage Stack: corosync Current DC: zfs-serv3 (version 1.1.16-94ff4df) - partition with quorum Last updated: Tue Jun 19 16:15:37 2018 Last change: Tue Jun 19 15:41:24 2018 by hacluster via crmd on zfs-serv4 2 nodes configured 5 resources configured Online: [ zfs-serv3 zfs-serv4 ] Full list of resources: vm_storage (ocf::heartbeat:ZFS): Stopped ha-ip (ocf::heartbeat:IPaddr2): Stopped resIPMI-zfs4 (stonith:external/ipmi):Started zfs-serv3 resIPMI-zfs3 (stonith:external/ipmi):Started zfs-serv4 nfs-server (systemd:nfs-server): Stopped Daemon Status: corosync: active/enabled pacemaker: active/enabled pcsd: active/enabled pcs config Cluster Name: zfs-vmstorage Corosync Nodes: zfs-serv3 zfs-serv4 Pacemaker Nodes: zfs-serv3 zfs-serv4 Resources: Resource: vm_storage (class=ocf provider=heartbeat type=ZFS) Attributes: pool=vm_storage importargs="-d /dev/disk/by-vdev/" Operations: monitor interval=5s timeout=30s (vm_storage-monitor-interval-5s) start interval=0s timeout=90 (vm_storage-start-interval-0s) stop interval=0s timeout=90 (vm_storage-stop-interval-0s) Resource: ha-ip (class=ocf provider=heartbeat type=IPaddr2) Attributes: ip=172.16.101.73 cidr_netmask=16 Operations: start interval=0s timeout=20s (ha-ip-start-interval-0s) stop interval=0s timeout=20s (ha-ip-stop-interval-0s) monitor interval=10s timeout=20s (ha-ip-monitor-interval-10s) Resource: nfs-server (class=systemd type=nfs-server) Operations: start interval=0s timeout=100 (nfs-server-start-interval-0s) stop interval=0s timeout=100 (nfs-server-stop-interval-0s) monitor interval=60 timeout=100 (nfs-server-monitor-interval-60) Stonith Devices: Resource: resIPMI-zfs4 (class=stonith type=external/ipmi) Attributes: hostname=ipmi-zfs-serv4 ipaddr=172.xx.xx.17 userid=USER passwd=GEHEIM interface=lan Operations: monitor interval=60s (resIPMI-zfs4-monitor-interval-60s) Resource: resIPMI-zfs3 (class=stonith type=external/ipmi) Attributes: hostname=ipmi-zfs-serv3 ipaddr=172.xx.xx.16 userid=USER passwd=GEHEIM interface=lan Operations: monitor interval=60s (resIPMI-zfs3-monitor-interval-60s) Fencing Levels: Location Constraints: Resource: resIPMI-zfs3 Disabled on: zfs-serv3 (score:-INFINITY) (id:location-resIPMI-zfs3-zfs-serv3--INFINITY) Resource: resIPMI-zfs4 Disabled on: zfs-serv4 (score:-INFINITY) (id:location-resIPMI-zfs4-zfs-serv4--INFINITY) Ordering Constraints: Resource Sets: set nfs-server vm_storage ha-ip action=start (id:pcs_rsc_set_nfs-server_vm_storage_ha-ip) (id:pcs_rsc_order_set_nfs-server_vm_storage_ha-ip) set ha-ip nfs-server vm_storage action=stop (id:pcs_rsc_set_ha-ip_nfs-server_vm_storage) (id:pcs_rsc_order_set_ha-ip_nfs-server_vm_storage) Colocation Constraints: Resource Sets: set ha-ip nfs-server vm_storage (id:colocation-ha-ip-nfs-server-INFINITY-0) setoptions score=INFINITY (id:colocation-ha-ip-nfs-server-INFINITY) Ticket Constraints: Alerts: No alerts defined Resources Defaults: resource-stickiness: 100 Operations Defaults: No defaults set Cluster Properties: cluster-infrastructure: corosync cluster-name: zfs-vmstorage dc-version: 1.1.16-94ff4df have-watchdog: false last-lrm-refresh: 1528814481 no-quorum-policy: stop stonith-enabled: false Quorum: Options: thanks for help! best regards Stefan ___ Users mailing list: Users@clusterlabs.org https://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
Re: [ClusterLabs] questions about fence_scsi
Hello Andrei, thanks for this hint, at the moment I try to solve this with a colocation. (but it doesn't work, see mailinglist) best regards > Gesendet: Freitag, 15. Juni 2018 um 10:32 Uhr > Von: "Andrei Borzenkov" > An: "Cluster Labs - All topics related to open-source clustering welcomed" > > Betreff: Re: [ClusterLabs] questions about fence_scsi > > On Fri, Jun 15, 2018 at 11:18 AM, Andrei Borzenkov > wrote: > > On Fri, Jun 15, 2018 at 10:14 AM, Stefan Krueger wrote: > >> Hello, > >> > >> so far as I understand I can use fence_scsi on a two node cluster, if the > >> fence running on one cluster the other cluster has no access to this > >> devices, correct? > > > > If I parse this sentence correctly - no, that's not correct to my best > > knowledge. All active nodes have access to shared resource - only when > > node fails will it be fenced (i.e. access to devices revoked), and > > unfenced again (i.e. access granted) when node comes back. > > > > If you really want to allow only one node at a time to access device, > you should look at something like sg_persist RA. See example in SLES > focumentation (scroll down to the end of page): > > https://www.suse.com/documentation/sle-ha-12/book_sleha/data/sec_ha_storage_protect_fencing.html > > > >> I've a 2node cluster with shared JBODs and configure fence_scsi, but I > >> still can use/mount all this devices on both nodes. Did I something wrong? > >> My Config: > >> > >> pcs resource > >> Resource Group: zfs-storage > >> vm_storage (ocf::heartbeat:ZFS): Started zfs-serv3 > >> ha-ip (ocf::heartbeat:IPaddr2): Started zfs-serv3 > >> root@zfs-serv4:~# pcs config > >> Cluster Name: zfs-vmstorage > >> Corosync Nodes: > >> zfs-serv3 zfs-serv4 > >> Pacemaker Nodes: > >> zfs-serv3 zfs-serv4 > >> > >> Resources: > >> Group: zfs-storage > >> Resource: vm_storage (class=ocf provider=heartbeat type=ZFS) > >>Attributes: pool=vm_storage importargs="-d /dev/disk/by-vdev/" > >>Operations: monitor interval=5s timeout=30s > >> (vm_storage-monitor-interval-5s) > >>start interval=0s timeout=90 (vm_storage-start-interval-0s) > >>stop interval=0s timeout=90 (vm_storage-stop-interval-0s) > >> Resource: ha-ip (class=ocf provider=heartbeat type=IPaddr2) > >>Attributes: ip=172.16.101.73 cidr_netmask=16 > >>Operations: start interval=0s timeout=20s (ha-ip-start-interval-0s) > >>stop interval=0s timeout=20s (ha-ip-stop-interval-0s) > >>monitor interval=10s timeout=20s > >> (ha-ip-monitor-interval-10s) > >> > >> Stonith Devices: > >> Resource: fence-vm_storage (class=stonith type=fence_scsi) > >> Attributes: pcmk_monitor_action=metadata > >> pcmk_host_list=172.16.101.74,172.16.101.75 devices=" > >> /dev/disk/by-vdev/j3d03-hdd /dev/disk/by-vdev/j4d03-hdd > >> /dev/disk/by-vdev/j3d04-hdd /dev/disk/by-vdev/j4d04-hdd > >> /dev/disk/by-vdev/j3d05-hdd /dev/disk/by-vdev/j4d05-hdd > >> /dev/disk/by-vdev/j3d06-hdd /dev/disk/by-vdev/j4d06-hdd > >> /dev/disk/by-vdev/j3d07-hdd /dev/disk/by-vdev/j4d07-hdd > >> /dev/disk/by-vdev/j3d08-hdd /dev/disk/by-vdev/j4d08-hdd > >> /dev/disk/by-vdev/j3d09-hdd /dev/disk/by-vdev/j4d09-hdd > >> /dev/disk/by-vdev/j3d10-hdd /dev/disk/by-vdev/j4d10-hdd > >> /dev/disk/by-vdev/j3d11-hdd /dev/disk/by-vdev/j4d11-hdd > >> /dev/disk/by-vdev/j3d12-hdd /dev/disk/by-vdev/j4d12-hdd > >> /dev/disk/by-vdev/j3d13-hdd /dev/disk/by-vdev/j4d13-hdd > >> /dev/disk/by-vdev/j3d14-hdd /dev/disk/by-vdev/j4d14-hdd > >> /dev/disk/by-vdev/j3d15-hdd /dev/disk/by-vdev/j4d15-hdd > >> /dev/disk/by-vdev/j3d16-hdd /dev/disk/by-vdev/j4d16-hdd > >> /dev/disk/by-vdev/j3d17-hdd /dev/disk/by-vdev/j4d17-hdd > >> /dev/disk/by-vdev/j3d18-hdd /dev/disk/by-vdev/j4d18-hdd / de > v/d > >> isk/by-vdev/j3d19-hdd /dev/disk/by-vdev/j4d19-hdd log > >> /dev/disk/by-vdev/j3d00-ssd /dev/disk/by-vdev/j4d00-ssd cache > >> /dev/disk/by-vdev/j3d02-ssd" > >> Meta Attrs: provides=unfencing > >> Operations: monitor interval=60s (fence-vm_storage-monitor-interval-60s) > >> Fencing Levels: > >> > >> Location Constraints: > >> Ordering Constraints: > >> Colocation Constraints: > >> Ticket Constraints: > >> > >> Alerts: > >>
[ClusterLabs] corosync doesn't start any resource
Hello, corosync doesn't start any ressource and I don't know why. I tried to stop/start the cluster, I also tried to reboot it but it doesn't help. also in the logs I dont find nothing that could be useful IMHO. It would be very nice if someone can help me. pcs status Cluster name: zfs-vmstorage Stack: corosync Current DC: zfs-serv3 (version 1.1.16-94ff4df) - partition with quorum Last updated: Fri Jun 15 14:42:32 2018 Last change: Fri Jun 15 14:17:23 2018 by root via cibadmin on zfs-serv3 2 nodes configured 3 resources configured Online: [ zfs-serv3 zfs-serv4 ] Full list of resources: nfs-server (systemd:nfs-server): Stopped vm_storage (ocf::heartbeat:ZFS): Stopped ha-ip (ocf::heartbeat:IPaddr2): Stopped Daemon Status: corosync: active/enabled pacemaker: active/enabled pcsd: active/enabled pcs config Cluster Name: zfs-vmstorage Corosync Nodes: zfs-serv3 zfs-serv4 Pacemaker Nodes: zfs-serv3 zfs-serv4 Resources: Resource: nfs-server (class=systemd type=nfs-server) Operations: start interval=0s timeout=100 (nfs-server-start-interval-0s) stop interval=0s timeout=100 (nfs-server-stop-interval-0s) monitor interval=60 timeout=100 (nfs-server-monitor-interval-60) Resource: vm_storage (class=ocf provider=heartbeat type=ZFS) Attributes: pool=vm_storage importargs="-d /dev/disk/by-vdev/" Operations: monitor interval=5s timeout=30s (vm_storage-monitor-interval-5s) start interval=0s timeout=90 (vm_storage-start-interval-0s) stop interval=0s timeout=90 (vm_storage-stop-interval-0s) Resource: ha-ip (class=ocf provider=heartbeat type=IPaddr2) Attributes: ip=172.16.101.73 cidr_netmask=16 Operations: start interval=0s timeout=20s (ha-ip-start-interval-0s) stop interval=0s timeout=20s (ha-ip-stop-interval-0s) monitor interval=10s timeout=20s (ha-ip-monitor-interval-10s) Stonith Devices: Fencing Levels: Location Constraints: Ordering Constraints: Resource Sets: set nfs-server vm_storage ha-ip action=start (id:pcs_rsc_set_nfs-server_vm_storage_ha-ip) (id:pcs_rsc_order_set_nfs-server_vm_storage_ha-ip) set ha-ip nfs-server vm_storage action=stop (id:pcs_rsc_set_ha-ip_nfs-server_vm_storage) (id:pcs_rsc_order_set_ha-ip_nfs-server_vm_storage) Colocation Constraints: Resource Sets: set ha-ip nfs-server vm_storage (id:colocation-ha-ip-nfs-server-INFINITY-0) setoptions score=INFINITY (id:colocation-ha-ip-nfs-server-INFINITY) Ticket Constraints: Alerts: No alerts defined Resources Defaults: resource-stickiness: 100 Operations Defaults: No defaults set Cluster Properties: cluster-infrastructure: corosync cluster-name: zfs-vmstorage dc-version: 1.1.16-94ff4df have-watchdog: false last-lrm-refresh: 1528814481 no-quorum-policy: ignore stonith-enabled: false Quorum: Options: and here are the Log-files https://paste.debian.net/hidden/9376add7/ best regards Stefan ___ Users mailing list: Users@clusterlabs.org https://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
[ClusterLabs] questions about fence_scsi
Hello, so far as I understand I can use fence_scsi on a two node cluster, if the fence running on one cluster the other cluster has no access to this devices, correct? I've a 2node cluster with shared JBODs and configure fence_scsi, but I still can use/mount all this devices on both nodes. Did I something wrong? My Config: pcs resource Resource Group: zfs-storage vm_storage (ocf::heartbeat:ZFS): Started zfs-serv3 ha-ip (ocf::heartbeat:IPaddr2): Started zfs-serv3 root@zfs-serv4:~# pcs config Cluster Name: zfs-vmstorage Corosync Nodes: zfs-serv3 zfs-serv4 Pacemaker Nodes: zfs-serv3 zfs-serv4 Resources: Group: zfs-storage Resource: vm_storage (class=ocf provider=heartbeat type=ZFS) Attributes: pool=vm_storage importargs="-d /dev/disk/by-vdev/" Operations: monitor interval=5s timeout=30s (vm_storage-monitor-interval-5s) start interval=0s timeout=90 (vm_storage-start-interval-0s) stop interval=0s timeout=90 (vm_storage-stop-interval-0s) Resource: ha-ip (class=ocf provider=heartbeat type=IPaddr2) Attributes: ip=172.16.101.73 cidr_netmask=16 Operations: start interval=0s timeout=20s (ha-ip-start-interval-0s) stop interval=0s timeout=20s (ha-ip-stop-interval-0s) monitor interval=10s timeout=20s (ha-ip-monitor-interval-10s) Stonith Devices: Resource: fence-vm_storage (class=stonith type=fence_scsi) Attributes: pcmk_monitor_action=metadata pcmk_host_list=172.16.101.74,172.16.101.75 devices=" /dev/disk/by-vdev/j3d03-hdd /dev/disk/by-vdev/j4d03-hdd /dev/disk/by-vdev/j3d04-hdd /dev/disk/by-vdev/j4d04-hdd /dev/disk/by-vdev/j3d05-hdd /dev/disk/by-vdev/j4d05-hdd /dev/disk/by-vdev/j3d06-hdd /dev/disk/by-vdev/j4d06-hdd /dev/disk/by-vdev/j3d07-hdd /dev/disk/by-vdev/j4d07-hdd /dev/disk/by-vdev/j3d08-hdd /dev/disk/by-vdev/j4d08-hdd /dev/disk/by-vdev/j3d09-hdd /dev/disk/by-vdev/j4d09-hdd /dev/disk/by-vdev/j3d10-hdd /dev/disk/by-vdev/j4d10-hdd /dev/disk/by-vdev/j3d11-hdd /dev/disk/by-vdev/j4d11-hdd /dev/disk/by-vdev/j3d12-hdd /dev/disk/by-vdev/j4d12-hdd /dev/disk/by-vdev/j3d13-hdd /dev/disk/by-vdev/j4d13-hdd /dev/disk/by-vdev/j3d14-hdd /dev/disk/by-vdev/j4d14-hdd /dev/disk/by-vdev/j3d15-hdd /dev/disk/by-vdev/j4d15-hdd /dev/disk/by-vdev/j3d16-hdd /dev/disk/by-vdev/j4d16-hdd /dev/disk/by-vdev/j3d17-hdd /dev/disk/by-vdev/j4d17-hdd /dev/disk/by-vdev/j3d18-hdd /dev/disk/by-vdev/j4d18-hdd /dev/d isk/by-vdev/j3d19-hdd /dev/disk/by-vdev/j4d19-hdd log /dev/disk/by-vdev/j3d00-ssd /dev/disk/by-vdev/j4d00-ssd cache /dev/disk/by-vdev/j3d02-ssd" Meta Attrs: provides=unfencing Operations: monitor interval=60s (fence-vm_storage-monitor-interval-60s) Fencing Levels: Location Constraints: Ordering Constraints: Colocation Constraints: Ticket Constraints: Alerts: No alerts defined Resources Defaults: resource-stickiness: 100 Operations Defaults: No defaults set Cluster Properties: cluster-infrastructure: corosync cluster-name: zfs-vmstorage dc-version: 1.1.16-94ff4df have-watchdog: false last-lrm-refresh: 1528814481 no-quorum-policy: ignore Quorum: Options: Thanks for help! Best regards Stefan ___ Users mailing list: Users@clusterlabs.org https://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
[ClusterLabs] ressource stopped unexpectly
Hello, I've a problem with my cluster. When I use 'pcs cluster standby serv3' it moves all ressources to serv4 that works fine, but when I restart a node the Ressource ha-ip become stopped and I don't know why. Can somebody give me an hint why this happen and how to resolve that? btw: i use this guide: https://github.com/ewwhite/zfs-ha/wiki the logfile is here (i guess it is too long for the mailinglist) https://paste.debian.net/hidden/2e001867/ thanks for help! best regards Stefan pcs status Cluster name: zfs-vmstorage Stack: corosync Current DC: zfs-serv3 (version 1.1.16-94ff4df) - partition with quorum Last updated: Tue Jun 12 16:56:45 2018 Last change: Tue Jun 12 16:44:52 2018 by hacluster via crm_attribute on zfs-serv3 2 nodes configured 3 resources configured Online: [ zfs-serv3 zfs-serv4 ] Full list of resources: fence-vm_storage (stonith:fence_scsi): Started zfs-serv3 Resource Group: zfs-storage vm_storage (ocf::heartbeat:ZFS): Started zfs-serv3 ha-ip (ocf::heartbeat:IPaddr2): Stopped Daemon Status: corosync: active/enabled pacemaker: active/enabled pcsd: active/enabled pcs config Cluster Name: zfs-vmstorage Corosync Nodes: zfs-serv3 zfs-serv4 Pacemaker Nodes: zfs-serv3 zfs-serv4 Resources: Group: zfs-storage Resource: vm_storage (class=ocf provider=heartbeat type=ZFS) Attributes: pool=vm_storage importargs="-d /dev/disk/by-vdev/" Operations: monitor interval=5s timeout=30s (vm_storage-monitor-interval-5s) start interval=0s timeout=90 (vm_storage-start-interval-0s) stop interval=0s timeout=90 (vm_storage-stop-interval-0s) Resource: ha-ip (class=ocf provider=heartbeat type=IPaddr2) Attributes: ip=172.16.101.73 cidr_netmask=16 Operations: start interval=0s timeout=20s (ha-ip-start-interval-0s) stop interval=0s timeout=20s (ha-ip-stop-interval-0s) monitor interval=10s timeout=20s (ha-ip-monitor-interval-10s) Stonith Devices: Resource: fence-vm_storage (class=stonith type=fence_scsi) Attributes: pcmk_monitor_action=metadata pcmk_host_list=172.16.101.74,172.16.101.75 devices=" /dev/disk/by-vdev/j3d03-hdd /dev/disk/by-vdev/j4d03-hdd /dev/disk/by-vdev/j3d04-hdd /dev/disk/by-vdev/j4d04-hdd /dev/disk/by-vdev/j3d05-hdd /dev/disk/by-vdev/j4d05-hdd /dev/disk/by-vdev/j3d06-hdd /dev/disk/by-vdev/j4d06-hdd /dev/disk/by-vdev/j3d07-hdd /dev/disk/by-vdev/j4d07-hdd /dev/disk/by-vdev/j3d08-hdd /dev/disk/by-vdev/j4d08-hdd /dev/disk/by-vdev/j3d09-hdd /dev/disk/by-vdev/j4d09-hdd /dev/disk/by-vdev/j3d10-hdd /dev/disk/by-vdev/j4d10-hdd /dev/disk/by-vdev/j3d11-hdd /dev/disk/by-vdev/j4d11-hdd /dev/disk/by-vdev/j3d12-hdd /dev/disk/by-vdev/j4d12-hdd /dev/disk/by-vdev/j3d13-hdd /dev/disk/by-vdev/j4d13-hdd /dev/disk/by-vdev/j3d14-hdd /dev/disk/by-vdev/j4d14-hdd /dev/disk/by-vdev/j3d15-hdd /dev/disk/by-vdev/j4d15-hdd /dev/disk/by-vdev/j3d16-hdd /dev/disk/by-vdev/j4d16-hdd /dev/disk/by-vdev/j3d17-hdd /dev/disk/by-vdev/j4d17-hdd /dev/disk/by-vdev/j3d18-hdd /dev/disk/by-vdev/j4d18-hdd /dev/d isk/by-vdev/j3d19-hdd /dev/disk/by-vdev/j4d19-hdd log /dev/disk/by-vdev/j3d00-ssd /dev/disk/by-vdev/j4d00-ssd cache /dev/disk/by-vdev/j3d02-ssd" Meta Attrs: provides=unfencing Operations: monitor interval=60s (fence-vm_storage-monitor-interval-60s) Fencing Levels: Location Constraints: Ordering Constraints: Colocation Constraints: Ticket Constraints: Alerts: No alerts defined Resources Defaults: resource-stickiness: 100 Operations Defaults: No defaults set Cluster Properties: cluster-infrastructure: corosync cluster-name: zfs-vmstorage dc-version: 1.1.16-94ff4df have-watchdog: false last-lrm-refresh: 1528814481 no-quorum-policy: ignore Quorum: Options: ___ Users mailing list: Users@clusterlabs.org https://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
[ClusterLabs] dont (re)start a ressource if there is already running
Hello, I've a configuration where 2nfs servers are running on different nodes and if one node goes down move a ressources to the other one, but dont try to restart the ressource, because it is already running. My config looks like this: node 1084815873: zfs-serv1 node 1084815874: zfs-serv2 primitive HA_IP-Serv1 IPaddr2 \ params ip=172.16.101.70 cidr_netmask=16 \ op monitor interval=20 timeout=30 on-fail=restart nic=bond0 \ meta target-role=Started primitive HA_IP-Serv2 IPaddr2 \ params ip=172.16.101.74 cidr_netmask=16 \ op monitor interval=10s nic=bond0 primitive nc_storage ZFS \ params pool=nc_storage importargs="-d /dev/disk/by-partlabel/" primitive nfs-serv1 systemd:nfs-server \ op monitor interval=5s primitive nfs-serv2 systemd:nfs-server \ op monitor interval=5s primitive vm_storage ZFS \ params pool=vm_storage importargs="-d /dev/disk/by-partlabel/" colocation compl_zfs-serv1 +inf: HA_IP-Serv1 nc_storage nfs-serv1 colocation compl_zfs-serv2 +inf: HA_IP-Serv2 vm_storage nfs-serv2 order start_serv1 nfs-serv1:start nc_storage:start HA_IP-Serv1:start symmetrical=false order start_serv2 nfs-serv2:start vm_storage:start HA_IP-Serv2:start symmetrical=false order stop_serv1 HA_IP-Serv1:stop nfs-serv1:stop nc_storage:stop symmetrical=false order stop_serv2 HA_IP-Serv2:stop nfs-serv2:stop vm_storage:stop symmetrical=false property cib-bootstrap-options: \ have-watchdog=false \ dc-version=1.1.16-94ff4df \ cluster-infrastructure=corosync \ cluster-name=debian \ no-quorum-policy=ignore \ default-resource-stickiness=100 \ stonith-enabled=false \ last-lrm-refresh=1509546667 So is it possible to check if the ressource nfs is already running on the other node? best regards stefan ___ Users mailing list: Users@clusterlabs.org http://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
Re: [ClusterLabs] different start/stop order
Hello Ken, thanks for this information! > You can make ordering constraints asymmetrical, > so they only apply in the listed direction: > > http://clusterlabs.org/doc/en-US/Pacemaker/1.1-pcs/html-single/Pacemaker_Explained/index.html#s-resource-ordering Ok but i see only how can i create a start order, but how can i create a different stop order? Best regards Stefan ___ Users mailing list: Users@clusterlabs.org http://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
[ClusterLabs] different start/stop order
Hello, is it possible to start a ressource in a different way then they stopped? eg: primitive HA_IP-Serv1 IPaddr2 \ params ip=172.16.101.70 cidr_netmask=16 \ op monitor interval=20 timeout=30 on-fail=restart nic=bond0 \ meta target-role=Started primitive HA_IP-Serv2 IPaddr2 \ params ip=172.16.101.74 cidr_netmask=16 \ op monitor interval=10s nic=bond0 primitive nc_storage ZFS \ params pool=nc_storage importargs="-d /dev/disk/by-partlabel/" primitive nfs-serv1 systemd:nfs-server primitive nfs-serv2 systemd:nfs-server group compl_zfs-serv1 nc_storage nfs-serv1 HA_IP-Serv1 I want that pacemaker stop it in this order: HA_IP-Serv1 nfs-serv1 nc_storage but start it in this order: nfs-serv1 nc_storage HA_IP-Serv1 it that possible and how can I achieve that? best regards Stefan ___ Users mailing list: Users@clusterlabs.org http://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
[ClusterLabs] can't move/migrate ressource
Hello, when i try to migrate a ressource from one server to an other (for example for maintenance), it don't work. a single ressource works fine, after that I create a group with 2 ressources and try to move that. my config is: crm conf show node 739272007: zfs-serv1 node 739272008: zfs-serv2 primitive HA_IP-Serv1 IPaddr2 \ params ip=172.16.101.70 cidr_netmask=16 \ op monitor interval=20 timeout=30 on-fail=restart nic=bond0 \ meta target-role=Started primitive HA_IP-Serv2 IPaddr2 \ params ip=172.16.101.74 cidr_netmask=16 \ op monitor interval=10s nic=bond0 primitive nc_storage ZFS \ params pool=nc_storage importargs="-d /dev/disk/by-partlabel/" group compl_zfs-serv1 nc_storage HA_IP-Serv1 location cli-prefer-HA_IP-Serv1 compl_zfs-serv1 role=Started inf: zfs-serv1 location cli-prefer-HA_IP-Serv2 HA_IP-Serv2 role=Started inf: zfs-serv2 location cli-prefer-compl_zfs-serv1 compl_zfs-serv1 role=Started inf: zfs-serv2 location cli-prefer-nc_storage compl_zfs-serv1 role=Started inf: zfs-serv1 property cib-bootstrap-options: \ have-watchdog=false \ dc-version=1.1.16-94ff4df \ cluster-infrastructure=corosync \ cluster-name=debian \ no-quorum-policy=ignore \ default-resource-stickiness=100 \ stonith-enabled=false \ last-lrm-refresh=1507702403 command: crm resource move compl_zfs-serv1 zfs-serv2 pacemakerlog from zfs-serv2: Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op: Diff: --- 0.106.0 2 Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op: Diff: +++ 0.107.0 cc224b15d0a796e040b026b7c2965770 Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op: -- /cib/configuration/constraints/rsc_location[@id='cli-prefer-compl_zfs-serv1'] Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op: + /cib: @epoch=107 Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_process_request: Completed cib_delete operation for section constraints: OK (rc=0, origin=zfs-serv1/crm_resource/3, version=0.107.0) Oct 11 13:55:58 [3561] zfs-serv2 crmd: info: abort_transition_graph: Transition aborted by deletion of rsc_location[@id='cli-prefer-compl_zfs-serv1']: Configuration change | cib=0.107.0 source=te_update_diff:444 path=/cib/configuration/constraints/rsc_location[@id='cli-prefer-compl_zfs-serv1'] complete=true Oct 11 13:55:58 [3561] zfs-serv2 crmd: notice: do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE | input=I_PE_CALC cause=C_FSA_INTERNAL origin=abort_transition_graph Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op: Diff: --- 0.107.0 2 Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op: Diff: +++ 0.108.0 (null) Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op: + /cib: @epoch=108 Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op: ++ /cib/configuration/constraints: Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_process_request: Completed cib_modify operation for section constraints: OK (rc=0, origin=zfs-serv1/crm_resource/4, version=0.108.0) Oct 11 13:55:58 [3561] zfs-serv2 crmd: info: abort_transition_graph: Transition aborted by rsc_location.cli-prefer-compl_zfs-serv1 'create': Configuration change | cib=0.108.0 source=te_update_diff:444 path=/cib/configuration/constraints complete=true Oct 11 13:55:58 [3560] zfs-serv2pengine: notice: unpack_config: On loss of CCM Quorum: Ignore Oct 11 13:55:58 [3560] zfs-serv2pengine: info: determine_online_status: Node zfs-serv2 is online Oct 11 13:55:58 [3560] zfs-serv2pengine: info: determine_online_status: Node zfs-serv1 is online Oct 11 13:55:58 [3560] zfs-serv2pengine: info: determine_op_status: Operation monitor found resource nc_storage active on zfs-serv2 Oct 11 13:55:58 [3560] zfs-serv2pengine: info: native_print: HA_IP-Serv2 (ocf::heartbeat:IPaddr2): Started zfs-serv2 Oct 11 13:55:58 [3560] zfs-serv2pengine: info: group_print: Resource Group: compl_zfs-serv1 Oct 11 13:55:58 [3560] zfs-serv2pengine: info: native_print: nc_storage (ocf::heartbeat:ZFS): Started zfs-serv1 Oct 11 13:55:58 [3560] zfs-serv2pengine: info: native_print: HA_IP-Serv1(ocf::heartbeat:IPaddr2): Started zfs-serv1 Oct 11 13:55:58 [3560] zfs-serv2pengine: info: LogActions: Leave HA_IP-Serv2 (Started zfs-serv2) Oct 11 13:55:58 [3560] zfs-serv2pengine: info: LogActions: Leave nc_storage (Started zfs-serv1) Oct 11 13:55:58 [3560] zfs-serv2pengine: info: LogActions: Leave HA_IP-Serv1 (Started zfs-serv1) Oct 11 13:55:58 [3560] zfs-serv2pengine: notice: process_pe_message: Calculated transition 8, saving inputs in /var/lib/pacemaker/pengine/pe-input-1348.bz2
Re: [ClusterLabs] trouble with IPaddr2
Hello Valentin, thanks for you help > Can you share more info on the network of zfs-serv2, for example: ip a? ip a s 1: lo:mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 2: enp4s0f0: mtu 1500 qdisc mq state UP group default qlen 1000 link/ether ac:1f:6b:xx:xx:xx brd ff:ff:ff:ff:ff:ff inet 172.16.22.126/16 brd 172.16.255.255 scope global enp4s0f0 valid_lft forever preferred_lft forever inet 172.16.101.74/16 brd 172.16.255.255 scope global secondary enp4s0f0 valid_lft forever preferred_lft forever inet6 fe80::ae1f:6bff::/64 scope link valid_lft forever preferred_lft forever 3: enp4s0f1: mtu 1500 qdisc noop state DOWN group default qlen 1000 link/ether ac:1f:6b:xx:xx:xx brd ff:ff:ff:ff:ff:ff 4: ens2f0: mtu 1500 qdisc mq master bond0 state UP group default qlen 1000 link/ether 3c:fd:fe:xx:xx:xx brd ff:ff:ff:ff:ff:ff 5: ens2f1: mtu 1500 qdisc mq master bond0 state UP group default qlen 1000 link/ether 3c:fd:fe:xx:xx:xx brd ff:ff:ff:ff:ff:ff 6: bond0: mtu 1500 qdisc noqueue state UP group default qlen 1000 link/ether 3c:fd:fe:xx:xx:xx brd ff:ff:ff:ff:ff:ff inet 172.16.101.72/16 brd 172.16.255.255 scope global bond0 valid_lft forever preferred_lft forever inet 172.16.101.74/16 brd 172.16.255.255 scope global secondary bond0 valid_lft forever preferred_lft forever inet6 fe80::3efd:feff::/64 scope link valid_lft forever preferred_lft forever ohh damn.. thanks a lot for this hint.. I delete all the IPs on enp4s0f0, and than it works.. but could you please explain why it now works? why he has a problem with this IPs? best regards ___ Users mailing list: Users@clusterlabs.org http://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
[ClusterLabs] trouble with IPaddr2
Hello, I've a simple setup with just 3 resources (at the moment), the ZFS resource also works fine.BUT my IPaddr2 don't work, and I dont know why and how to resolve that. my config: conf sh node 739272007: zfs-serv1 node 739272008: zfs-serv2 primitive HA_IP-Serv1 IPaddr2 \ params ip=172.16.101.70 cidr_netmask=16 \ op monitor interval=20 timeout=30 on-fail=restart nic=bond0 \ meta target-role=Started primitive HA_IP-Serv2 IPaddr2 \ params ip=172.16.101.74 cidr_netmask=16 \ op monitor interval=10s nic=bond0 primitive nc_storage ZFS \ params pool=nc_storage importargs="-d /dev/disk/by-partlabel/" location cli-prefer-HA_IP-Serv1 HA_IP-Serv1 role=Started inf: zfs-serv1 location cli-prefer-HA_IP-Serv2 HA_IP-Serv2 role=Started inf: zfs-serv2 location cli-prefer-nc_storage nc_storage role=Started inf: zfs-serv2 property cib-bootstrap-options: \ have-watchdog=false \ dc-version=1.1.16-94ff4df \ cluster-infrastructure=corosync \ cluster-name=debian \ no-quorum-policy=ignore \ default-resource-stickiness=100 \ stonith-enabled=false \ last-lrm-refresh=1507702403 command: resource move HA_IP-Serv1 zfs-serv2 pacemaker log from zfs-serv2: Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_process_request: Completed cib_delete operation for section constraints: OK (rc=0, origin=zfs-serv1/crm_resource/3, version=0.82.44) Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: Diff: --- 0.82.44 2 Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: Diff: +++ 0.83.0 (null) Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: + /cib: @epoch=83, @num_updates=0 Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: + /cib/configuration/constraints/rsc_location[@id='cli-prefer-HA_IP-Serv1']: @node=zfs-serv2 Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_process_request: Completed cib_modify operation for section constraints: OK (rc=0, origin=zfs-serv1/crm_resource/4, version=0.83.0) Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: Diff: --- 0.83.0 2 Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: Diff: +++ 0.83.1 (null) Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: + /cib: @num_updates=1 Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: + /cib/status/node_state[@id='739272007']/lrm[@id='739272007']/lrm_resources/lrm_resource[@id='HA_IP-Serv1']/lrm_rsc_op[@id='HA_IP-Serv1_last_0']: @operation_key=HA_IP-Serv1_stop_0, @operation=stop, @crm-debug-origin=do_update_resource, @transition-key=8:1574:0:d4b03c3c-1a4e-4609-86ca-675fa4a2ec8f, @transition-magic=0:0;8:1574:0:d4b03c3c-1a4e-4609-86ca-675fa4a2ec8f, @call-id=55, @last-run=1507702772, @last-rc-change=1507702772, @exec Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=zfs-serv1/crmd/1853, version=0.83.1) Oct 11 08:19:32 [23938] zfs-serv2 crmd: info: do_lrm_rsc_op: Performing key=9:1574:0:d4b03c3c-1a4e-4609-86ca-675fa4a2ec8f op=HA_IP-Serv1_start_0 Oct 11 08:19:32 [23935] zfs-serv2 lrmd: info: log_execute: executing - rsc:HA_IP-Serv1 action:start call_id:17 Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_file_backup: Archived previous version as /var/lib/pacemaker/cib/cib-39.raw Oct 11 08:19:32 [23935] zfs-serv2 lrmd: info: log_finished: finished - rsc:HA_IP-Serv1 action:start call_id:17 pid:27672 exit-code:0 exec-time:48ms queue-time:0ms Oct 11 08:19:32 [23938] zfs-serv2 crmd: info: action_synced_wait: Managed IPaddr2_meta-data_0 process 27735 exited with rc=0 Oct 11 08:19:32 [23938] zfs-serv2 crmd: notice: process_lrm_event: Result of start operation for HA_IP-Serv1 on zfs-serv2: 0 (ok) | call=17 key=HA_IP-Serv1_start_0 confirmed=true cib-update=15 Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_process_request: Forwarding cib_modify operation for section status to all (origin=local/crmd/15) Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: Diff: --- 0.83.1 2 Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: Diff: +++ 0.83.2 (null) Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: + /cib: @num_updates=2 Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: + /cib/status/node_state[@id='739272008']/lrm[@id='739272008']/lrm_resources/lrm_resource[@id='HA_IP-Serv1']/lrm_rsc_op[@id='HA_IP-Serv1_last_0']: @operation_key=HA_IP-Serv1_start_0, @operation=start, @transition-key=9:1574:0:d4b03c3c-1a4e-4609-86ca-675fa4a2ec8f, @transition-magic=0:0;9:1574:0:d4b03c3c-1a4e-4609-86ca-675fa4a2ec8f, @call-id=17, @rc-code=0, @last-run=1507702772,