On 02/16/2017 01:55 PM, Oscar Segarra wrote: > Hi Klaus, > > Thanks a lot, I will try to delete the stop monitor. > > Nevertheless, I have 6 domains configured exactly the same... Is there > any reason why just this domain has this behaviour ?
None I can see here at least. Does the problem persist as is if you shut down the whole cluster and bring it up again? Just to sort out history issues ... > > Thanks a lot. > > 2017-02-16 11:12 GMT+01:00 Klaus Wenninger <kwenn...@redhat.com > <mailto:kwenn...@redhat.com>>: > > On 02/16/2017 11:02 AM, Oscar Segarra wrote: > > Hi Kaluss > > > > Which is your proposal to fix this behavior? > > First you can try to remove the monitor op for role=stopped. > Then the startup-probing will probably still fail but for that > the behaviour is different. > The startup-probing can be disabled globally via cluster-property > enable-startup-probes that is defaulting to true. > But be aware that the cluster then wouldn't be able to react > properly if services are already up when pacemaker is starting. > It should be possible to disable the probing on a per resource > or node basis as well iirc. But I can't tell you out of my mind > how that worked - there was a discussion a few weeks ago > on the list iirc. > > Regards, > Klaus > > > > > Thanks a lot! > > > > > > El 16 feb. 2017 10:57 a. m., "Klaus Wenninger" > <kwenn...@redhat.com <mailto:kwenn...@redhat.com> > > <mailto:kwenn...@redhat.com <mailto:kwenn...@redhat.com>>> escribió: > > > > On 02/16/2017 09:05 AM, Oscar Segarra wrote: > > > Hi, > > > > > > In my environment I have deployed 5 VirtualDomains as one can > > see below: > > > [root@vdicnode01 ~]# pcs status > > > Cluster name: vdic-cluster > > > Stack: corosync > > > Current DC: vdicnode01-priv (version > 1.1.15-11.el7_3.2-e174ec8) - > > > partition with quorum > > > Last updated: Thu Feb 16 09:02:53 2017 Last > change: Thu Feb > > > 16 08:20:53 2017 by root via crm_attribute on vdicnode02-priv > > > > > > 2 nodes and 14 resources configured: 5 resources DISABLED > and 0 > > > BLOCKED from being started due to failures > > > > > > Online: [ vdicnode01-priv vdicnode02-priv ] > > > > > > Full list of resources: > > > > > > nfs-vdic-mgmt-vm-vip (ocf::heartbeat:IPaddr): > Started > > > vdicnode01-priv > > > Clone Set: nfs_setup-clone [nfs_setup] > > > Started: [ vdicnode01-priv vdicnode02-priv ] > > > Clone Set: nfs-mon-clone [nfs-mon] > > > Started: [ vdicnode01-priv vdicnode02-priv ] > > > Clone Set: nfs-grace-clone [nfs-grace] > > > Started: [ vdicnode01-priv vdicnode02-priv ] > > > vm-vdicone01 (ocf::heartbeat:VirtualDomain): FAILED > (disabled)[ > > > vdicnode02-priv vdicnode01-priv ] > > > vm-vdicsunstone01 (ocf::heartbeat:VirtualDomain): FAILED > > > vdicnode01-priv (disabled) > > > vm-vdicdb01 (ocf::heartbeat:VirtualDomain): FAILED > (disabled)[ > > > vdicnode02-priv vdicnode01-priv ] > > > vm-vdicudsserver (ocf::heartbeat:VirtualDomain): FAILED > > > (disabled)[ vdicnode02-priv vdicnode01-priv ] > > > vm-vdicudstuneler (ocf::heartbeat:VirtualDomain): FAILED > > > vdicnode01-priv (disabled) > > > Clone Set: nfs-vdic-images-vip-clone [nfs-vdic-images-vip] > > > Stopped: [ vdicnode01-priv vdicnode02-priv ] > > > > > > Failed Actions: > > > * vm-vdicone01_monitor_20000 on vdicnode02-priv 'not > installed' (5): > > > call=2322, status=complete, exitreason='Configuration file > > > /mnt/nfs-vdic-mgmt-vm/vdicone01.xml does not exist or is not > > readable.', > > > last-rc-change='Thu Feb 16 09:02:07 2017', queued=0ms, > exec=21ms > > > * vm-vdicsunstone01_monitor_20000 on vdicnode02-priv 'not > installed' > > > (5): call=2310, status=complete, exitreason='Configuration > file > > > /mnt/nfs-vdic-mgmt-vm/vdicsunstone01.xml does not exist or > is not > > > readable.', > > > last-rc-change='Thu Feb 16 09:02:07 2017', queued=0ms, > exec=37ms > > > * vm-vdicdb01_monitor_20000 on vdicnode02-priv 'not > installed' (5): > > > call=2320, status=complete, exitreason='Configuration file > > > /mnt/nfs-vdic-mgmt-vm/vdicdb01.xml does not exist or is not > > readable.', > > > last-rc-change='Thu Feb 16 09:02:07 2017', queued=0ms, > exec=35ms > > > * vm-vdicudsserver_monitor_20000 on vdicnode02-priv 'not > installed' > > > (5): call=2321, status=complete, exitreason='Configuration > file > > > /mnt/nfs-vdic-mgmt-vm/vdicudsserver.xml does not exist or > is not > > > readable.', > > > last-rc-change='Thu Feb 16 09:02:07 2017', queued=0ms, > exec=42ms > > > * vm-vdicudstuneler_monitor_20000 on vdicnode01-priv 'not > installed' > > > (5): call=1987183, status=complete, > exitreason='Configuration file > > > /mnt/nfs-vdic-mgmt-vm/vdicudstuneler.xml does not exist or > is not > > > readable.', > > > last-rc-change='Thu Feb 16 04:00:25 2017', queued=0ms, > exec=30ms > > > * vm-vdicdb01_monitor_20000 on vdicnode01-priv 'not > installed' (5): > > > call=2550049, status=complete, exitreason='Configuration file > > > /mnt/nfs-vdic-mgmt-vm/vdicdb01.xml does not exist or is not > > readable.', > > > last-rc-change='Thu Feb 16 08:13:37 2017', queued=0ms, > exec=44ms > > > * nfs-mon_monitor_10000 on vdicnode01-priv 'unknown error' > (1): > > > call=1984009, status=Timed Out, exitreason='none', > > > last-rc-change='Thu Feb 16 04:24:30 2017', queued=0ms, > exec=0ms > > > * vm-vdicsunstone01_monitor_20000 on vdicnode01-priv 'not > installed' > > > (5): call=2552050, status=complete, > exitreason='Configuration file > > > /mnt/nfs-vdic-mgmt-vm/vdicsunstone01.xml does not exist or > is not > > > readable.', > > > last-rc-change='Thu Feb 16 08:14:07 2017', queued=0ms, > exec=22ms > > > * vm-vdicone01_monitor_20000 on vdicnode01-priv 'not > installed' (5): > > > call=2620052, status=complete, exitreason='Configuration file > > > /mnt/nfs-vdic-mgmt-vm/vdicone01.xml does not exist or is not > > readable.', > > > last-rc-change='Thu Feb 16 09:02:53 2017', queued=0ms, > exec=45ms > > > * vm-vdicudsserver_monitor_20000 on vdicnode01-priv 'not > installed' > > > (5): call=2550052, status=complete, > exitreason='Configuration file > > > /mnt/nfs-vdic-mgmt-vm/vdicudsserver.xml does not exist or > is not > > > readable.', > > > last-rc-change='Thu Feb 16 08:13:37 2017', queued=0ms, > exec=48ms > > > > > > > > > Al VirtualDomain resources are configured the same: > > > > > > [root@vdicnode01 cluster]# pcs resource show vm-vdicone01 > > > Resource: vm-vdicone01 (class=ocf provider=heartbeat > > type=VirtualDomain) > > > Attributes: hypervisor=qemu:///system > > > config=/mnt/nfs-vdic-mgmt-vm/vdicone01.xml > > > migration_network_suffix=tcp:// migration_transport=ssh > > > Meta Attrs: allow-migrate=true target-role=Stopped > > > Utilization: cpu=1 hv_memory=512 > > > Operations: start interval=0s timeout=90 > > > (vm-vdicone01-start-interval-0s) > > > stop interval=0s timeout=90 > > (vm-vdicone01-stop-interval-0s) > > > monitor interval=20s role=Stopped > > > (vm-vdicone01-monitor-interval-20s) > > > monitor interval=30s > > (vm-vdicone01-monitor-interval-30s) > > > [root@vdicnode01 cluster]# pcs resource show vm-vdicdb01 > > > Resource: vm-vdicdb01 (class=ocf provider=heartbeat > > type=VirtualDomain) > > > Attributes: hypervisor=qemu:///system > > > config=/mnt/nfs-vdic-mgmt-vm/vdicdb01.xml > > > migration_network_suffix=tcp:// migration_transport=ssh > > > Meta Attrs: allow-migrate=true target-role=Stopped > > > Utilization: cpu=1 hv_memory=512 > > > Operations: start interval=0s timeout=90 > > (vm-vdicdb01-start-interval-0s) > > > stop interval=0s timeout=90 > > (vm-vdicdb01-stop-interval-0s) > > > monitor interval=20s role=Stopped > > > (vm-vdicdb01-monitor-interval-20s) > > > monitor interval=30s > > (vm-vdicdb01-monitor-interval-30s) > > > > > > > > > > > > Nevertheless, one of the virtual domains is logging hardly and > > > fulfilling my hard disk: > > > > > > VirtualDomain(vm-vdicone01)[116359]: > 2017/02/16_08:52:27 INFO: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not > readable, > > > resource considered stopped. > > > VirtualDomain(vm-vdicone01)[116401]: > 2017/02/16_08:52:27 ERROR: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml > does not > > exist > > > or is not readable. > > > VirtualDomain(vm-vdicone01)[116423]: > 2017/02/16_08:52:27 INFO: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not > readable, > > > resource considered stopped. > > > VirtualDomain(vm-vdicone01)[116444]: > 2017/02/16_08:52:27 ERROR: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml > does not > > exist > > > or is not readable. > > > VirtualDomain(vm-vdicone01)[116466]: > 2017/02/16_08:52:27 INFO: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not > readable, > > > resource considered stopped. > > > VirtualDomain(vm-vdicone01)[116487]: > 2017/02/16_08:52:27 ERROR: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml > does not > > exist > > > or is not readable. > > > VirtualDomain(vm-vdicone01)[116509]: > 2017/02/16_08:52:27 INFO: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not > readable, > > > resource considered stopped. > > > VirtualDomain(vm-vdicone01)[116530]: > 2017/02/16_08:52:27 ERROR: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml > does not > > exist > > > or is not readable. > > > VirtualDomain(vm-vdicone01)[116552]: > 2017/02/16_08:52:27 INFO: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not > readable, > > > resource considered stopped. > > > VirtualDomain(vm-vdicone01)[116573]: > 2017/02/16_08:52:27 ERROR: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml > does not > > exist > > > or is not readable. > > > VirtualDomain(vm-vdicone01)[116595]: > 2017/02/16_08:52:27 INFO: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not > readable, > > > resource considered stopped. > > > VirtualDomain(vm-vdicone01)[116616]: > 2017/02/16_08:52:27 ERROR: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml > does not > > exist > > > or is not readable. > > > VirtualDomain(vm-vdicone01)[116638]: > 2017/02/16_08:52:27 INFO: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not > readable, > > > resource considered stopped. > > > VirtualDomain(vm-vdicone01)[116659]: > 2017/02/16_08:52:27 ERROR: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml > does not > > exist > > > or is not readable. > > > VirtualDomain(vm-vdicone01)[116681]: > 2017/02/16_08:52:27 INFO: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not > readable, > > > resource considered stopped. > > > VirtualDomain(vm-vdicone01)[116702]: > 2017/02/16_08:52:27 ERROR: > > > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml > does not > > exist > > > or is not readable. > > > [root@vdicnode01 cluster]# pcs status > > > > > > > > > Note! Is it normal the error as I have not mounted the nfs > > > resource /mnt/nfs-vdic-mgmt-vm/vdicone01.xml yet. > > > > Well that is probably the explanation already: > > The resource should be stopped. The config-file is not > available. > > But the resource needs the config-file to verify that it is > really > > stopped. > > So the probe is failing and as you have a monitoring op for > > role="stopped" it > > is doing that over and over again. > > > > > > > > ¿Is there any explanation for this heavy logging? > > > > > > Thanks a lot! > > > > > > > > > _______________________________________________ > > > Users mailing list: Users@clusterlabs.org > <mailto:Users@clusterlabs.org> > > <mailto:Users@clusterlabs.org <mailto:Users@clusterlabs.org>> > > > http://lists.clusterlabs.org/mailman/listinfo/users > <http://lists.clusterlabs.org/mailman/listinfo/users> > > <http://lists.clusterlabs.org/mailman/listinfo/users > <http://lists.clusterlabs.org/mailman/listinfo/users>> > > > > > > Project Home: http://www.clusterlabs.org > > > Getting started: > > http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf> > > <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf>> > > > Bugs: http://bugs.clusterlabs.org > > > > > > > > _______________________________________________ > > Users mailing list: Users@clusterlabs.org > <mailto:Users@clusterlabs.org> > > <mailto:Users@clusterlabs.org <mailto:Users@clusterlabs.org>> > > http://lists.clusterlabs.org/mailman/listinfo/users > <http://lists.clusterlabs.org/mailman/listinfo/users> > > <http://lists.clusterlabs.org/mailman/listinfo/users > <http://lists.clusterlabs.org/mailman/listinfo/users>> > > > > Project Home: http://www.clusterlabs.org > > Getting started: > > http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf> > > <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf>> > > Bugs: http://bugs.clusterlabs.org > > > > > > > _______________________________________________ > Users mailing list: Users@clusterlabs.org > <mailto:Users@clusterlabs.org> > http://lists.clusterlabs.org/mailman/listinfo/users > <http://lists.clusterlabs.org/mailman/listinfo/users> > > Project Home: http://www.clusterlabs.org > Getting started: > http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf> > Bugs: http://bugs.clusterlabs.org > >
_______________________________________________ Users mailing list: Users@clusterlabs.org http://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org