On Tue, Aug 10, 2021 at 12:05 PM Milan Zamazal <mzama...@redhat.com> wrote: > > Yedidyah Bar David <d...@redhat.com> writes: > > > On Tue, Aug 3, 2021 at 10:27 PM Michal Skrivanek < > > michal.skriva...@redhat.com> wrote: > > > >> > >> > >> On 3. 8. 2021, at 11:43, Yedidyah Bar David <d...@redhat.com> wrote: > >> > >> On Tue, Aug 3, 2021 at 10:05 AM Yedidyah Bar David <d...@redhat.com> > >> wrote: > >> > >> > >> On Tue, Aug 3, 2021 at 7:50 AM <jenk...@jenkins.phx.ovirt.org> wrote: > >> > >> > >> Project: > >> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/ > >> Build: > >> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2126/ > >> Build Number: 2126 > >> Build Status: Failure > >> Triggered By: Started by timer > >> > >> ------------------------------------- > >> Changes Since Last Success: > >> ------------------------------------- > >> Changes for Build #2126 > >> [Michal Skrivanek] basic: skipping just the VNC console part of > >> test_virtual_machines > >> > >> > >> > >> > >> ----------------- > >> Failed Tests: > >> ----------------- > >> 2 tests failed. > >> FAILED: > >> > >> he-basic-suite-master.test-scenarios.test_012_local_maintenance_sdk.test_local_maintenance > >> > >> Error Message: > >> ovirtsdk4.Error: Failed to read response: [(<pycurl.Curl object at > >> 0x5555faf11228>, 7, 'Failed to connect to 192.168.200.99 port 443: > >> Connection refused')] > >> > >> > >> This looks very similar to the issue we have with dns/dig failures > >> that cause the engine VM to go down, and it's similar, but different. > >> > >> dig didn't fail (it now uses TCP), but something else caused the agent > >> to stop the engine VM - a combination of high cpu load and low free > >> memory, after restarting the engine VM as part of test_008. > >> > >> > >> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2126/artifact/exported-artifacts/test_logs/ost-he-basic-suite-master-host-0/var/log/ovirt-hosted-engine-ha/agent.log > >> : > >> > >> > >> ===================================================================================== > >> MainThread::INFO::2021-08-03 > >> > >> 06:46:55,068::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Current state ReinitializeFSM (score: 0) > >> MainThread::INFO::2021-08-03 > >> > >> 06:47:04,089::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check) > >> Global maintenance detected > >> MainThread::INFO::2021-08-03 > >> > >> 06:47:04,169::brokerlink::73::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify) > >> Success, was notification of state_transition > >> (ReinitializeFSM-GlobalMaintenance) sent? ignored > >> MainThread::INFO::2021-08-03 > >> > >> 06:47:05,249::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Current state GlobalMaintenance (score: 3400) > >> MainThread::INFO::2021-08-03 > >> > >> 06:47:14,439::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check) > >> Global maintenance detected > >> MainThread::INFO::2021-08-03 > >> > >> 06:47:25,526::states::176::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score) > >> Penalizing score by 814 due to cpu load > >> MainThread::INFO::2021-08-03 > >> > >> 06:47:25,527::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Current state GlobalMaintenance (score: 2586) > >> MainThread::INFO::2021-08-03 > >> > >> 06:47:25,537::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check) > >> Global maintenance detected > >> MainThread::INFO::2021-08-03 > >> > >> 06:47:26,029::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Current state GlobalMaintenance (score: 2586) > >> MainThread::INFO::2021-08-03 > >> > >> 06:47:35,050::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check) > >> Global maintenance detected > >> MainThread::INFO::2021-08-03 > >> > >> 06:47:35,576::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Current state GlobalMaintenance (score: 2586) > >> MainThread::INFO::2021-08-03 > >> > >> 06:47:45,597::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check) > >> Global maintenance detected > >> MainThread::INFO::2021-08-03 > >> > >> 06:47:46,521::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Current state GlobalMaintenance (score: 2586) > >> MainThread::INFO::2021-08-03 > >> > >> 06:47:55,577::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check) > >> Global maintenance detected > >> MainThread::INFO::2021-08-03 > >> > >> 06:47:56,559::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Current state GlobalMaintenance (score: 2586) > >> MainThread::INFO::2021-08-03 > >> > >> 06:47:56,559::hosted_engine::525::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Best remote host ost-he-basic-suite-master-host-1 (id: 2, score: 3400) > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:05,633::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check) > >> Global maintenance detected > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:06,188::states::176::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score) > >> Penalizing score by 820 due to cpu load > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:06,188::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Current state GlobalMaintenance (score: 2580) > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:16,256::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check) > >> Global maintenance detected > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:16,950::states::176::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score) > >> Penalizing score by 831 due to cpu load > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:16,951::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Current state GlobalMaintenance (score: 2569) > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:26,053::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check) > >> Global maintenance detected > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:26,999::states::176::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score) > >> Penalizing score by 839 due to cpu load > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:26,999::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Current state GlobalMaintenance (score: 2561) > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:36,026::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check) > >> Global maintenance detected > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:36,802::states::176::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score) > >> Penalizing score by 844 due to cpu load > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:36,802::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Current state GlobalMaintenance (score: 2556) > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:45,827::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check) > >> Global maintenance detected > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:46,401::states::176::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score) > >> Penalizing score by 849 due to cpu load > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:46,401::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Current state GlobalMaintenance (score: 2551) > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:56,588::brokerlink::73::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify) > >> Success, was notification of state_transition > >> (GlobalMaintenance-ReinitializeFSM) sent? ignored > >> MainThread::INFO::2021-08-03 > >> > >> 06:48:58,685::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Current state ReinitializeFSM (score: 0) > >> MainThread::INFO::2021-08-03 > >> > >> 06:49:05,729::brokerlink::73::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify) > >> Success, was notification of state_transition > >> (ReinitializeFSM-EngineStarting) sent? ignored > >> MainThread::INFO::2021-08-03 > >> > >> 06:49:06,146::states::176::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score) > >> Penalizing score by 830 due to cpu load > >> MainThread::INFO::2021-08-03 > >> > >> 06:49:06,146::states::72::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_penalize_memory) > >> Penalizing score by 400 due to free memory 1782 being lower than > >> required 3171 > >> MainThread::INFO::2021-08-03 > >> > >> 06:49:06,146::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Current state EngineStarting (score: 2170) > >> MainThread::INFO::2021-08-03 > >> > >> 06:49:06,150::state_decorators::95::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check) > >> Timeout cleared while transitioning <class > >> 'ovirt_hosted_engine_ha.agent.states.EngineStarting'> -> <class > >> 'ovirt_hosted_engine_ha.agent.states.EngineUp'> > >> MainThread::INFO::2021-08-03 > >> > >> 06:49:06,172::brokerlink::73::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify) > >> Success, was notification of state_transition > >> (EngineStarting-EngineUp) sent? ignored > >> MainThread::INFO::2021-08-03 > >> > >> 06:49:06,178::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop) > >> Current state EngineUp (score: 2570) > >> MainThread::ERROR::2021-08-03 > >> > >> 06:49:16,197::states::398::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(consume) > >> Host ost-he-basic-suite-master-host-1 (id 2) score is significantly > >> better than local score, shutting down VM on this host > >> > >> ===================================================================================== > >> > >> I looked a bit at /var/log/messages of the host, and while there is > >> quite some noise there, can't tell specifically what might have caused > >> the high cpu load. > >> > >> Also looked at logs of previous runs, and there is indeed a similar > >> pattern, where the cpuload on the host causes the agent to penalize > >> the score, but according to the agent log the load goes down faster, > >> up to the point of being quite low when the engine is up, and only > >> then we exit global maintenance, allowing the agent to take actions. > >> > >> I now ran it again, but think that otherwise I'll ignore this for now, > >> unless we see more similar failures. If we do, we might want to > >> check/monitor/log the cpu load on the hosts, and/or change test_008 to > >> wait, after the engine is up, until the cpu load on the host goes down > >> a bit. Before the port to pytest we waited a hard-coded 5 minutes and > >> I > >> changed that then to only wait until the engine VM is not migrating, > >> and this worked more-or-less ok. We might need to refine that, > >> although I'd rather not introduce another arbitrary long delays but > >> wait for some condition. > >> > >> I also noticed that this ran was started at "03:04 AM" (UTC), by > >> timer, whereas all previous timer-based runs started at 01:55 AM, > >> somewhat earlier - perhaps there are some other things that run at > >> that time that cause load. > >> > >> > >> Checked some more and found another difference: The total cpu use was > >> not very different between the failed and the last successful runs, > >> but load caused by the engine vm was different between them - in the > >> failed run, engine kept being '0.0' for a longer time, compared with > >> the successful run. Since the agent tries to ignore the load by the > >> engine vm, the final result is significantly different. The load-by > >> -engine-vm is calculated based on the result of a VDSM getStats call > >> on the engine VM. Indeed, in vdsm.log, cpuUser and cpuSys are both > >> starting with '0.00' for both runs, but the failed one is '0.00' for > >> a longer time - 45 seconds between VM.create and first-non-zero > >> cpuUser on the successful run, vs 93 seconds for the failed one. > >> Does this make sense (from VDSM's POV)? If so, perhaps there is > >> a way to get from VDSM the answer to this question: "Is this 0.00 > >> CPU usage a result of still-initial, non-accurate, non-complete > >> monitoring? Or is it really not using the CPU at all?" and use it > >> in the agent somehow. > >> > >> > >> I think it’s futile to try to understand that on jenkins.ovirt.org runs. > >> It’s too unpredictable since it runs in a shared environment, you can’t > >> know what’s going on on the same physical host. > >> > > > > I understand that I can't get/expect stable performance/load there. That's > > not my question. I asked about VDSM. I don't know the code there well. I > > assume that when it starts a VM, it initializes something somewhere with > > 'cpuUser: 0.00', > > Yes. > > > and once its gets _some_ data, whatever that might be, it starts > > reporting (based on) this data. > > Yes. It retrieves the data in 15 seconds intervals by default and it > needs 2 retrievals to compute the data (the difference between the 2 > samples). It gets the data from libvirt for running VMs. And then it > waits until Engine calls getStats, which I believe is also every 15 > seconds. So under normal circumstances it should take about 15-45 > seconds after the VM starts being running in libvirt before first real > data is reported to Engine. > > I can think about the following circumstances when it could be delayed: > > - The VM is detected as non-responsive in libvirt. Then the VM is not > queried for its data this time and the following 40 seconds. > > - The executor gets overfilled. > > - Engine doesn't ask for the data for some reason. > > - There is something very wrong in the system. > > None of these should happen if everything is all right but sometimes not > everything is all right, especially when there is something stuck in > QEMU monitoring (e.g. because of temporary storage problems). > > > My question was if there is a reliable way to know if '0.00' is just > > real data, or initial-not-yet-real data. > > From this data itself not but from the accompanied data probably yes. > I'd say that if memory stats are present then it's a real value.
That's indeed what I see also in the referenced vdsm.log. Now pushed: https://gerrit.ovirt.org/c/ovirt-hosted-engine-ha/+/116145 Please review. Thanks! -- Didi _______________________________________________ Devel mailing list -- devel@ovirt.org To unsubscribe send an email to devel-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/devel@ovirt.org/message/LP4OFM4H3VYFY3OVHIHSWLI7MT6P4UTY/