On Thu, 15 Nov 2018 at 13:11, Dafna Ron <d...@redhat.com> wrote: > I am checking the failed jobs > However, Please note that I think you are confusing issues. > Currently, we (CI) have a problem in the job that syncs the package to the > snapshot repo. this jobs run nightly and we had no way of knowing it would > fail until today. > Before today, we had several regressions which lasted for two weeks which > means no package was build at all. > So different issues > > It should be fixed now
> > On Thu, Nov 15, 2018 at 10:54 AM Dan Kenigsberg <dan...@redhat.com> wrote: > >> On Thu, Nov 15, 2018 at 12:45 PM Eyal Edri <ee...@redhat.com> wrote: >> > >> > >> > >> > On Thu, Nov 15, 2018 at 12:43 PM Dan Kenigsberg <dan...@redhat.com> >> wrote: >> >> >> >> On Wed, Nov 14, 2018 at 5:07 PM Dan Kenigsberg <dan...@redhat.com> >> wrote: >> >> > >> >> > On Wed, Nov 14, 2018 at 12:42 PM Dominik Holler <dhol...@redhat.com> >> wrote: >> >> > > >> >> > > On Wed, 14 Nov 2018 11:24:10 +0100 >> >> > > Michal Skrivanek <mskri...@redhat.com> wrote: >> >> > > >> >> > > > > On 14 Nov 2018, at 10:50, Dominik Holler <dhol...@redhat.com> >> wrote: >> >> > > > > >> >> > > > > On Wed, 14 Nov 2018 09:27:39 +0100 >> >> > > > > Dominik Holler <dhol...@redhat.com> wrote: >> >> > > > > >> >> > > > >> On Tue, 13 Nov 2018 13:01:09 +0100 >> >> > > > >> Martin Perina <mper...@redhat.com> wrote: >> >> > > > >> >> >> > > > >>> On Tue, Nov 13, 2018 at 12:49 PM Michal Skrivanek < >> mskri...@redhat.com> >> >> > > > >>> wrote: >> >> > > > >>> >> >> > > > >>>> >> >> > > > >>>> >> >> > > > >>>> On 13 Nov 2018, at 12:20, Dominik Holler < >> dhol...@redhat.com> wrote: >> >> > > > >>>> >> >> > > > >>>> On Tue, 13 Nov 2018 11:56:37 +0100 >> >> > > > >>>> Martin Perina <mper...@redhat.com> wrote: >> >> > > > >>>> >> >> > > > >>>> On Tue, Nov 13, 2018 at 11:02 AM Dafna Ron <d...@redhat.com> >> wrote: >> >> > > > >>>> >> >> > > > >>>> Martin? can you please look at the patch that Dominik sent? >> >> > > > >>>> We need to resolve this as we have not had an engine build >> for the last 11 >> >> > > > >>>> days >> >> > > > >>>> >> >> > > > >>>> >> >> > > > >>>> Yesterday I've merged Dominik's revert patch >> >> > > > >>>> https://gerrit.ovirt.org/95377 >> >> > > > >>>> which should switch cluster level back to 4.2. Below >> mentioned change >> >> > > > >>>> https://gerrit.ovirt.org/95310 is relevant only to cluster >> level 4.3, am I >> >> > > > >>>> right Michal? >> >> > > > >>>> >> >> > > > >>>> The build mentioned >> >> > > > >>>> >> >> > > > >>>> >> https://jenkins.ovirt.org/view/Change%20queue%20jobs/job/ovirt-master_change-queue-tester/11121/ >> >> > > > >>>> is from yesterday. Are we sure that it was executed only >> after #95377 was >> >> > > > >>>> merged? I'd like to see the results from latest >> >> > > > >>>> >> >> > > > >>>> >> https://jenkins.ovirt.org/view/Change%20queue%20jobs/job/ovirt-master_change-queue-tester/11127/ >> >> > > > >>>> but unfortunately it already waits more than an hour for >> available hosts >> >> > > > >>>> ... >> >> > > > >>>> >> >> > > > >>>> >> >> > > > >>>> >> >> > > > >>>> >> >> > > > >>>> >> >> > > > >>>> https://gerrit.ovirt.org/#/c/95283/ results in >> >> > > > >>>> >> >> > > > >>>> >> http://jenkins.ovirt.org/job/ovirt-engine_master_build-artifacts-el7-x86_64/8071/ >> >> > > > >>>> which is used in >> >> > > > >>>> >> >> > > > >>>> >> https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/3489/parameters/ >> >> > > > >>>> results in run_vms succeeding. >> >> > > > >>>> >> >> > > > >>>> The next merged change >> >> > > > >>>> https://gerrit.ovirt.org/#/c/95310/ results in >> >> > > > >>>> >> >> > > > >>>> >> http://jenkins.ovirt.org/job/ovirt-engine_master_build-artifacts-el7-x86_64/8072/ >> >> > > > >>>> which is used in >> >> > > > >>>> >> >> > > > >>>> >> https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/3490/parameters/ >> >> > > > >>>> results in run_vms failing with >> >> > > > >>>> 2018-11-12 17:35:10,109-05 INFO >> >> > > > >>>> [org.ovirt.engine.core.bll.RunVmOnceCommand] (default >> task-1) >> >> > > > >>>> [6930b632-5593-4481-bf2a-a1d8b14a583a] Running command: >> RunVmOnceCommand >> >> > > > >>>> internal: false. Entities affected : ID: >> >> > > > >>>> d10aa133-b9b6-455d-8137-ab822d1c1971 Type: VMAction group >> RUN_VM with role >> >> > > > >>>> type USER >> >> > > > >>>> 2018-11-12 17:35:10,113-05 DEBUG >> >> > > > >>>> >> [org.ovirt.engine.core.common.di.interceptor.DebugLoggingInterceptor] >> >> > > > >>>> (default task-1) [6930b632-5593-4481-bf2a-a1d8b14a583a] >> method: >> >> > > > >>>> getVmManager, params: >> [d10aa133-b9b6-455d-8137-ab822d1c1971], timeElapsed: >> >> > > > >>>> 4ms >> >> > > > >>>> 2018-11-12 17:35:10,128-05 DEBUG >> >> > > > >>>> >> [org.ovirt.engine.core.common.di.interceptor.DebugLoggingInterceptor] >> >> > > > >>>> (default task-1) [6930b632-5593-4481-bf2a-a1d8b14a583a] >> method: >> >> > > > >>>> getAllForClusterWithStatus, params: >> [2ca9ccd8-61f0-470c-ba3f-07766202f260, >> >> > > > >>>> Up], timeElapsed: 7ms >> >> > > > >>>> 2018-11-12 17:35:10,129-05 INFO >> >> > > > >>>> [org.ovirt.engine.core.bll.scheduling.SchedulingManager] >> (default task-1) >> >> > > > >>>> [6930b632-5593-4481-bf2a-a1d8b14a583a] Candidate host >> >> > > > >>>> 'lago-basic-suite-master-host-1' >> ('282860ab-8873-4702-a2be-100a6da111af') >> >> > > > >>>> was filtered out by 'VAR__FILTERTYPE__INTERNAL' filter >> 'CPU-Level' >> >> > > > >>>> (correlation id: 6930b632-5593-4481-bf2a-a1d8b14a583a) >> >> > > > >>>> 2018-11-12 17:35:10,129-05 INFO >> >> > > > >>>> [org.ovirt.engine.core.bll.scheduling.SchedulingManager] >> (default task-1) >> >> > > > >>>> [6930b632-5593-4481-bf2a-a1d8b14a583a] Candidate host >> >> > > > >>>> 'lago-basic-suite-master-host-0' >> ('c48eca36-ea98-46b2-8473-f184833e68a8') >> >> > > > >>>> was filtered out by 'VAR__FILTERTYPE__INTERNAL' filter >> 'CPU-Level' >> >> > > > >>>> (correlation id: 6930b632-5593-4481-bf2a-a1d8b14a583a) >> >> > > > >>>> 2018-11-12 17:35:10,130-05 ERROR >> [org.ovirt.engine.core.bll.RunVmCommand] >> >> > > > >>>> (default task-1) [6930b632-5593-4481-bf2a-a1d8b14a583a] >> Can't find VDS to >> >> > > > >>>> run the VM 'd10aa133-b9b6-455d-8137-ab822d1c1971' on, so >> this VM will not >> >> > > > >>>> be run. >> >> > > > >>>> in >> >> > > > >>>> >> >> > > > >>>> >> https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/3490/artifact/exported-artifacts/test_logs/basic-suite-master/post-004_basic_sanity.py/lago-basic-suite-master-engine/_var_log/ovirt-engine/engine.log/*view*/ >> >> > > > >>>> >> >> > > > >>>> Is this helpful for you? >> >> > > > >>>> >> >> > > > >>>> >> >> > > > >>>> >> >> > > > >>>> actually, there ire two issues >> >> > > > >>>> 1) cluster is still 4.3 even after Martin’s revert. >> >> > > > >>>> >> >> > > > >>> >> >> > > > >>> https://gerrit.ovirt.org/#/c/95409/ should align cluster >> level with dc level >> >> > > > >>> >> >> > > > >> >> >> > > > >> This change aligns the cluster level, but >> >> > > > >> >> https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/3502/parameters/ >> >> > > > >> consuming build result from >> >> > > > >> >> https://jenkins.ovirt.org/view/Change%20queue%20jobs/job/ovirt-master_change-queue-tester/11121/ >> >> > > > >> looks like that this does not solve the issue: >> >> > > > >> File >> "/home/jenkins/workspace/ovirt-system-tests_manual/ovirt-system-tests/basic-suite-master/test-scenarios/004_basic_sanity.py", >> line 698, in run_vms >> >> > > > >> api.vms.get(VM0_NAME).start(start_params) >> >> > > > >> File >> "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/brokers.py", line >> 31193, in start >> >> > > > >> headers={"Correlation-Id":correlation_id} >> >> > > > >> File >> "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/proxy.py", line >> 122, in request >> >> > > > >> persistent_auth=self.__persistent_auth >> >> > > > >> File >> "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/connectionspool.py", >> line 79, in do_request >> >> > > > >> persistent_auth) >> >> > > > >> File >> "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/connectionspool.py", >> line 162, in __do_request >> >> > > > >> raise errors.RequestError(response_code, response_reason, >> response_body) >> >> > > > >> RequestError: >> >> > > > >> status: 400 >> >> > > > >> reason: Bad Request >> >> > > > >> >> >> > > > >> engine.log: >> >> > > > >> 2018-11-14 03:10:36,802-05 INFO >> [org.ovirt.engine.core.bll.scheduling.SchedulingManager] (default task-3) >> [99e282ea-577a-4dab-857b-285b1df5e6f6] Candidate host >> 'lago-basic-suite-master-host-0' ('4dbfb937-ac4b-4cef-8ae3-124944829add') >> was filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'CPU-Level' >> (correlation id: 99e282ea-577a-4dab-857b-285b1df5e6f6) >> >> > > > >> 2018-11-14 03:10:36,802-05 INFO >> [org.ovirt.engine.core.bll.scheduling.SchedulingManager] (default task-3) >> [99e282ea-577a-4dab-857b-285b1df5e6f6] Candidate host >> 'lago-basic-suite-master-host-1' ('731e5055-706e-4310-a062-045e32ffbfeb') >> was filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'CPU-Level' >> (correlation id: 99e282ea-577a-4dab-857b-285b1df5e6f6) >> >> > > > >> 2018-11-14 03:10:36,802-05 ERROR >> [org.ovirt.engine.core.bll.RunVmCommand] (default task-3) >> [99e282ea-577a-4dab-857b-285b1df5e6f6] Can't find VDS to run the VM >> 'dc1e1e92-1e5c-415e-8ac2-b919017adf40' on, so this VM will not be run. >> >> > > > >> >> >> > > > >> >> >> > > > > >> >> > > > > >> >> > > > > https://gerrit.ovirt.org/#/c/95283/ results in >> >> > > > > >> http://jenkins.ovirt.org/job/ovirt-engine_master_build-artifacts-el7-x86_64/8071/ >> >> > > > > which is used in >> >> > > > > >> https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/3504/parameters/ >> >> > > > > results in run_vms succeeding. >> >> > > > > >> >> > > > > The next merged change >> >> > > > > https://gerrit.ovirt.org/#/c/95310/ results in >> >> > > > > >> http://jenkins.ovirt.org/job/ovirt-engine_master_build-artifacts-el7-x86_64/8072/ >> >> > > > > which is used in >> >> > > > > >> https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/3505/parameters/ >> >> > > > > results in run_vms failing with >> >> > > > > File >> "/home/jenkins/workspace/ovirt-system-tests_manual/ovirt-system-tests/basic-suite-master/test-scenarios/004_basic_sanity.py", >> line 698, in run_vms >> >> > > > > api.vms.get(VM0_NAME).start(start_params) >> >> > > > > File >> "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/brokers.py", line >> 31193, in start >> >> > > > > headers={"Correlation-Id":correlation_id} >> >> > > > > File >> "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/proxy.py", line >> 122, in request >> >> > > > > persistent_auth=self.__persistent_auth >> >> > > > > File >> "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/connectionspool.py", >> line 79, in do_request >> >> > > > > persistent_auth) >> >> > > > > File >> "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/connectionspool.py", >> line 162, in __do_request >> >> > > > > raise errors.RequestError(response_code, response_reason, >> response_body) >> >> > > > > RequestError: >> >> > > > > status: 400 >> >> > > > > reason: Bad Request >> >> > > > > >> >> > > > > >> >> > > > > So even if the Cluster Level should be 4.2 now, >> >> > > > > still https://gerrit.ovirt.org/#/c/95310/ seems influence the >> behavior. >> >> > > > >> >> > > > I really do not see how it can affect 4.2. >> >> > > >> >> > > Me neither. >> >> > > >> >> > > > Are you sure the cluster is really 4.2? Sadly it’s not being >> logged at all >> >> > > >> >> > > screenshot from local execution https://imgur.com/a/yiWBw3c >> >> > > >> >> > > > But if it really seem to matter (and since it needs a fix anyway >> for 4.3) feel free to revert it of course >> >> > > > >> >> > > >> >> > > I will post a revert change and check if this changes the behavior. >> >> > >> >> > Dominik, thanks for the research and for Martin's and your >> >> > reverts/fixes. Finally Engine passes OST >> >> > >> https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/11153/ >> >> > and QE can expect a build tomorrow, after 2 weeks of droughts. >> >> >> >> unfortunately, the drought continues. >> > >> > >> > Sorry, missing the content or meaning, what does drought means? >> >> Pardon my flowery language. I mean 2 weeks of no ovirt-engine builds. >> >> > >> >> >> >> Barrak tells me that something is broken in the nightly cron job >> >> copying the the tested repo onto the master-snapshot one. >> > >> > >> > Dafna, can you check this? >> > >> >> >> >> >> >> +Edri: please make it a priority to have it fixed. >> > >> > >> > >> > -- >> > >> > Eyal edri >> > >> > >> > MANAGER >> > >> > RHV/CNV DevOps >> > >> > EMEA VIRTUALIZATION R&D >> > >> > >> > Red Hat EMEA >> > >> > TRIED. TESTED. TRUSTED. >> > phone: +972-9-7692018 >> > irc: eedri (on #tlv #rhev-dev #rhev-integ) >> > -- Barak Korren RHV DevOps team , RHCE, RHCi Red Hat EMEA redhat.com | TRIED. TESTED. TRUSTED. | redhat.com/trusted
_______________________________________________ Infra mailing list -- infra@ovirt.org To unsubscribe send an email to infra-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/infra@ovirt.org/message/MMWXTDQD6BBRPCZEY3BC4LYHRVKNXYGZ/