As IL are on independence day, anyone else can merge? https://gerrit.ovirt.org/#/c/99845/
On Thu, May 9, 2019 at 11:30 AM Dafna Ron <d...@redhat.com> wrote: > Thanks Andrej. > I will follow the patch and update. > Dafna > > On Thu, May 9, 2019 at 11:23 AM Andrej Krejcir <akrej...@redhat.com> > wrote: > >> Hi, >> >> Ok, I have posted the reverting patch: >> https://gerrit.ovirt.org/#/c/99845/ >> >> I'm still investigating what is the problem. Sorry for the delay, we had >> a public holiday yesturday. >> >> >> Andrej >> >> On Thu, 9 May 2019 at 11:20, Dafna Ron <d...@redhat.com> wrote: >> >>> Hi, >>> >>> I have not heard back on this issue and ovirt-engine has been broken for >>> the past 3 days. >>> >>> As this does not seem a simple debug and fix I suggest reverting the >>> patch and investigating later. >>> >>> thanks, >>> Dafna >>> >>> >>> >>> On Wed, May 8, 2019 at 9:42 AM Dafna Ron <d...@redhat.com> wrote: >>> >>>> Any news? >>>> >>>> Thanks, >>>> Dafna >>>> >>>> >>>> On Tue, May 7, 2019 at 4:57 PM Dafna Ron <d...@redhat.com> wrote: >>>> >>>>> thanks for the quick reply and investigation. >>>>> Please update me if I can help any further and if you find the cause >>>>> and have a patch let me know. >>>>> Note that ovirt-engine project is broken and if we cannot find the >>>>> cause relatively fast we should consider reverting the patch to allow a >>>>> new >>>>> package to be built in CQ with other changes that were submitted. >>>>> >>>>> Thanks, >>>>> Dafna >>>>> >>>>> >>>>> On Tue, May 7, 2019 at 4:42 PM Andrej Krejcir <akrej...@redhat.com> >>>>> wrote: >>>>> >>>>>> After running a few OSTs manually, it seems that the patch is the >>>>>> cause. Investigating... >>>>>> >>>>>> On Tue, 7 May 2019 at 14:58, Andrej Krejcir <akrej...@redhat.com> >>>>>> wrote: >>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> The issue is probably not caused by the patch. >>>>>>> >>>>>>> This log line means that the VM does not exist in the DB: >>>>>>> >>>>>>> 2019-05-07 06:02:04,215-04 WARN >>>>>>> [org.ovirt.engine.core.bll.MigrateMultipleVmsCommand] >>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [33485140] >>>>>>> Validation >>>>>>> of action 'MigrateMultipleVms' failed for user admin@internal-authz. >>>>>>> Reasons: ACTION_TYPE_FAILED_VMS_NOT_FOUND >>>>>>> >>>>>>> I will investigate more, why the VM is missing. >>>>>>> >>>>>>> On Tue, 7 May 2019 at 14:07, Dafna Ron <d...@redhat.com> wrote: >>>>>>> >>>>>>>> Hi, >>>>>>>> >>>>>>>> We are failing test upgrade_hosts on >>>>>>>> upgrade-from-release-suite-master. >>>>>>>> From the logs I can see that we are calling migrate vm when we have >>>>>>>> only one host and the vm seem to have been shut down before the >>>>>>>> maintenance >>>>>>>> call is issued. >>>>>>>> >>>>>>>> Can you please look into this? >>>>>>>> >>>>>>>> suspected patch reported as root cause by CQ is: >>>>>>>> >>>>>>>> https://gerrit.ovirt.org/#/c/98920/ - core: Add MigrateMultipleVms >>>>>>>> command and use it for host maintenance >>>>>>>> >>>>>>>> >>>>>>>> logs are found here: >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> http://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/14021/artifact/upgrade-from-release-suite.el7.x86_64/test_logs/upgrade-from-release-suite-master/post-004_basic_sanity.py/ >>>>>>>> >>>>>>>> >>>>>>>> I can see the issue is vm migration when putting host in >>>>>>>> maintenance: >>>>>>>> >>>>>>>> >>>>>>>> 2019-05-07 06:02:04,170-04 INFO >>>>>>>> [org.ovirt.engine.core.bll.MaintenanceVdsCommand] >>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) >>>>>>>> [05592db2-f859-487b-b779-4b32eec5bab >>>>>>>> 3] Running command: MaintenanceVdsCommand internal: true. Entities >>>>>>>> affected : ID: 38e1379b-c3b6-4a2e-91df-d1f346e414a9 Type: VDS >>>>>>>> 2019-05-07 06:02:04,215-04 WARN >>>>>>>> [org.ovirt.engine.core.bll.MigrateMultipleVmsCommand] >>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [33485140] >>>>>>>> Validation >>>>>>>> of action >>>>>>>> 'MigrateMultipleVms' failed for user admin@internal-authz. >>>>>>>> Reasons: ACTION_TYPE_FAILED_VMS_NOT_FOUND >>>>>>>> 2019-05-07 06:02:04,221-04 ERROR >>>>>>>> [org.ovirt.engine.core.bll.MaintenanceVdsCommand] >>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [33485140] >>>>>>>> Failed to >>>>>>>> migrate one or >>>>>>>> more VMs. >>>>>>>> 2019-05-07 06:02:04,227-04 ERROR >>>>>>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [33485140] EVEN >>>>>>>> T_ID: VDS_MAINTENANCE_FAILED(17), Failed to switch Host >>>>>>>> lago-upgrade-from-release-suite-master-host-0 to Maintenance mode. >>>>>>>> 2019-05-07 06:02:04,239-04 INFO >>>>>>>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] Lock >>>>>>>> Acquired to object 'Eng >>>>>>>> ineLock:{exclusiveLocks='[38e1379b-c3b6-4a2e-91df-d1f346e414a9=VDS]', >>>>>>>> sharedLocks=''}' >>>>>>>> 2019-05-07 06:02:04,242-04 INFO >>>>>>>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] >>>>>>>> Running >>>>>>>> command: ActivateVds >>>>>>>> Command internal: true. Entities affected : ID: >>>>>>>> 38e1379b-c3b6-4a2e-91df-d1f346e414a9 Type: VDSAction group >>>>>>>> MANIPULATE_HOST >>>>>>>> with role type ADMIN >>>>>>>> 2019-05-07 06:02:04,243-04 INFO >>>>>>>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] Before >>>>>>>> acquiring lock in ord >>>>>>>> er to prevent monitoring for host >>>>>>>> 'lago-upgrade-from-release-suite-master-host-0' from data-center >>>>>>>> 'test-dc' >>>>>>>> 2019-05-07 06:02:04,243-04 INFO >>>>>>>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] Lock >>>>>>>> acquired, from now a mo >>>>>>>> nitoring of host will be skipped for host >>>>>>>> 'lago-upgrade-from-release-suite-master-host-0' from data-center >>>>>>>> 'test-dc' >>>>>>>> 2019-05-07 06:02:04,252-04 INFO >>>>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] >>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] START, >>>>>>>> SetVdsStatu >>>>>>>> sVDSCommand(HostName = >>>>>>>> lago-upgrade-from-release-suite-master-host-0, >>>>>>>> SetVdsStatusVDSCommandParameters:{hostId='38e1379b-c3b6-4a2e-91df-d1f346e414a9', >>>>>>>> status='Unassigned', n >>>>>>>> onOperationalReason='NONE', stopSpmFailureLogged='false', >>>>>>>> maintenanceReason='null'}), log id: 2c8aa211 >>>>>>>> 2019-05-07 06:02:04,256-04 INFO >>>>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] >>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] >>>>>>>> FINISH, >>>>>>>> SetVdsStat >>>>>>>> usVDSCommand, return: , log id: 2c8aa211 >>>>>>>> 2019-05-07 06:02:04,261-04 INFO >>>>>>>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] >>>>>>>> Activate >>>>>>>> host finished. Lock >>>>>>>> released. Monitoring can run now for host >>>>>>>> 'lago-upgrade-from-release-suite-master-host-0' from data-center >>>>>>>> 'test-dc' >>>>>>>> 2019-05-07 06:02:04,265-04 INFO >>>>>>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] EVEN >>>>>>>> T_ID: VDS_ACTIVATE(16), Activation of host >>>>>>>> lago-upgrade-from-release-suite-master-host-0 initiated by >>>>>>>> admin@internal-authz. >>>>>>>> 2019-05-07 06:02:04,266-04 INFO >>>>>>>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] Lock >>>>>>>> freed >>>>>>>> to object 'Engine >>>>>>>> Lock:{exclusiveLocks='[38e1379b-c3b6-4a2e-91df-d1f346e414a9=VDS]', >>>>>>>> sharedLocks=''}' >>>>>>>> 2019-05-07 06:02:04,484-04 ERROR >>>>>>>> [org.ovirt.engine.core.bll.hostdeploy.HostUpgradeCallback] >>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-96) >>>>>>>> [05592db2-f859-487b-b779-4b32 >>>>>>>> eec5bab3] Host 'lago-upgrade-from-release-suite-master-host-0' >>>>>>>> failed to move to maintenance mode. Upgrade process is terminated. >>>>>>>> >>>>>>>> I can see there was only one vm running: >>>>>>>> >>>>>>>> >>>>>>>> drwxrwxr-x. 2 dron dron 1024 May 7 11:49 qemu >>>>>>>> [dron@dron post-004_basic_sanity.py]$ ls -l >>>>>>>> lago-upgrade-from-release-suite-master-host-0/_var_log/libvirt/qemu/ >>>>>>>> total 6 >>>>>>>> -rw-rw-r--. 1 dron dron 4466 May 7 10:12 vm-with-iface.log >>>>>>>> >>>>>>>> and I can see that there was an attempt to terminate it with an >>>>>>>> error that it does not exist: >>>>>>>> >>>>>>>> >>>>>>>> stroyVmVDSCommandParameters:{hostId='38e1379b-c3b6-4a2e-91df-d1f346e414a9', >>>>>>>> vmId='dfbd75e2-a9cb-4fca-8788-a16954db4abf', secondsToWait='0', >>>>>>>> gracefully='false', reason='', ig >>>>>>>> noreNoVm='false'}), log id: 24278e9b >>>>>>>> 2019-05-07 06:01:41,082-04 INFO >>>>>>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] (default >>>>>>>> task-1) [105f7555-517b-4bf9-b86e-6eb42375de20] START, DestroyVDSComma >>>>>>>> nd(HostName = lago-upgrade-from-release-suite-master-host-0, >>>>>>>> DestroyVmVDSCommandParameters:{hostId='38e1379b-c3b6-4a2e-91df-d1f346e414a9', >>>>>>>> vmId='dfbd75e2-a9cb-4fca-8788-a169 >>>>>>>> 54db4abf', secondsToWait='0', gracefully='false', reason='', >>>>>>>> ignoreNoVm='false'}), log id: 78bba2f8 >>>>>>>> 2019-05-07 06:01:42,090-04 INFO >>>>>>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] (default >>>>>>>> task-1) [105f7555-517b-4bf9-b86e-6eb42375de20] FINISH, DestroyVDSComm >>>>>>>> and, return: , log id: 78bba2f8 >>>>>>>> 2019-05-07 06:01:42,090-04 INFO >>>>>>>> [org.ovirt.engine.core.vdsbroker.DestroyVmVDSCommand] (default task-1) >>>>>>>> [105f7555-517b-4bf9-b86e-6eb42375de20] FINISH, DestroyVmVDSCommand, r >>>>>>>> eturn: , log id: 24278e9b >>>>>>>> 2019-05-07 06:01:42,094-04 INFO >>>>>>>> [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] >>>>>>>> (ForkJoinPool-1-worker-4) [] VM 'dfbd75e2-a9cb-4fca-8788-a16954db4abf' >>>>>>>> was >>>>>>>> reported >>>>>>>> as Down on VDS >>>>>>>> '38e1379b-c3b6-4a2e-91df-d1f346e414a9'(lago-upgrade-from-release-suite-master-host-0) >>>>>>>> 2019-05-07 06:01:42,096-04 INFO >>>>>>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] >>>>>>>> (ForkJoinPool-1-worker-4) [] START, DestroyVDSCommand(HostName = >>>>>>>> lago-upgrade- >>>>>>>> from-release-suite-master-host-0, >>>>>>>> DestroyVmVDSCommandParameters:{hostId='38e1379b-c3b6-4a2e-91df-d1f346e414a9', >>>>>>>> vmId='dfbd75e2-a9cb-4fca-8788-a16954db4abf', secondsToWait='0 >>>>>>>> ', gracefully='false', reason='', ignoreNoVm='true'}), log id: >>>>>>>> 1dbd31eb >>>>>>>> 2019-05-07 06:01:42,114-04 INFO >>>>>>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] >>>>>>>> (ForkJoinPool-1-worker-4) [] Failed to destroy VM >>>>>>>> 'dfbd75e2-a9cb-4fca-8788-a16 >>>>>>>> 954db4abf' because VM does not exist, ignoring >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>
_______________________________________________ Devel mailing list -- devel@ovirt.org To unsubscribe send an email to devel-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/devel@ovirt.org/message/OQ22IENFIVB2SHBJ2WOPO3KECPVZ4NPU/