we have a passing ovirt-engine build today.
Thank you all for a fast response.
Dafna


On Thu, May 9, 2019 at 12:43 PM Sandro Bonazzola <sbona...@redhat.com>
wrote:

>
>
> Il giorno gio 9 mag 2019 alle ore 12:59 Dafna Ron <d...@redhat.com> ha
> scritto:
>
>> As IL are on independence day, anyone else can merge?
>> https://gerrit.ovirt.org/#/c/99845/
>>
>>
> I have merge rights but I need at least CI to pass. Waiting on jenkins.
>
>
>>
>> On Thu, May 9, 2019 at 11:30 AM Dafna Ron <d...@redhat.com> wrote:
>>
>>> Thanks Andrej.
>>> I will follow the patch and update.
>>> Dafna
>>>
>>> On Thu, May 9, 2019 at 11:23 AM Andrej Krejcir <akrej...@redhat.com>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> Ok, I have posted the reverting patch:
>>>> https://gerrit.ovirt.org/#/c/99845/
>>>>
>>>> I'm still investigating what is the problem. Sorry for the delay, we
>>>> had a public holiday yesturday.
>>>>
>>>>
>>>> Andrej
>>>>
>>>> On Thu, 9 May 2019 at 11:20, Dafna Ron <d...@redhat.com> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> I have not heard back on this issue and ovirt-engine has been broken
>>>>> for the past 3 days.
>>>>>
>>>>> As this does not seem a simple debug and fix I suggest reverting the
>>>>> patch and investigating later.
>>>>>
>>>>> thanks,
>>>>> Dafna
>>>>>
>>>>>
>>>>>
>>>>> On Wed, May 8, 2019 at 9:42 AM Dafna Ron <d...@redhat.com> wrote:
>>>>>
>>>>>> Any news?
>>>>>>
>>>>>> Thanks,
>>>>>> Dafna
>>>>>>
>>>>>>
>>>>>> On Tue, May 7, 2019 at 4:57 PM Dafna Ron <d...@redhat.com> wrote:
>>>>>>
>>>>>>> thanks for the quick reply and investigation.
>>>>>>> Please update me if I can help any further and if you find the cause
>>>>>>> and have a patch let me know.
>>>>>>> Note that ovirt-engine project is broken and if we cannot find the
>>>>>>> cause relatively fast we should consider reverting the patch to allow a 
>>>>>>> new
>>>>>>> package to be built in CQ with other changes that were submitted.
>>>>>>>
>>>>>>> Thanks,
>>>>>>> Dafna
>>>>>>>
>>>>>>>
>>>>>>> On Tue, May 7, 2019 at 4:42 PM Andrej Krejcir <akrej...@redhat.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> After running a few OSTs manually, it seems that the patch is the
>>>>>>>> cause. Investigating...
>>>>>>>>
>>>>>>>> On Tue, 7 May 2019 at 14:58, Andrej Krejcir <akrej...@redhat.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Hi,
>>>>>>>>>
>>>>>>>>> The issue is probably not caused by the patch.
>>>>>>>>>
>>>>>>>>> This log line means that the VM does not exist in the DB:
>>>>>>>>>
>>>>>>>>> 2019-05-07 06:02:04,215-04 WARN
>>>>>>>>> [org.ovirt.engine.core.bll.MigrateMultipleVmsCommand]
>>>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [33485140] 
>>>>>>>>> Validation
>>>>>>>>> of action 'MigrateMultipleVms' failed for user admin@internal-authz.
>>>>>>>>> Reasons: ACTION_TYPE_FAILED_VMS_NOT_FOUND
>>>>>>>>>
>>>>>>>>> I will investigate more, why the VM is missing.
>>>>>>>>>
>>>>>>>>> On Tue, 7 May 2019 at 14:07, Dafna Ron <d...@redhat.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hi,
>>>>>>>>>>
>>>>>>>>>> We are failing test upgrade_hosts on
>>>>>>>>>> upgrade-from-release-suite-master.
>>>>>>>>>> From the logs I can see that we are calling migrate vm when we
>>>>>>>>>> have only one host and the vm seem to have been shut down before the
>>>>>>>>>> maintenance call is issued.
>>>>>>>>>>
>>>>>>>>>> Can you please look into this?
>>>>>>>>>>
>>>>>>>>>> suspected patch reported as root cause by CQ is:
>>>>>>>>>>
>>>>>>>>>> https://gerrit.ovirt.org/#/c/98920/ - core: Add
>>>>>>>>>> MigrateMultipleVms command and use it for host maintenance
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> logs are found here:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> http://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/14021/artifact/upgrade-from-release-suite.el7.x86_64/test_logs/upgrade-from-release-suite-master/post-004_basic_sanity.py/
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> I can see the issue is vm migration when putting host in
>>>>>>>>>> maintenance:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> 2019-05-07 06:02:04,170-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.bll.MaintenanceVdsCommand]
>>>>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2)
>>>>>>>>>> [05592db2-f859-487b-b779-4b32eec5bab
>>>>>>>>>> 3] Running command: MaintenanceVdsCommand internal: true.
>>>>>>>>>> Entities affected : ID: 38e1379b-c3b6-4a2e-91df-d1f346e414a9 Type: 
>>>>>>>>>> VDS
>>>>>>>>>> 2019-05-07 06:02:04,215-04 WARN
>>>>>>>>>> [org.ovirt.engine.core.bll.MigrateMultipleVmsCommand]
>>>>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [33485140] 
>>>>>>>>>> Validation
>>>>>>>>>> of action
>>>>>>>>>> 'MigrateMultipleVms' failed for user admin@internal-authz.
>>>>>>>>>> Reasons: ACTION_TYPE_FAILED_VMS_NOT_FOUND
>>>>>>>>>> 2019-05-07 06:02:04,221-04 ERROR
>>>>>>>>>> [org.ovirt.engine.core.bll.MaintenanceVdsCommand]
>>>>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [33485140] 
>>>>>>>>>> Failed to
>>>>>>>>>> migrate one or
>>>>>>>>>> more VMs.
>>>>>>>>>> 2019-05-07 06:02:04,227-04 ERROR
>>>>>>>>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>>>>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [33485140] EVEN
>>>>>>>>>> T_ID: VDS_MAINTENANCE_FAILED(17), Failed to switch Host
>>>>>>>>>> lago-upgrade-from-release-suite-master-host-0 to Maintenance mode.
>>>>>>>>>> 2019-05-07 06:02:04,239-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>>>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] Lock
>>>>>>>>>> Acquired to object 'Eng
>>>>>>>>>> ineLock:{exclusiveLocks='[38e1379b-c3b6-4a2e-91df-d1f346e414a9=VDS]',
>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>> 2019-05-07 06:02:04,242-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>>>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] 
>>>>>>>>>> Running
>>>>>>>>>> command: ActivateVds
>>>>>>>>>> Command internal: true. Entities affected : ID:
>>>>>>>>>> 38e1379b-c3b6-4a2e-91df-d1f346e414a9 Type: VDSAction group 
>>>>>>>>>> MANIPULATE_HOST
>>>>>>>>>> with role type ADMIN
>>>>>>>>>> 2019-05-07 06:02:04,243-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>>>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] 
>>>>>>>>>> Before
>>>>>>>>>> acquiring lock in ord
>>>>>>>>>> er to prevent monitoring for host
>>>>>>>>>> 'lago-upgrade-from-release-suite-master-host-0' from data-center 
>>>>>>>>>> 'test-dc'
>>>>>>>>>> 2019-05-07 06:02:04,243-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>>>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] Lock
>>>>>>>>>> acquired, from now a mo
>>>>>>>>>> nitoring of host will be skipped for host
>>>>>>>>>> 'lago-upgrade-from-release-suite-master-host-0' from data-center 
>>>>>>>>>> 'test-dc'
>>>>>>>>>> 2019-05-07 06:02:04,252-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] 
>>>>>>>>>> START,
>>>>>>>>>> SetVdsStatu
>>>>>>>>>> sVDSCommand(HostName =
>>>>>>>>>> lago-upgrade-from-release-suite-master-host-0,
>>>>>>>>>> SetVdsStatusVDSCommandParameters:{hostId='38e1379b-c3b6-4a2e-91df-d1f346e414a9',
>>>>>>>>>> status='Unassigned', n
>>>>>>>>>> onOperationalReason='NONE', stopSpmFailureLogged='false',
>>>>>>>>>> maintenanceReason='null'}), log id: 2c8aa211
>>>>>>>>>> 2019-05-07 06:02:04,256-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] 
>>>>>>>>>> FINISH,
>>>>>>>>>> SetVdsStat
>>>>>>>>>> usVDSCommand, return: , log id: 2c8aa211
>>>>>>>>>> 2019-05-07 06:02:04,261-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>>>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] 
>>>>>>>>>> Activate
>>>>>>>>>> host finished. Lock
>>>>>>>>>> released. Monitoring can run now for host
>>>>>>>>>> 'lago-upgrade-from-release-suite-master-host-0' from data-center 
>>>>>>>>>> 'test-dc'
>>>>>>>>>> 2019-05-07 06:02:04,265-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>>>>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] EVEN
>>>>>>>>>> T_ID: VDS_ACTIVATE(16), Activation of host
>>>>>>>>>> lago-upgrade-from-release-suite-master-host-0 initiated by
>>>>>>>>>> admin@internal-authz.
>>>>>>>>>> 2019-05-07 06:02:04,266-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>>>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) [70840477] 
>>>>>>>>>> Lock freed
>>>>>>>>>> to object 'Engine
>>>>>>>>>> Lock:{exclusiveLocks='[38e1379b-c3b6-4a2e-91df-d1f346e414a9=VDS]',
>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>> 2019-05-07 06:02:04,484-04 ERROR
>>>>>>>>>> [org.ovirt.engine.core.bll.hostdeploy.HostUpgradeCallback]
>>>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-96)
>>>>>>>>>> [05592db2-f859-487b-b779-4b32
>>>>>>>>>> eec5bab3] Host 'lago-upgrade-from-release-suite-master-host-0'
>>>>>>>>>> failed to move to maintenance mode. Upgrade process is terminated.
>>>>>>>>>>
>>>>>>>>>> I can see there was only one vm running:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> drwxrwxr-x. 2 dron dron 1024 May 7 11:49 qemu
>>>>>>>>>> [dron@dron post-004_basic_sanity.py]$ ls -l
>>>>>>>>>> lago-upgrade-from-release-suite-master-host-0/_var_log/libvirt/qemu/
>>>>>>>>>> total 6
>>>>>>>>>> -rw-rw-r--. 1 dron dron 4466 May 7 10:12 vm-with-iface.log
>>>>>>>>>>
>>>>>>>>>> and I can see that there was an attempt to terminate it with an
>>>>>>>>>> error that it does not exist:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> stroyVmVDSCommandParameters:{hostId='38e1379b-c3b6-4a2e-91df-d1f346e414a9',
>>>>>>>>>> vmId='dfbd75e2-a9cb-4fca-8788-a16954db4abf', secondsToWait='0',
>>>>>>>>>> gracefully='false', reason='', ig
>>>>>>>>>> noreNoVm='false'}), log id: 24278e9b
>>>>>>>>>> 2019-05-07 06:01:41,082-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] 
>>>>>>>>>> (default
>>>>>>>>>> task-1) [105f7555-517b-4bf9-b86e-6eb42375de20] START, DestroyVDSComma
>>>>>>>>>> nd(HostName = lago-upgrade-from-release-suite-master-host-0,
>>>>>>>>>> DestroyVmVDSCommandParameters:{hostId='38e1379b-c3b6-4a2e-91df-d1f346e414a9',
>>>>>>>>>> vmId='dfbd75e2-a9cb-4fca-8788-a169
>>>>>>>>>> 54db4abf', secondsToWait='0', gracefully='false', reason='',
>>>>>>>>>> ignoreNoVm='false'}), log id: 78bba2f8
>>>>>>>>>> 2019-05-07 06:01:42,090-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] 
>>>>>>>>>> (default
>>>>>>>>>> task-1) [105f7555-517b-4bf9-b86e-6eb42375de20] FINISH, DestroyVDSComm
>>>>>>>>>> and, return: , log id: 78bba2f8
>>>>>>>>>> 2019-05-07 06:01:42,090-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.DestroyVmVDSCommand] (default 
>>>>>>>>>> task-1)
>>>>>>>>>> [105f7555-517b-4bf9-b86e-6eb42375de20] FINISH, DestroyVmVDSCommand, r
>>>>>>>>>> eturn: , log id: 24278e9b
>>>>>>>>>> 2019-05-07 06:01:42,094-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>>>>>>>>>> (ForkJoinPool-1-worker-4) [] VM 
>>>>>>>>>> 'dfbd75e2-a9cb-4fca-8788-a16954db4abf' was
>>>>>>>>>> reported
>>>>>>>>>> as Down on VDS
>>>>>>>>>> '38e1379b-c3b6-4a2e-91df-d1f346e414a9'(lago-upgrade-from-release-suite-master-host-0)
>>>>>>>>>> 2019-05-07 06:01:42,096-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>>>>>>>>>> (ForkJoinPool-1-worker-4) [] START, DestroyVDSCommand(HostName =
>>>>>>>>>> lago-upgrade-
>>>>>>>>>> from-release-suite-master-host-0,
>>>>>>>>>> DestroyVmVDSCommandParameters:{hostId='38e1379b-c3b6-4a2e-91df-d1f346e414a9',
>>>>>>>>>> vmId='dfbd75e2-a9cb-4fca-8788-a16954db4abf', secondsToWait='0
>>>>>>>>>> ', gracefully='false', reason='', ignoreNoVm='true'}), log id:
>>>>>>>>>> 1dbd31eb
>>>>>>>>>> 2019-05-07 06:01:42,114-04 INFO
>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>>>>>>>>>> (ForkJoinPool-1-worker-4) [] Failed to destroy VM
>>>>>>>>>> 'dfbd75e2-a9cb-4fca-8788-a16
>>>>>>>>>> 954db4abf' because VM does not exist, ignoring
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> _______________________________________________
>> Devel mailing list -- devel@ovirt.org
>> To unsubscribe send an email to devel-le...@ovirt.org
>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>> oVirt Code of Conduct:
>> https://www.ovirt.org/community/about/community-guidelines/
>> List Archives:
>> https://lists.ovirt.org/archives/list/devel@ovirt.org/message/OQ22IENFIVB2SHBJ2WOPO3KECPVZ4NPU/
>>
>
>
> --
>
> Sandro Bonazzola
>
> MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
>
> Red Hat EMEA <https://www.redhat.com/>
>
> sbona...@redhat.com
> <https://red.ht/sig>
> <https://redhat.com/summit>
>
_______________________________________________
Devel mailing list -- devel@ovirt.org
To unsubscribe send an email to devel-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/devel@ovirt.org/message/AIS2UOHJJBVXQOB3QWEXD4BMJ2I6BCWY/

Reply via email to