On Fri, May 10, 2019 at 12:04 PM Milan Zamazal <mzama...@redhat.com> wrote:

> Marcin Sobczyk <msobc...@redhat.com> writes:
>
> > Hi,
> >
> > the mentioned patch only touches our yum plugin - I don't think it's
> > related to the failure.
> >
> > VDSM fails on 'VM.create' call - Milan, could you please take a quick
> > look at it?
>
> I merged a patch to master yesterday that disables legacy pre-XML Engine
> (< 4.2) support.  Engines and clusters < 4.2 are no longer supported in
> master.
>
> I can see the VM is run with a legacy configuration rather than with
> `xml' parameter.  I wonder why -- is perhaps the cluster level < 4.2?
> If so then the cluster level must be updated before the VM is run.
>

This is on master branch. but it is happening on upgrade suite.
however, the suite is upgrading from 4.3 to 4.4:

pre-reposync-config.repo -> ../common/yum-repos/ovirt-4.3-cq.repo
reposync-config.repo -> ../common/yum-repos/ovirt-master-cq.repo



> > Regards, Marcin
> >
> > On 5/10/19 11:36 AM, Dafna Ron wrote:
> >> Hi,
> >>
> >> We are failing upgrade-from-release-suite.el7.x86_64 /
> >> 004_basic_sanity.vm_run
> >>
> >> The issue is an unexpected exception in vdsm.
> >>
> >> root cause based on CQ is this patch:
> >> https://gerrit.ovirt.org/#/c/99854/ - yum: Allow downloading only
> >> 'vdsm' package
> >>
> >> Logs can be found here:
> >>
> >>
> http://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/14079/artifact/upgrade-from-release-suite.el7.x86_64/test_logs/upgrade-from-release-suite-master/post-004_basic_sanity.py/
> >>
> >> Marcin, can you please take a look?
> >>
> >> error:
> >>
> >> vdsm:
> >>
> >> 2019-05-10 05:06:38,329-0400 ERROR (jsonrpc/0) [api] FINISH create
> >> error=Unexpected exception (api:131)
> >> Traceback (most recent call last):
> >>   File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line
> >> 124, in method
> >>     ret = func(*args, **kwargs)
> >>   File "/usr/lib/python2.7/site-packages/vdsm/API.py", line 245, in
> create
> >>     raise exception.UnexpectedError()
> >> UnexpectedError: Unexpected exception
> >> 2019-05-10 05:06:38,330-0400 INFO  (jsonrpc/0) [api.virt] FINISH
> >> create return={'status': {'message': 'Unexpected exception', 'code':
> >> 16}} from=::ffff:192.168.201.2,39486,
> >> flow_id=83e4f0c4-a39a-45aa-891f-4022765e1a87,
> >> vmId=397013a0-b7d4-4d38-86c3-e944cebd75a7 (api:54)
> >> 2019-05-10 05:06:38,331-0400 INFO  (jsonrpc/0)
> >> [jsonrpc.JsonRpcServer] RPC call VM.create failed (error 16) in 0.00
> >> seconds (__init__:312)
> >> 2019-05-10 05:06:38,629-0400 INFO  (jsonrpc/7) [vdsm.api] FINISH
> >> getStorageDomainInfo return={'info': {'uuid':
> >> '24498263-8985-46a9-9161-f65ee776cb7f', 'vgMetadataDevice':
> >> '360014051f333158820a4cc6ab3be5b55', 'vguuid':
> >> 'rXTLNc-JwEz-0dtL-Z2EC-ASqn-gzkg-jDJa1b', 'metadataDevice':
> >> '360014051f333158820a4cc6ab3be5b55', 'state': 'OK', 'version': '4',
> >> 'role': 'Master', 'type': 'ISCSI', 'class': 'Data', 'pool':
> >> ['3e330122-c587-4078-b9a4-13dbb697c5cc'], 'name': 'iscsi'}}
> >> from=::ffff:192.168.201.2,39486, flow_id=3f76d32a,
> >> task_id=084e7ee5-3c43-4905-833f-9a522043a862 (api:54)
> >>
> >> engine:
> >>
> >> 2019-05-10 05:06:38,328-04 ERROR
> >> [org.ovirt.engine.core.vdsbroker.CreateVDSCommand] (default task-1)
> >> [83e4f0c4-a39a-45aa-891f-4022765e1a87] VDS::create Failed creating
> >> vm 'v
> >> m0' in vds = '566eda0f-3ef0-4791-a618-1e649af4b0be' error =
> >> 'org.ovirt.engine.core.vdsbroker.vdsbroker.VDSErrorException:
> >> VDSGenericException: VDSErrorException: Failed to C
> >> reateBrokerVDS, error = Unexpected exception, code = 16'
> >> 2019-05-10 05:06:38,328-04 INFO
> >> [org.ovirt.engine.core.vdsbroker.CreateVDSCommand] (default task-1)
> >> [83e4f0c4-a39a-45aa-891f-4022765e1a87] FINISH, CreateVDSCommand,
> >> return:
> >>  Down, log id: 3cb7c62d
> >> 2019-05-10 05:06:38,328-04 WARN
> >> [org.ovirt.engine.core.bll.RunVmCommand] (default task-1)
> >> [83e4f0c4-a39a-45aa-891f-4022765e1a87] Failed to run VM 'vm0':
> >> EngineException: or
> >> g.ovirt.engine.core.vdsbroker.vdsbroker.VDSErrorException:
> >> VDSGenericException: VDSErrorException: Failed to CreateBrokerVDS,
> >> error = Unexpected exception, code = 16 (Failed
> >>  with error unexpected and code 16)
> >> 2019-05-10 05:06:38,328-04 INFO
> >> [org.ovirt.engine.core.bll.RunVmOnceCommand] (default task-1)
> >> [83e4f0c4-a39a-45aa-891f-4022765e1a87] Lock freed to object
> >> 'EngineLock:{exclu
> >> siveLocks='[397013a0-b7d4-4d38-86c3-e944cebd75a7=VM]', sharedLocks=''}'
> >> 2019-05-10 05:06:38,329-04 INFO
> >> [org.ovirt.engine.core.bll.RunVmCommand] (default task-1)
> >> [83e4f0c4-a39a-45aa-891f-4022765e1a87] Trying to rerun VM 'vm0'
> >> 2019-05-10 05:06:38,373-04 WARN
> >> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> >> (default task-1) [83e4f0c4-a39a-45aa-891f-4022765e1a87] EVENT_ID:
> >> USE
> >> R_INITIATED_RUN_VM_FAILED(151), Failed to run VM vm0 on Host
> >> lago-upgrade-from-release-suite-master-host-0.
> >> 2019-05-10 05:06:38,381-04 INFO
> >> [org.ovirt.engine.core.bll.RunVmOnceCommand] (default task-1)
> >> [83e4f0c4-a39a-45aa-891f-4022765e1a87] Lock Acquired to object
> >> 'EngineLock:{ex
> >> clusiveLocks='[397013a0-b7d4-4d38-86c3-e944cebd75a7=VM]',
> sharedLocks=''}'
> >> 2019-05-10 05:06:38,393-04 INFO
> >> [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand]
> >> (default task-1) [83e4f0c4-a39a-45aa-891f-4022765e1a87] START,
> >> IsVmDuringIn
> >> itiatingVDSCommand(
> >>
> IsVmDuringInitiatingVDSCommandParameters:{vmId='397013a0-b7d4-4d38-86c3-e944cebd75a7'}),
> >> log id: 96be328
> >> 2019-05-10 05:06:38,393-04 INFO
> >> [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand]
> >> (default task-1) [83e4f0c4-a39a-45aa-891f-4022765e1a87] FINISH,
> >> IsVmDuringI
> >> nitiatingVDSCommand, return: false, log id: 96be328
> >> 2019-05-10 05:06:38,403-04 WARN
> >> [org.ovirt.engine.core.bll.RunVmOnceCommand] (default task-1)
> >> [83e4f0c4-a39a-45aa-891f-4022765e1a87] Validation of action
> >> 'RunVmOnce' failed
> >>  for user admin@internal-authz. Reasons:
> >>
> VAR__ACTION__RUN,VAR__TYPE__VM,VAR__ACTION__RUN,VAR__TYPE__VM,SCHEDULING_NO_HOSTS
> >> 2019-05-10 05:06:38,404-04 INFO
> >> [org.ovirt.engine.core.bll.RunVmOnceCommand] (default task-1)
> >> [83e4f0c4-a39a-45aa-891f-4022765e1a87] Lock freed to object
> >> 'EngineLock:{exclu
> >> siveLocks='[397013a0-b7d4-4d38-86c3-e944cebd75a7=VM]', sharedLocks=''}'
> >> 2019-05-10 05:06:38,413-04 ERROR
> >> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> >> (default task-1) [83e4f0c4-a39a-45aa-891f-4022765e1a87] EVENT_ID:
> >> USE
> >> R_FAILED_RUN_VM(54), Failed to run VM vm0  (User: admin@internal-authz
> ).
> >> 2019-05-10 05:06:38,462-04 ERROR
> >> [org.ovirt.engine.api.restapi.resource.AbstractBackendResource]
> >> (default task-1) [] Operation Failed: [There are no hosts to
> >> use. Check that the cluster contains at least one host in Up state.]
> >> 2019-05-10 05:06:38,466-04 INFO
> >> [org.ovirt.engine.core.bll.ProcessDownVmCommand]
> >> (EE-ManagedThreadFactory-engine-Thread-35) [3f40b3e8] Running
> >> command: ProcessDownVmCommand internal: true.
> >>
> >>
> >> Thanks,
> >> Dafna
> >>
>
_______________________________________________
Devel mailing list -- devel@ovirt.org
To unsubscribe send an email to devel-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/devel@ovirt.org/message/3LAEUTPBREDZDT6YMY2FOY6PCFYY6W6K/

Reply via email to