On February 17, 2020 1:55:13 PM GMT+02:00, "Vrgotic, Marko" 
<m.vrgo...@activevideo.com> wrote:
>Good day Strahil,
>
>
>
>I believe I found the causing link:
>
>
>
>HostedEngine.log-20200216:-cpu
>SandyBridge,pcid=on,spec-ctrl=on,ssbd=on,md-clear=on,vme=on,hypervisor=on,arat=on,xsaveopt=on
>\
>
>HostedEngine.log-20200216:2020-02-13T17:58:38.674630Z qemu-kvm:
>warning: host doesn't support requested feature: CPUID.07H:EDX.md-clear
>[bit 10]
>
>HostedEngine.log-20200216:2020-02-13T17:58:38.676205Z qemu-kvm:
>warning: host doesn't support requested feature: CPUID.07H:EDX.md-clear
>[bit 10]
>
>HostedEngine.log-20200216:2020-02-13T17:58:38.676901Z qemu-kvm:
>warning: host doesn't support requested feature: CPUID.07H:EDX.md-clear
>[bit 10]
>
>HostedEngine.log-20200216:2020-02-13T17:58:38.677616Z qemu-kvm:
>warning: host doesn't support requested feature: CPUID.07H:EDX.md-clear
>[bit 10]
>
>
>
>The "md-clear" CPU seem to be removed as feature due to spectre
>vulnerabilities.
>
>
>
>However, when I check the CPU Type/flags of the VMs on the same Host as
>where Engine is currently, as well as on the other hosts, the md-clear
>seems to be only present on the HostedEngine:
>
>
>
>  *   HostedEngine:
>
>FromwebUI:
>Intel SandyBridge IBRS SSBD Family
>
>
>
>Via virsh:
>
>#virsh dumpxml
>
><cpu mode='custom' match='exact' check='full'>
>
>    <model fallback='forbid'>SandyBridge</model>
>
>    <topology sockets='16' cores='4' threads='1'/>
>
>    <feature policy='require' name='pcid'/>
>
>    <feature policy='require' name='spec-ctrl'/>
>
>    <feature policy='require' name='ssbd'/>
>
>    <feature policy='require' name='md-clear'/>
>
>    <feature policy='require' name='vme'/>
>
>    <feature policy='require' name='hypervisor'/>
>
>    <feature policy='require' name='arat'/>
>
>    <feature policy='require' name='xsaveopt'/>
>
>    <numa>
>
>      <cell id='0' cpus='0-3' memory='16777216' unit='KiB'/>
>
>    </numa>
>
></cpu>
>
>
>
>
>
>  *   OtherVMs:
>
>From webUI:
>(SandyBridge,+pcid,+spec-ctrl,+ssbd)
>
>
>
>Via virsh:
>
>#virsh dumpxml
>
><cpu mode='custom' match='exact' check='full'>
>
>    <model fallback='forbid'>SandyBridge</model>
>
>    <topology sockets='16' cores='1' threads='1'/>
>
>    <feature policy='require' name='pcid'/>
>
>    <feature policy='require' name='spec-ctrl'/>
>
>    <feature policy='require' name='ssbd'/>
>
>    <feature policy='require' name='vme'/>
>
>    <feature policy='require' name='hypervisor'/>
>
>    <feature policy='require' name='arat'/>
>
>    <feature policy='require' name='xsaveopt'/>
>
>    <numa>
>
>      <cell id='0' cpus='0-3' memory='4194304' unit='KiB'/>
>
>    </numa>
>
>  </cpu>
>
>
>
>
>
>Strahil, knowing this, do you propose different approach or shall I
>just proceed with initially suggested workaround?
>
>
>
>Kindly awaiting your eply.
>
>
>
>-----
>kind regards/met vriendelijke groeten
>
>Marko Vrgotic
>Sr. System Engineer @ System Administration
>
>ActiveVideo
>o: +31 (35) 6774131
>e: m.vrgo...@activevideo.com<mailto:m.vrgo...@activevideo.com>
>w: www.activevideo.com<http://www.activevideo.com>
>
>ActiveVideo Networks BV. Mediacentrum 3745 Joop van den Endeplein
>1.1217 WJ Hilversum, The Netherlands. The information contained in this
>message may be legally privileged and confidential. It is intended to
>be read only by the individual or entity to whom it is addressed or by
>their designee. If the reader of this message is not the intended
>recipient, you are on notice that any distribution of this message, in
>any form, is strictly prohibited.  If you have received this message in
>error, please immediately notify the sender and/or ActiveVideo
>Networks, LLC by telephone at +1 408.931.9200 and delete or destroy any
>copy of this message.
>
>
>
>
>
>
>
>
>
>
>On 16/02/2020, 15:28, "Strahil Nikolov" <hunter86...@yahoo.com> wrote:
>
>
>
>    ssh root@engine "poweroff"
>
>ssh host-that-holded-engine "virsh undefine HostedEngine; virsh list
>--all"
>
>
>
>    Lot's of virsh - less vdsm :)
>
>
>
>    Good luck
>
>
>
>    Best Regards,
>
>    Strahil Nikolov
>
>
>
>
>
>В неделя, 16 февруари 2020 г., 16:01:44 ч. Гринуич+2, Vrgotic, Marko
><m.vrgo...@activevideo.com> написа:
>
>
>
>
>
>
>
>
>
>
>
>    Hi Strahil,
>
>
>
>   Regarding step 3:  Stop and undefine the VM on the last working host
>
>One question: How do I undefine HostedEngine from last Host?
>Hosted-engine command does not provide such option, or it's just not
>obvious.
>
>
>
>    Kindly awaiting your reply.
>
>
>
>
>
>    -----
>
>    kind regards/met vriendelijke groeten
>
>
>
>    Marko Vrgotic
>
>    ActiveVideo
>
>
>
>
>
> On 14/02/2020, 18:44, "Strahil Nikolov" <hunter86...@yahoo.com> wrote:
>
>
>
>On February 14, 2020 4:19:53 PM GMT+02:00, "Vrgotic, Marko"
><m.vrgo...@activevideo.com> wrote:
>
>        >Good answer Strahil,
>
>        >
>
>        >Thank you, I forgot.
>
>        >
>
>        >Libvirt logs are actually showing the reason why:
>
>        >
>
>        >2020-02-14T12:33:51.847970Z qemu-kvm: -drive
>
>>file=/var/run/vdsm/storage/054c43fc-1924-4106-9f80-0f2ac62b9886/b019c5fa-8fb5-4bfc-8339-f5b7f590a051/f1ce8ba6-2d3b-4309-bca0-e6a00ce74c75,format=raw,if=none,id=drive-ua-b019c5fa-8fb5-4bfc-8339-f5b7f590a051,serial=b019c5fa-8fb5-4bfc-8339-f5b7f590a051,werror=stop,rerror=stop,cache=none,aio=threads:
>
>        >'serial' is deprecated, please use the corresponding option of
>
>        >'-device' instead
>
>     >Spice-Message: 04:33:51.856: setting TLS option 'CipherString' to
>
>        >'kECDHE+FIPS:kDHE+FIPS:kRSA+FIPS:!eNULL:!aNULL' from
>
>        >/etc/pki/tls/spice.cnf configuration file
>
>  >2020-02-14T12:33:51.863449Z qemu-kvm: warning: CPU(s) not present in
>
> >any NUMA nodes: CPU 4 [socket-id: 1, core-id: 0, thread-id: 0], CPU 5
>
>>[socket-id: 1, core-id: 1, thread-id: 0], CPU 6 [socket-id: 1,
>core-id:
>
>>2, thread-id: 0], CPU 7 [socket-id: 1, core-id: 3, thread-id: 0], CPU
>8
>
>>[socket-id: 2, core-id: 0, thread-id: 0], CPU 9 [socket-id: 2,
>core-id:
>
>>1, thread-id: 0], CPU 10 [socket-id: 2, core-id: 2, thread-id: 0], CPU
>
>    >11 [socket-id: 2, core-id: 3, thread-id: 0], CPU 12 [socket-id: 3,
>
>>core-id: 0, thread-id: 0], CPU 13 [socket-id: 3, core-id: 1,
>thread-id:
>
>>0], CPU 14 [socket-id: 3, core-id: 2, thread-id: 0], CPU 15
>[socket-id:
>
>       >3, core-id: 3, thread-id: 0], CPU 16 [socket-id: 4, core-id: 0,
>
>>thread-id: 0], CPU 17 [socket-id: 4, core-id: 1, thread-id: 0], CPU 18
>
>       >[socket-id: 4, core-id: 2, thread-id: 0], CPU 19 [socket-id: 4,
>
>>core-id: 3, thread-id: 0], CPU 20 [socket-id: 5, core-id: 0,
>thread-id:
>
>>0], CPU 21 [socket-id: 5, core-id: 1, thread-id: 0], CPU 22
>[socket-id:
>
>       >5, core-id: 2, thread-id: 0], CPU 23 [socket-id: 5, core-id: 3,
>
>>thread-id: 0], CPU 24 [socket-id: 6, core-id: 0, thread-id: 0], CPU 25
>
>       >[socket-id: 6, core-id: 1, thread-id: 0], CPU 26 [socket-id: 6,
>
>>core-id: 2, thread-id: 0], CPU 27 [socket-id: 6, core-id: 3,
>thread-id:
>
>>0], CPU 28 [socket-id: 7, core-id: 0, thread-id: 0], CPU 29
>[socket-id:
>
>       >7, core-id: 1, thread-id: 0], CPU 30 [socket-id: 7, core-id: 2,
>
>>thread-id: 0], CPU 31 [socket-id: 7, core-id: 3, thread-id: 0], CPU 32
>
>       >[socket-id: 8, core-id: 0, thread-id: 0], CPU 33 [socket-id: 8,
>
>>core-id: 1, thread-id: 0], CPU 34 [socket-id: 8, core-id: 2,
>thread-id:
>
>>0], CPU 35 [socket-id: 8, core-id: 3, thread-id: 0], CPU 36
>[socket-id:
>
>       >9, core-id: 0, thread-id: 0], CPU 37 [socket-id: 9, core-id: 1,
>
>>thread-id: 0], CPU 38 [socket-id: 9, core-id: 2, thread-id: 0], CPU 39
>
>      >[socket-id: 9, core-id: 3, thread-id: 0], CPU 40 [socket-id: 10,
>
>        >core-id: 0, thread-id: 0], CPU 41 [socket-id: 10, core-id: 1,
>
>>thread-id: 0], CPU 42 [socket-id: 10, core-id: 2, thread-id: 0], CPU
>43
>
>     >[socket-id: 10, core-id: 3, thread-id: 0], CPU 44 [socket-id: 11,
>
>        >core-id: 0, thread-id: 0], CPU 45 [socket-id: 11, core-id: 1,
>
>>thread-id: 0], CPU 46 [socket-id: 11, core-id: 2, thread-id: 0], CPU
>47
>
>     >[socket-id: 11, core-id: 3, thread-id: 0], CPU 48 [socket-id: 12,
>
>        >core-id: 0, thread-id: 0], CPU 49 [socket-id: 12, core-id: 1,
>
>>thread-id: 0], CPU 50 [socket-id: 12, core-id: 2, thread-id: 0], CPU
>51
>
>     >[socket-id: 12, core-id: 3, thread-id: 0], CPU 52 [socket-id: 13,
>
>        >core-id: 0, thread-id: 0], CPU 53 [socket-id: 13, core-id: 1,
>
>>thread-id: 0], CPU 54 [socket-id: 13, core-id: 2, thread-id: 0], CPU
>55
>
>     >[socket-id: 13, core-id: 3, thread-id: 0], CPU 56 [socket-id: 14,
>
>        >core-id: 0, thread-id: 0], CPU 57 [socket-id: 14, core-id: 1,
>
>>thread-id: 0], CPU 58 [socket-id: 14, core-id: 2, thread-id: 0], CPU
>59
>
>     >[socket-id: 14, core-id: 3, thread-id: 0], CPU 60 [socket-id: 15,
>
>        >core-id: 0, thread-id: 0], CPU 61 [socket-id: 15, core-id: 1,
>
>>thread-id: 0], CPU 62 [socket-id: 15, core-id: 2, thread-id: 0], CPU
>63
>
>        >[socket-id: 15, core-id: 3, thread-id: 0]
>
>>2020-02-14T12:33:51.863475Z qemu-kvm: warning: All CPU(s) up to
>maxcpus
>
>  >should be described in NUMA config, ability to start up with partial
>
>        >NUMA mappings is obsoleted and will be removed in future
>
>   >2020-02-14T12:33:51.863973Z qemu-kvm: warning: host doesn't support
>
>        >requested feature: CPUID.07H:EDX.md-clear [bit 10]
>
>   >2020-02-14T12:33:51.865066Z qemu-kvm: warning: host doesn't support
>
>        >requested feature: CPUID.07H:EDX.md-clear [bit 10]
>
>   >2020-02-14T12:33:51.865547Z qemu-kvm: warning: host doesn't support
>
>        >requested feature: CPUID.07H:EDX.md-clear [bit 10]
>
>   >2020-02-14T12:33:51.865996Z qemu-kvm: warning: host doesn't support
>
>        >requested feature: CPUID.07H:EDX.md-clear [bit 10]
>
>        >2020-02-14 12:33:51.932+0000: shutting down, reason=failed
>
>        >
>
>        >But then I wonder if the following is related to error above:
>
>        >
>
>    >Before I started upgrading Host by Host, all Hosts in Cluster were
>
>    >showing CPU Family type: " Intel SandyBridge IBRS SSBD MDS Family"
>
>  >After first Host was upgraded, his CPU Family type was changed to: "
>
>     >Intel SandyBridge IBRS SSBD Family" and that forced me to have do
>
>      >"downgrade" Cluster family type to " Intel SandyBridge IBRS SSBD
>
> >Family" in order to be able to Activate the Host back inside Cluster.
>
>>Following further, each Host CPU family type changed after Upgrade
>from
>
>>"" Intel SandyBridge IBRS SSBD MDS Family" to "" Intel SandyBridge
>IBRS
>
>        >SSBD Family" , except one where HostedEngine is currently one.
>
>        >
>
>>Could this possibly be the reason why I cannot Migrate the
>HostedEngine
>
>        >now and how to solve it?
>
>        >
>
>        >Kindly awaiting your reply.
>
>        >
>
>        >
>
>        >-----
>
>        >kind regards/met vriendelijke groeten
>
>        >
>
>        >Marko Vrgotic
>
>        >Sr. System Engineer @ System Administration
>
>        >
>
>        >ActiveVideo
>
>        >o: +31 (35) 6774131
>
>        >e: m.vrgo...@activevideo.com
>
>        >w: www.activevideo.com <http://www.activevideo.com>
>
>        >
>
>     >ActiveVideo Networks BV. Mediacentrum 3745 Joop van den Endeplein
>
>>1.1217 WJ Hilversum, The Netherlands. The information contained in
>this
>
> >message may be legally privileged and confidential. It is intended to
>
>>be read only by the individual or entity to whom it is addressed or by
>
>     >their designee. If the reader of this message is not the intended
>
>>recipient, you are on notice that any distribution of this message, in
>
>>any form, is strictly prohibited.  If you have received this message
>in
>
>        >error, please immediately notify the sender and/or ActiveVideo
>
>>Networks, LLC by telephone at +1 408.931.9200 and delete or destroy
>any
>
>        >copy of this message.
>
>        >
>
>        >
>
>        >
>
>        >
>
>        >
>
>        >
>
>        >
>
>>On 14/02/2020, 14:01, "Strahil Nikolov" <hunter86...@yahoo.com> wrote:
>
>        >
>
>        >On February 14, 2020 2:47:04 PM GMT+02:00, "Vrgotic, Marko"
>
>        ><m.vrgo...@activevideo.com> wrote:
>
>        >    >Dear oVirt,
>
>        >    >
>
>> >I have problem migrating HostedEngine, only HA VM server, to other
>HA
>
>        >    >nodes.
>
>        >    >
>
>        >    >Bit of background story:
>
>        >    >
>
>        >    >  *  We have oVirt SHE 4.3.5
>
>        >    >  *  Three Nodes act as HA pool for SHE
>
>        >    >  *  Node 3 is currently Hosting SHE
>
>        >    >  *  Actions:
>
>>>*  Put Node1 in Maintenance mode, all VMs were successfully migrated,
>
>        >    >than Upgrade packages, Activate Host – all looks good
>
>>>*  Put Node2 in Maintenance mode, all VMs were successfully migrated,
>
>        >    >than Upgrade packages, Activate Host – all looks good
>
>        >    >
>
>        >    >Not the problem:
>
> >    >Try to set  Node3 in Maintenance mode, all VMs were successfully
>
>        >    >migrated, except HostedEngine.
>
>        >    >
>
>  >    >When attempting Migration of the VM HostedEngine, it fails with
>
>        >    >following error message:
>
>        >    >
>
>        >    >2020-02-14 12:33:49,960Z INFO
>
>  >    >[org.ovirt.engine.core.bll.MigrateVmCommand] (default task-265)
>
>   >    >[16f4559e-e262-4c9d-80b4-ec81c2cbf950] Lock Acquired to object
>
>>>'EngineLock:{exclusiveLocks='[66b6d489-ceb8-486a-951a-355e21f13627=VM]',
>
>        >    >sharedLocks=''}'
>
>        >    >2020-02-14 12:33:49,984Z INFO
>
>>    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager] (default
>
>  >    >task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] Candidate host
>
>>  >'ovirt-sj-04.ictv.com' ('d98843da-bd81-46c9-9425-065b196ac59d') was
>
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>
>        >    >id: null)
>
>        >    >2020-02-14 12:33:49,984Z INFO
>
>>    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager] (default
>
>  >    >task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] Candidate host
>
>>  >'ovirt-sj-05.ictv.com' ('e3176705-9fb0-41d6-8721-367dfa2e62bd') was
>
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>
>        >    >id: null)
>
>        >    >2020-02-14 12:33:49,997Z INFO
>
>  >    >[org.ovirt.engine.core.bll.MigrateVmCommand] (default task-265)
>
>        >    >[16f4559e-e262-4c9d-80b4-ec81c2cbf950] Running command:
>
>       >    >MigrateVmCommand internal: false. Entities affected :  ID:
>
>>  >66b6d489-ceb8-486a-951a-355e21f13627 Type: VMAction group
>MIGRATE_VM
>
>        >    >with role type USER
>
>        >    >2020-02-14 12:33:50,008Z INFO
>
>>    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager] (default
>
>  >    >task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] Candidate host
>
>>  >'ovirt-sj-04.ictv.com' ('d98843da-bd81-46c9-9425-065b196ac59d') was
>
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>
>        >    >id: 16f4559e-e262-4c9d-80b4-ec81c2cbf950)
>
>        >    >2020-02-14 12:33:50,008Z INFO
>
>>    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager] (default
>
>  >    >task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] Candidate host
>
>>  >'ovirt-sj-05.ictv.com' ('e3176705-9fb0-41d6-8721-367dfa2e62bd') was
>
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>
>        >    >id: 16f4559e-e262-4c9d-80b4-ec81c2cbf950)
>
>        >    >2020-02-14 12:33:50,033Z INFO
>
>>>[org.ovirt.engine.core.vdsbroker.MigrateVDSCommand] (default
>task-265)
>
> >    >[16f4559e-e262-4c9d-80b4-ec81c2cbf950] START, MigrateVDSCommand(
>
>>>MigrateVDSCommandParameters:{hostId='f8d27efb-1527-45f0-97d6-d34a86abaaa2',
>
>        >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627',
>
>        >    >srcHost='ovirt-sj-03.ictv.com',
>
>        >    >dstVdsId='9808f434-5cd4-48b5-8bbc-e639e391c6a5',
>
>  >    >dstHost='ovirt-sj-01.ictv.com:54321', migrationMethod='ONLINE',
>
>>  >tunnelMigration='false', migrationDowntime='0',
>autoConverge='true',
>
>>  >migrateCompressed='false', consoleAddress='null',
>maxBandwidth='40',
>
>        >    >enableGuestEvents='true', maxIncomingMigrations='2',
>
>        >    >maxOutgoingMigrations='2',
>
>   >    >convergenceSchedule='[init=[{name=setDowntime, params=[100]}],
>
>        >>stalling=[{limit=1, action={name=setDowntime, params=[150]}},
>
>        >{limit=2,
>
>        >    >action={name=setDowntime, params=[200]}}, {limit=3,
>
>        >    >action={name=setDowntime, params=[300]}}, {limit=4,
>
>        >    >action={name=setDowntime, params=[400]}}, {limit=6,
>
>        >    >action={name=setDowntime, params=[500]}}, {limit=-1,
>
>> >action={name=abort, params=[]}}]]', dstQemu='10.210.13.11'}), log
>id:
>
>        >    >5c126a47
>
>        >    >2020-02-14 12:33:50,036Z INFO
>
>>  >[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
>
> >    >(default task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] START,
>
>        >    >MigrateBrokerVDSCommand(HostName = ovirt-sj-03.ictv.com,
>
>>>MigrateVDSCommandParameters:{hostId='f8d27efb-1527-45f0-97d6-d34a86abaaa2',
>
>        >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627',
>
>        >    >srcHost='ovirt-sj-03.ictv.com',
>
>        >    >dstVdsId='9808f434-5cd4-48b5-8bbc-e639e391c6a5',
>
>  >    >dstHost='ovirt-sj-01.ictv.com:54321', migrationMethod='ONLINE',
>
>>  >tunnelMigration='false', migrationDowntime='0',
>autoConverge='true',
>
>>  >migrateCompressed='false', consoleAddress='null',
>maxBandwidth='40',
>
>        >    >enableGuestEvents='true', maxIncomingMigrations='2',
>
>        >    >maxOutgoingMigrations='2',
>
>   >    >convergenceSchedule='[init=[{name=setDowntime, params=[100]}],
>
>        >>stalling=[{limit=1, action={name=setDowntime, params=[150]}},
>
>        >{limit=2,
>
>        >    >action={name=setDowntime, params=[200]}}, {limit=3,
>
>        >    >action={name=setDowntime, params=[300]}}, {limit=4,
>
>        >    >action={name=setDowntime, params=[400]}}, {limit=6,
>
>        >    >action={name=setDowntime, params=[500]}}, {limit=-1,
>
>> >action={name=abort, params=[]}}]]', dstQemu='10.210.13.11'}), log
>id:
>
>        >    >a0f776d
>
>        >    >2020-02-14 12:33:50,043Z INFO
>
>>  >[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
>
>>    >(default task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] FINISH,
>
>        >    >MigrateBrokerVDSCommand, return: , log id: a0f776d
>
>        >    >2020-02-14 12:33:50,046Z INFO
>
>>>[org.ovirt.engine.core.vdsbroker.MigrateVDSCommand] (default
>task-265)
>
>>    >[16f4559e-e262-4c9d-80b4-ec81c2cbf950] FINISH, MigrateVDSCommand,
>
>        >    >return: MigratingFrom, log id: 5c126a47
>
>        >    >2020-02-14 12:33:50,052Z INFO
>
>>>[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>
>>  >(default task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] EVENT_ID:
>
>>  >VM_MIGRATION_START(62), Migration started (VM: HostedEngine,
>Source:
>
>   >    >ovirt-sj-03.ictv.com, Destination: ovirt-sj-01.ictv.com, User:
>
>        >    >mvrgo...@ictv.com@ictv.com-authz).
>
>        >    >2020-02-14 12:33:52,893Z INFO
>
>        >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>>>(ForkJoinPool-1-worker-8) [] VM
>'66b6d489-ceb8-486a-951a-355e21f13627'
>
>        >    >was reported as Down on VDS
>
>     >    >'9808f434-5cd4-48b5-8bbc-e639e391c6a5'(ovirt-sj-01.ictv.com)
>
>        >    >2020-02-14 12:33:52,893Z INFO
>
>    >    >[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>
> >    >(ForkJoinPool-1-worker-8) [] START, DestroyVDSCommand(HostName =
>
>        >    >ovirt-sj-01.ictv.com,
>
>>>DestroyVmVDSCommandParameters:{hostId='9808f434-5cd4-48b5-8bbc-e639e391c6a5',
>
>  >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627', secondsToWait='0',
>
>>  >gracefully='false', reason='', ignoreNoVm='true'}), log id:
>7532a8c0
>
>        >    >2020-02-14 12:33:53,217Z INFO
>
>    >    >[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>
>        >    >(ForkJoinPool-1-worker-8) [] Failed to destroy VM
>
>>    >'66b6d489-ceb8-486a-951a-355e21f13627' because VM does not exist,
>
>        >    >ignoring
>
>        >    >2020-02-14 12:33:53,217Z INFO
>
>    >    >[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>
>> >(ForkJoinPool-1-worker-8) [] FINISH, DestroyVDSCommand, return: ,
>log
>
>        >    >id: 7532a8c0
>
>        >    >2020-02-14 12:33:53,217Z INFO
>
>        >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>        >    >(ForkJoinPool-1-worker-8) [] VM
>
>> >'66b6d489-ceb8-486a-951a-355e21f13627'(HostedEngine) was
>unexpectedly
>
>        >    >detected as 'Down' on VDS
>
>>>'9808f434-5cd4-48b5-8bbc-e639e391c6a5'(ovirt-sj-01.ictv.com)
>(expected
>
>        >    >on 'f8d27efb-1527-45f0-97d6-d34a86abaaa2')
>
>        >    >2020-02-14 12:33:53,217Z ERROR
>
>        >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>>  >(ForkJoinPool-1-worker-8) [] Migration of VM 'HostedEngine' to host
>
>  >    >'ovirt-sj-01.ictv.com' failed: VM destroyed during the startup.
>
>        >    >2020-02-14 12:33:53,219Z INFO
>
>        >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>        >    >(ForkJoinPool-1-worker-15) [] VM
>
>  >    >'66b6d489-ceb8-486a-951a-355e21f13627'(HostedEngine) moved from
>
>        >    >'MigratingFrom' --> 'Up'
>
>        >    >2020-02-14 12:33:53,219Z INFO
>
>        >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>        >    >(ForkJoinPool-1-worker-15) [] Adding VM
>
>>  >'66b6d489-ceb8-486a-951a-355e21f13627'(HostedEngine) to re-run list
>
>        >    >2020-02-14 12:33:53,221Z ERROR
>
>       >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmsMonitoring]
>
>        >    >(ForkJoinPool-1-worker-15) [] Rerun VM
>
>        >    >'66b6d489-ceb8-486a-951a-355e21f13627'. Called from VDS
>
>        >    >'ovirt-sj-03.ictv.com'
>
>        >    >2020-02-14 12:33:53,259Z INFO
>
>>  >[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand]
>
>        >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] START,
>
>        >    >MigrateStatusVDSCommand(HostName = ovirt-sj-03.ictv.com,
>
>>>MigrateStatusVDSCommandParameters:{hostId='f8d27efb-1527-45f0-97d6-d34a86abaaa2',
>
>  >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627'}), log id: 62bac076
>
>        >    >2020-02-14 12:33:53,265Z INFO
>
>>  >[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand]
>
>        >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] FINISH,
>
>        >    >MigrateStatusVDSCommand, return: , log id: 62bac076
>
>        >    >2020-02-14 12:33:53,277Z WARN
>
>>>[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>
>      >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] EVENT_ID:
>
>>  >VM_MIGRATION_TRYING_RERUN(128), Failed to migrate VM HostedEngine
>to
>
>   >    >Host ovirt-sj-01.ictv.com . Trying to migrate to another Host.
>
>        >    >2020-02-14 12:33:53,330Z INFO
>
>        >    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
>
> >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] Candidate host
>
>>  >'ovirt-sj-04.ictv.com' ('d98843da-bd81-46c9-9425-065b196ac59d') was
>
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>
>        >    >id: null)
>
>        >    >2020-02-14 12:33:53,330Z INFO
>
>        >    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
>
> >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] Candidate host
>
>>  >'ovirt-sj-05.ictv.com' ('e3176705-9fb0-41d6-8721-367dfa2e62bd') was
>
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>
>        >    >id: null)
>
>        >    >2020-02-14 12:33:53,345Z INFO
>
>        >    >[org.ovirt.engine.core.bll.MigrateVmCommand]
>
>>    >(EE-ManagedThreadFactory-engine-Thread-377323) [] Running
>command:
>
>       >    >MigrateVmCommand internal: false. Entities affected :  ID:
>
>>  >66b6d489-ceb8-486a-951a-355e21f13627 Type: VMAction group
>MIGRATE_VM
>
>        >    >with role type USER
>
>        >    >2020-02-14 12:33:53,356Z INFO
>
>        >    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
>
> >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] Candidate host
>
>>  >'ovirt-sj-04.ictv.com' ('d98843da-bd81-46c9-9425-065b196ac59d') was
>
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>
>        >    >id: 16f4559e-e262-4c9d-80b4-ec81c2cbf950)
>
>        >    >2020-02-14 12:33:53,356Z INFO
>
>        >    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
>
> >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] Candidate host
>
>>  >'ovirt-sj-05.ictv.com' ('e3176705-9fb0-41d6-8721-367dfa2e62bd') was
>
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>
>        >    >id: 16f4559e-e262-4c9d-80b4-ec81c2cbf950)
>
>        >    >2020-02-14 12:33:53,380Z INFO
>
>        >    >[org.ovirt.engine.core.vdsbroker.MigrateVDSCommand]
>
>        >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] START,
>
>        >    >MigrateVDSCommand(
>
>>>MigrateVDSCommandParameters:{hostId='f8d27efb-1527-45f0-97d6-d34a86abaaa2',
>
>        >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627',
>
>        >    >srcHost='ovirt-sj-03.ictv.com',
>
>        >    >dstVdsId='33e8ff78-e396-4f40-b43c-685bfaaee9af',
>
>  >    >dstHost='ovirt-sj-02.ictv.com:54321', migrationMethod='ONLINE',
>
>>  >tunnelMigration='false', migrationDowntime='0',
>autoConverge='true',
>
>>  >migrateCompressed='false', consoleAddress='null',
>maxBandwidth='40',
>
>        >    >enableGuestEvents='true', maxIncomingMigrations='2',
>
>        >    >maxOutgoingMigrations='2',
>
>   >    >convergenceSchedule='[init=[{name=setDowntime, params=[100]}],
>
>        >>stalling=[{limit=1, action={name=setDowntime, params=[150]}},
>
>        >{limit=2,
>
>        >    >action={name=setDowntime, params=[200]}}, {limit=3,
>
>        >    >action={name=setDowntime, params=[300]}}, {limit=4,
>
>        >    >action={name=setDowntime, params=[400]}}, {limit=6,
>
>        >    >action={name=setDowntime, params=[500]}}, {limit=-1,
>
>> >action={name=abort, params=[]}}]]', dstQemu='10.210.13.12'}), log
>id:
>
>        >    >d99059f
>
>        >    >2020-02-14 12:33:53,380Z INFO
>
>>  >[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
>
>        >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] START,
>
>        >    >MigrateBrokerVDSCommand(HostName = ovirt-sj-03.ictv.com,
>
>>>MigrateVDSCommandParameters:{hostId='f8d27efb-1527-45f0-97d6-d34a86abaaa2',
>
>        >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627',
>
>        >    >srcHost='ovirt-sj-03.ictv.com',
>
>        >    >dstVdsId='33e8ff78-e396-4f40-b43c-685bfaaee9af',
>
>  >    >dstHost='ovirt-sj-02.ictv.com:54321', migrationMethod='ONLINE',
>
>>  >tunnelMigration='false', migrationDowntime='0',
>autoConverge='true',
>
>>  >migrateCompressed='false', consoleAddress='null',
>maxBandwidth='40',
>
>        >    >enableGuestEvents='true', maxIncomingMigrations='2',
>
>        >    >maxOutgoingMigrations='2',
>
>   >    >convergenceSchedule='[init=[{name=setDowntime, params=[100]}],
>
>        >>stalling=[{limit=1, action={name=setDowntime, params=[150]}},
>
>        >{limit=2,
>
>        >    >action={name=setDowntime, params=[200]}}, {limit=3,
>
>        >    >action={name=setDowntime, params=[300]}}, {limit=4,
>
>        >    >action={name=setDowntime, params=[400]}}, {limit=6,
>
>        >    >action={name=setDowntime, params=[500]}}, {limit=-1,
>
>> >action={name=abort, params=[]}}]]', dstQemu='10.210.13.12'}), log
>id:
>
>        >    >6f0483ac
>
>        >    >2020-02-14 12:33:53,386Z INFO
>
>>  >[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
>
>        >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] FINISH,
>
>        >    >MigrateBrokerVDSCommand, return: , log id: 6f0483ac
>
>        >    >2020-02-14 12:33:53,388Z INFO
>
>        >    >[org.ovirt.engine.core.vdsbroker.MigrateVDSCommand]
>
>        >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] FINISH,
>
>        >    >MigrateVDSCommand, return: MigratingFrom, log id: d99059f
>
>        >    >2020-02-14 12:33:53,391Z INFO
>
>>>[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>
>      >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] EVENT_ID:
>
>>  >VM_MIGRATION_START(62), Migration started (VM: HostedEngine,
>Source:
>
>   >    >ovirt-sj-03.ictv.com, Destination: ovirt-sj-02.ictv.com, User:
>
>        >    >mvrgo...@ictv.com@ictv.com-authz).
>
>        >    >2020-02-14 12:33:55,108Z INFO
>
>>    >[org.ovirt.engine.core.vdsbroker.monitoring.VmsStatisticsFetcher]
>
>> >(EE-ManagedThreadFactory-engineScheduled-Thread-96) [] Fetched 10
>VMs
>
>        >    >from VDS '33e8ff78-e396-4f40-b43c-685bfaaee9af'
>
>        >    >2020-02-14 12:33:55,110Z INFO
>
>        >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>        >    >(EE-ManagedThreadFactory-engineScheduled-Thread-96) [] VM
>
>       >    >'66b6d489-ceb8-486a-951a-355e21f13627' is migrating to VDS
>
>> >'33e8ff78-e396-4f40-b43c-685bfaaee9af'(ovirt-sj-02.ictv.com)
>ignoring
>
>        >    >it in the refresh until migration is done
>
>        >    >2020-02-14 12:33:57,224Z INFO
>
>        >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>        >>(ForkJoinPool-1-worker-15) [] VM
>
>        >'66b6d489-ceb8-486a-951a-355e21f13627'
>
>        >    >was reported as Down on VDS
>
>     >    >'33e8ff78-e396-4f40-b43c-685bfaaee9af'(ovirt-sj-02.ictv.com)
>
>        >    >2020-02-14 12:33:57,225Z INFO
>
>    >    >[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>
>>    >(ForkJoinPool-1-worker-15) [] START, DestroyVDSCommand(HostName =
>
>        >    >ovirt-sj-02.ictv.com,
>
>>>DestroyVmVDSCommandParameters:{hostId='33e8ff78-e396-4f40-b43c-685bfaaee9af',
>
>  >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627', secondsToWait='0',
>
>>  >gracefully='false', reason='', ignoreNoVm='true'}), log id:
>1dec553e
>
>        >    >2020-02-14 12:33:57,672Z INFO
>
>    >    >[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>
>        >    >(ForkJoinPool-1-worker-15) [] Failed to destroy VM
>
>>    >'66b6d489-ceb8-486a-951a-355e21f13627' because VM does not exist,
>
>        >    >ignoring
>
>        >    >2020-02-14 12:33:57,672Z INFO
>
>    >    >[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>
>>>(ForkJoinPool-1-worker-15) [] FINISH, DestroyVDSCommand, return: ,
>log
>
>        >    >id: 1dec553e
>
>        >    >2020-02-14 12:33:57,672Z INFO
>
>        >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>        >    >(ForkJoinPool-1-worker-15) [] VM
>
>> >'66b6d489-ceb8-486a-951a-355e21f13627'(HostedEngine) was
>unexpectedly
>
>        >    >detected as 'Down' on VDS
>
>>>'33e8ff78-e396-4f40-b43c-685bfaaee9af'(ovirt-sj-02.ictv.com)
>(expected
>
>        >    >on 'f8d27efb-1527-45f0-97d6-d34a86abaaa2')
>
>        >    >2020-02-14 12:33:57,672Z ERROR
>
>        >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>>  >(ForkJoinPool-1-worker-15) [] Migration of VM 'HostedEngine' to
>host
>
>  >    >'ovirt-sj-02.ictv.com' failed: VM destroyed during the startup.
>
>        >    >2020-02-14 12:33:57,674Z INFO
>
>        >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>        >    >(ForkJoinPool-1-worker-8) [] VM
>
>  >    >'66b6d489-ceb8-486a-951a-355e21f13627'(HostedEngine) moved from
>
>        >    >'MigratingFrom' --> 'Up'
>
>        >    >2020-02-14 12:33:57,674Z INFO
>
>        >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>        >    >(ForkJoinPool-1-worker-8) [] Adding VM
>
>>  >'66b6d489-ceb8-486a-951a-355e21f13627'(HostedEngine) to re-run list
>
>        >    >2020-02-14 12:33:57,676Z ERROR
>
>       >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmsMonitoring]
>
>        >    >(ForkJoinPool-1-worker-8) [] Rerun VM
>
>        >    >'66b6d489-ceb8-48
>
>
>
>        I am afraid that your suspicions  are  right.
>
>
>
>        What is  the host cpu and the HostedEngine's xml?
>
>
>
>Have  you checked the xml on any working VM ? What cpu flags  do the
>working VMs have ?
>
>
>
>  How to solve - I think I have a solution , but you might not like it.
>
>
>
>        1. Get current VM xml with virsh
>
>2. Set all nodes in maintenance 'hosted-engine --set-maintenance 
>--mode=global'
>
>        3. Stop and undefine the VM on the last working host
>
>4. Edit the xml from step 1 and add/remove the flags  that are
>different from the other (working) VMs
>
>        5. Define the HostedEngine on any of the updated hosts
>
>        6. Start the HostedEngine via  virsh.
>
>        7. Try with different cpu flags until the engine starts.
>
>8. Leave the engine for at least 12 hours , so it will have enough time
>to update  it's  own configuration.
>
>9.  Remove the maintenance  and migrate the engine to the other 
>upgraded  host
>
>        10.  Patch the last HostedEngine's host
>
>
>
>I have done this procedure  in order to recover my engine (except
>changing the cpu flags).
>
>
>
>        Note: You may hit some hiccups:
>
>        A) virsh alias
>
>alias virsh='virsh -c
>qemu:///system?authfile=/etc/ovirt-hosted-engine/virsh_auth.conf'
>
>
>
>        B) HostedEngine network missing:
>
>        [root@ovirt1 ~]# virsh net-dumpxml vdsm-ovirtmgmt
>
>        <network>
>
>          <name>vdsm-ovirtmgmt</name>
>
>          <uuid>986c27cf-a1ec-44d8-ae61-ee09ce75c886</uuid>
>
>          <forward mode='bridge'/>
>
>          <bridge name='ovirtmgmt'/>
>
>        </network>
>
>
>
>        Define in xml and add  it via:
>
>        virsh net-define somefile.xml
>
>        C) Missing disk
>
>        Vdsm is creating symlinks like these:
>
>        [root@ovirt1 808423f9-8a5c-40cd-bc9f-2568c85b8c74]# pwd
>
>        /var/run/vdsm/storage/808423f9-8a5c-40cd-bc9f-2568c85b8c74
>
>        [root@ovirt1 808423f9-8a5c-40cd-bc9f-2568c85b8c74]# ls -l
>
>        total 20
>
>lrwxrwxrwx. 1 vdsm kvm 129 Feb  2 19:05
>2c74697a-8bd9-4472-8a98-bf624f3462d5 ->
>/rhev/data-center/mnt/glusterSD/gluster1:_engine/808423f9-8a5c-40cd-bc9f-2568c85b8c74/images/2c74697a-8bd9-4472-8a98-bf624f3462d5
>
>lrwxrwxrwx. 1 vdsm kvm 129 Feb  2 19:09
>3ec27d6d-921c-4348-b799-f50543b6f919 ->
>/rhev/data-center/mnt/glusterSD/gluster1:_engine/808423f9-8a5c-40cd-bc9f-2568c85b8c74/images/3ec27d6d-921c-4348-b799-f50543b6f919
>
>lrwxrwxrwx. 1 vdsm kvm 129 Feb  2 19:09
>441abdc8-6cb1-49a4-903f-a1ec0ed88429 ->
>/rhev/data-center/mnt/glusterSD/gluster1:_engine/808423f9-8a5c-40cd-bc9f-2568c85b8c74/images/441abdc8-6cb1-49a4-903f-a1ec0ed88429
>
>lrwxrwxrwx. 1 vdsm kvm 129 Feb  2 19:09
>94ade632-6ecc-4901-8cec-8e39f3d69cb0 ->
>/rhev/data-center/mnt/glusterSD/gluster1:_engine/808423f9-8a5c-40cd-bc9f-2568c85b8c74/images/94ade632-6ecc-4901-8cec-8e39f3d69cb0
>
>lrwxrwxrwx. 1 vdsm kvm 129 Feb  2 19:05
>fe62a281-51e9-4b23-87b3-2deb52357304 ->
>/rhev/data-center/mnt/glusterSD/gluster1:_engine/808423f9-8a5c-40cd-bc9f-2568c85b8c74/images/fe62a281-51e9-4b23-87b3-2deb52357304
>
>        [root@ovirt1 808423f9-8a5c-40cd-bc9f-2568c85b8c74]#
>
>
>
>Just create the link,  so it  points to correct  destinationand power
>up again.
>
>
>
>
>
>        Good  luck !
>
>
>
>        Best Regards,
>
>        Strahil Nikolov

Hi Marko,

If the  other  VMs work without issues  -> it's worth trying.

Best Regards,
Strahil Nikolov
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/IJEC3B3STD26EMCEEUP72JMKDEBGPJYK/

Reply via email to