On February 16, 2020 11:40:37 AM GMT+02:00, "Vrgotic, Marko" 
<m.vrgo...@activevideo.com> wrote:
>Hi Strahil,
>
>Thank you for proposed steps. If this is now the only way to go,
>without having to do restore, I will take it.
>Before I start, I will first gather information and later share the
>outcome.
>
>What does not make me happy, is the cause.
>What is the reason CPU family type of the processor would be changed
>upon upgrade? Can you please share with me?
>At least for the Hosts which are part of HA, the upgrade should check
>and if possible give a warning if any of HA cluster crucial parameters
>will be changed.
>
>
>-----
>kind regards/met vriendelijke groeten
> 
>Marko Vrgotic
>ActiveVideo
>
>
>On 14/02/2020, 18:44, "Strahil Nikolov" <hunter86...@yahoo.com> wrote:
>
>On February 14, 2020 4:19:53 PM GMT+02:00, "Vrgotic, Marko"
><m.vrgo...@activevideo.com> wrote:
>    >Good answer Strahil,
>    >
>    >Thank you, I forgot.
>    >
>    >Libvirt logs are actually showing the reason why:
>    >
>    >2020-02-14T12:33:51.847970Z qemu-kvm: -drive
>>file=/var/run/vdsm/storage/054c43fc-1924-4106-9f80-0f2ac62b9886/b019c5fa-8fb5-4bfc-8339-f5b7f590a051/f1ce8ba6-2d3b-4309-bca0-e6a00ce74c75,format=raw,if=none,id=drive-ua-b019c5fa-8fb5-4bfc-8339-f5b7f590a051,serial=b019c5fa-8fb5-4bfc-8339-f5b7f590a051,werror=stop,rerror=stop,cache=none,aio=threads:
>    >'serial' is deprecated, please use the corresponding option of
>    >'-device' instead
>    >Spice-Message: 04:33:51.856: setting TLS option 'CipherString' to
>    >'kECDHE+FIPS:kDHE+FIPS:kRSA+FIPS:!eNULL:!aNULL' from
>    >/etc/pki/tls/spice.cnf configuration file
>  >2020-02-14T12:33:51.863449Z qemu-kvm: warning: CPU(s) not present in
> >any NUMA nodes: CPU 4 [socket-id: 1, core-id: 0, thread-id: 0], CPU 5
>>[socket-id: 1, core-id: 1, thread-id: 0], CPU 6 [socket-id: 1,
>core-id:
>>2, thread-id: 0], CPU 7 [socket-id: 1, core-id: 3, thread-id: 0], CPU
>8
>>[socket-id: 2, core-id: 0, thread-id: 0], CPU 9 [socket-id: 2,
>core-id:
>>1, thread-id: 0], CPU 10 [socket-id: 2, core-id: 2, thread-id: 0], CPU
>    >11 [socket-id: 2, core-id: 3, thread-id: 0], CPU 12 [socket-id: 3,
>>core-id: 0, thread-id: 0], CPU 13 [socket-id: 3, core-id: 1,
>thread-id:
>>0], CPU 14 [socket-id: 3, core-id: 2, thread-id: 0], CPU 15
>[socket-id:
>    >3, core-id: 3, thread-id: 0], CPU 16 [socket-id: 4, core-id: 0,
>>thread-id: 0], CPU 17 [socket-id: 4, core-id: 1, thread-id: 0], CPU 18
>    >[socket-id: 4, core-id: 2, thread-id: 0], CPU 19 [socket-id: 4,
>>core-id: 3, thread-id: 0], CPU 20 [socket-id: 5, core-id: 0,
>thread-id:
>>0], CPU 21 [socket-id: 5, core-id: 1, thread-id: 0], CPU 22
>[socket-id:
>    >5, core-id: 2, thread-id: 0], CPU 23 [socket-id: 5, core-id: 3,
>>thread-id: 0], CPU 24 [socket-id: 6, core-id: 0, thread-id: 0], CPU 25
>    >[socket-id: 6, core-id: 1, thread-id: 0], CPU 26 [socket-id: 6,
>>core-id: 2, thread-id: 0], CPU 27 [socket-id: 6, core-id: 3,
>thread-id:
>>0], CPU 28 [socket-id: 7, core-id: 0, thread-id: 0], CPU 29
>[socket-id:
>    >7, core-id: 1, thread-id: 0], CPU 30 [socket-id: 7, core-id: 2,
>>thread-id: 0], CPU 31 [socket-id: 7, core-id: 3, thread-id: 0], CPU 32
>    >[socket-id: 8, core-id: 0, thread-id: 0], CPU 33 [socket-id: 8,
>>core-id: 1, thread-id: 0], CPU 34 [socket-id: 8, core-id: 2,
>thread-id:
>>0], CPU 35 [socket-id: 8, core-id: 3, thread-id: 0], CPU 36
>[socket-id:
>    >9, core-id: 0, thread-id: 0], CPU 37 [socket-id: 9, core-id: 1,
>>thread-id: 0], CPU 38 [socket-id: 9, core-id: 2, thread-id: 0], CPU 39
>    >[socket-id: 9, core-id: 3, thread-id: 0], CPU 40 [socket-id: 10,
>    >core-id: 0, thread-id: 0], CPU 41 [socket-id: 10, core-id: 1,
>>thread-id: 0], CPU 42 [socket-id: 10, core-id: 2, thread-id: 0], CPU
>43
>    >[socket-id: 10, core-id: 3, thread-id: 0], CPU 44 [socket-id: 11,
>    >core-id: 0, thread-id: 0], CPU 45 [socket-id: 11, core-id: 1,
>>thread-id: 0], CPU 46 [socket-id: 11, core-id: 2, thread-id: 0], CPU
>47
>    >[socket-id: 11, core-id: 3, thread-id: 0], CPU 48 [socket-id: 12,
>    >core-id: 0, thread-id: 0], CPU 49 [socket-id: 12, core-id: 1,
>>thread-id: 0], CPU 50 [socket-id: 12, core-id: 2, thread-id: 0], CPU
>51
>    >[socket-id: 12, core-id: 3, thread-id: 0], CPU 52 [socket-id: 13,
>    >core-id: 0, thread-id: 0], CPU 53 [socket-id: 13, core-id: 1,
>>thread-id: 0], CPU 54 [socket-id: 13, core-id: 2, thread-id: 0], CPU
>55
>    >[socket-id: 13, core-id: 3, thread-id: 0], CPU 56 [socket-id: 14,
>    >core-id: 0, thread-id: 0], CPU 57 [socket-id: 14, core-id: 1,
>>thread-id: 0], CPU 58 [socket-id: 14, core-id: 2, thread-id: 0], CPU
>59
>    >[socket-id: 14, core-id: 3, thread-id: 0], CPU 60 [socket-id: 15,
>    >core-id: 0, thread-id: 0], CPU 61 [socket-id: 15, core-id: 1,
>>thread-id: 0], CPU 62 [socket-id: 15, core-id: 2, thread-id: 0], CPU
>63
>    >[socket-id: 15, core-id: 3, thread-id: 0]
>>2020-02-14T12:33:51.863475Z qemu-kvm: warning: All CPU(s) up to
>maxcpus
>  >should be described in NUMA config, ability to start up with partial
>    >NUMA mappings is obsoleted and will be removed in future
>   >2020-02-14T12:33:51.863973Z qemu-kvm: warning: host doesn't support
>    >requested feature: CPUID.07H:EDX.md-clear [bit 10]
>   >2020-02-14T12:33:51.865066Z qemu-kvm: warning: host doesn't support
>    >requested feature: CPUID.07H:EDX.md-clear [bit 10]
>   >2020-02-14T12:33:51.865547Z qemu-kvm: warning: host doesn't support
>    >requested feature: CPUID.07H:EDX.md-clear [bit 10]
>   >2020-02-14T12:33:51.865996Z qemu-kvm: warning: host doesn't support
>    >requested feature: CPUID.07H:EDX.md-clear [bit 10]
>    >2020-02-14 12:33:51.932+0000: shutting down, reason=failed
>    >
>    >But then I wonder if the following is related to error above:
>    >
>    >Before I started upgrading Host by Host, all Hosts in Cluster were
>    >showing CPU Family type: " Intel SandyBridge IBRS SSBD MDS Family"
>  >After first Host was upgraded, his CPU Family type was changed to: "
>    >Intel SandyBridge IBRS SSBD Family" and that forced me to have do
>    >"downgrade" Cluster family type to " Intel SandyBridge IBRS SSBD
> >Family" in order to be able to Activate the Host back inside Cluster.
>>Following further, each Host CPU family type changed after Upgrade
>from
>>"" Intel SandyBridge IBRS SSBD MDS Family" to "" Intel SandyBridge
>IBRS
>    >SSBD Family" , except one where HostedEngine is currently one. 
>    >
>>Could this possibly be the reason why I cannot Migrate the
>HostedEngine
>    >now and how to solve it?
>    >
>    >Kindly awaiting your reply.
>    >
>    >
>    >-----
>    >kind regards/met vriendelijke groeten
>    > 
>    >Marko Vrgotic
>    >Sr. System Engineer @ System Administration
>    >
>    >ActiveVideo
>    >o: +31 (35) 6774131
>    >e: m.vrgo...@activevideo.com
>    >w: www.activevideo.com <http://www.activevideo.com>
>    > 
>    >ActiveVideo Networks BV. Mediacentrum 3745 Joop van den Endeplein
>>1.1217 WJ Hilversum, The Netherlands. The information contained in
>this
> >message may be legally privileged and confidential. It is intended to
>>be read only by the individual or entity to whom it is addressed or by
>    >their designee. If the reader of this message is not the intended
>>recipient, you are on notice that any distribution of this message, in
>>any form, is strictly prohibited.  If you have received this message
>in
>    >error, please immediately notify the sender and/or ActiveVideo
>>Networks, LLC by telephone at +1 408.931.9200 and delete or destroy
>any
>    >copy of this message.
>    > 
>    >
>    >
>    >
>    >
>    > 
>    >
>>On 14/02/2020, 14:01, "Strahil Nikolov" <hunter86...@yahoo.com> wrote:
>    >
>    >On February 14, 2020 2:47:04 PM GMT+02:00, "Vrgotic, Marko"
>    ><m.vrgo...@activevideo.com> wrote:
>    >    >Dear oVirt,
>    >    >
>> >I have problem migrating HostedEngine, only HA VM server, to other
>HA
>    >    >nodes.
>    >    >
>    >    >Bit of background story:
>    >    >
>    >    >  *   We have oVirt SHE 4.3.5
>    >    >  *   Three Nodes act as HA pool for SHE
>    >    >  *   Node 3 is currently Hosting SHE
>    >    >  *   Actions:
>>>*   Put Node1 in Maintenance mode, all VMs were successfully
>migrated,
>    >    >than Upgrade packages, Activate Host – all looks good
>>>*   Put Node2 in Maintenance mode, all VMs were successfully
>migrated,
>    >    >than Upgrade packages, Activate Host – all looks good
>    >    >
>    >    >Not the problem:
> >    >Try to set  Node3 in Maintenance mode, all VMs were successfully
>    >    >migrated, except HostedEngine.
>    >    >
>  >    >When attempting Migration of the VM HostedEngine, it fails with
>    >    >following error message:
>    >    >
>    >    >2020-02-14 12:33:49,960Z INFO 
>  >    >[org.ovirt.engine.core.bll.MigrateVmCommand] (default task-265)
>   >    >[16f4559e-e262-4c9d-80b4-ec81c2cbf950] Lock Acquired to object
>>>'EngineLock:{exclusiveLocks='[66b6d489-ceb8-486a-951a-355e21f13627=VM]',
>    >    >sharedLocks=''}'
>    >    >2020-02-14 12:33:49,984Z INFO 
>>    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager] (default
>  >    >task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] Candidate host
>>   >'ovirt-sj-04.ictv.com' ('d98843da-bd81-46c9-9425-065b196ac59d')
>was
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>    >    >id: null)
>    >    >2020-02-14 12:33:49,984Z INFO 
>>    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager] (default
>  >    >task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] Candidate host
>>   >'ovirt-sj-05.ictv.com' ('e3176705-9fb0-41d6-8721-367dfa2e62bd')
>was
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>    >    >id: null)
>    >    >2020-02-14 12:33:49,997Z INFO 
>  >    >[org.ovirt.engine.core.bll.MigrateVmCommand] (default task-265)
>    >    >[16f4559e-e262-4c9d-80b4-ec81c2cbf950] Running command:
>    >    >MigrateVmCommand internal: false. Entities affected :  ID:
>>  >66b6d489-ceb8-486a-951a-355e21f13627 Type: VMAction group
>MIGRATE_VM
>    >    >with role type USER
>    >    >2020-02-14 12:33:50,008Z INFO 
>>    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager] (default
>  >    >task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] Candidate host
>>   >'ovirt-sj-04.ictv.com' ('d98843da-bd81-46c9-9425-065b196ac59d')
>was
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>    >    >id: 16f4559e-e262-4c9d-80b4-ec81c2cbf950)
>    >    >2020-02-14 12:33:50,008Z INFO 
>>    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager] (default
>  >    >task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] Candidate host
>>   >'ovirt-sj-05.ictv.com' ('e3176705-9fb0-41d6-8721-367dfa2e62bd')
>was
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>    >    >id: 16f4559e-e262-4c9d-80b4-ec81c2cbf950)
>    >    >2020-02-14 12:33:50,033Z INFO 
>>>[org.ovirt.engine.core.vdsbroker.MigrateVDSCommand] (default
>task-265)
> >    >[16f4559e-e262-4c9d-80b4-ec81c2cbf950] START, MigrateVDSCommand(
>>>MigrateVDSCommandParameters:{hostId='f8d27efb-1527-45f0-97d6-d34a86abaaa2',
>    >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627',
>    >    >srcHost='ovirt-sj-03.ictv.com',
>    >    >dstVdsId='9808f434-5cd4-48b5-8bbc-e639e391c6a5',
>  >    >dstHost='ovirt-sj-01.ictv.com:54321', migrationMethod='ONLINE',
>>  >tunnelMigration='false', migrationDowntime='0',
>autoConverge='true',
>>  >migrateCompressed='false', consoleAddress='null',
>maxBandwidth='40',
>    >    >enableGuestEvents='true', maxIncomingMigrations='2',
>    >    >maxOutgoingMigrations='2',
>   >    >convergenceSchedule='[init=[{name=setDowntime, params=[100]}],
>    >>stalling=[{limit=1, action={name=setDowntime, params=[150]}},
>    >{limit=2,
>    >    >action={name=setDowntime, params=[200]}}, {limit=3,
>    >    >action={name=setDowntime, params=[300]}}, {limit=4,
>    >    >action={name=setDowntime, params=[400]}}, {limit=6,
>    >    >action={name=setDowntime, params=[500]}}, {limit=-1,
>> >action={name=abort, params=[]}}]]', dstQemu='10.210.13.11'}), log
>id:
>    >    >5c126a47
>    >    >2020-02-14 12:33:50,036Z INFO 
>>  
>>[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
> >    >(default task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] START,
>    >    >MigrateBrokerVDSCommand(HostName = ovirt-sj-03.ictv.com,
>>>MigrateVDSCommandParameters:{hostId='f8d27efb-1527-45f0-97d6-d34a86abaaa2',
>    >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627',
>    >    >srcHost='ovirt-sj-03.ictv.com',
>    >    >dstVdsId='9808f434-5cd4-48b5-8bbc-e639e391c6a5',
>  >    >dstHost='ovirt-sj-01.ictv.com:54321', migrationMethod='ONLINE',
>>  >tunnelMigration='false', migrationDowntime='0',
>autoConverge='true',
>>  >migrateCompressed='false', consoleAddress='null',
>maxBandwidth='40',
>    >    >enableGuestEvents='true', maxIncomingMigrations='2',
>    >    >maxOutgoingMigrations='2',
>   >    >convergenceSchedule='[init=[{name=setDowntime, params=[100]}],
>    >>stalling=[{limit=1, action={name=setDowntime, params=[150]}},
>    >{limit=2,
>    >    >action={name=setDowntime, params=[200]}}, {limit=3,
>    >    >action={name=setDowntime, params=[300]}}, {limit=4,
>    >    >action={name=setDowntime, params=[400]}}, {limit=6,
>    >    >action={name=setDowntime, params=[500]}}, {limit=-1,
>> >action={name=abort, params=[]}}]]', dstQemu='10.210.13.11'}), log
>id:
>    >    >a0f776d
>    >    >2020-02-14 12:33:50,043Z INFO 
>>  
>>[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
>>    >(default task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] FINISH,
>    >    >MigrateBrokerVDSCommand, return: , log id: a0f776d
>    >    >2020-02-14 12:33:50,046Z INFO 
>>>[org.ovirt.engine.core.vdsbroker.MigrateVDSCommand] (default
>task-265)
>>    >[16f4559e-e262-4c9d-80b4-ec81c2cbf950] FINISH, MigrateVDSCommand,
>    >    >return: MigratingFrom, log id: 5c126a47
>    >    >2020-02-14 12:33:50,052Z INFO 
>>>[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>>   >(default task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950]
>EVENT_ID:
>>  >VM_MIGRATION_START(62), Migration started (VM: HostedEngine,
>Source:
>   >    >ovirt-sj-03.ictv.com, Destination: ovirt-sj-01.ictv.com, User:
>    >    >mvrgo...@ictv.com@ictv.com-authz).
>    >    >2020-02-14 12:33:52,893Z INFO 
>    >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>>>(ForkJoinPool-1-worker-8) [] VM
>'66b6d489-ceb8-486a-951a-355e21f13627'
>    >    >was reported as Down on VDS
>    >    >'9808f434-5cd4-48b5-8bbc-e639e391c6a5'(ovirt-sj-01.ictv.com)
>    >    >2020-02-14 12:33:52,893Z INFO 
>    >    >[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
> >    >(ForkJoinPool-1-worker-8) [] START, DestroyVDSCommand(HostName =
>    >    >ovirt-sj-01.ictv.com,
>>>DestroyVmVDSCommandParameters:{hostId='9808f434-5cd4-48b5-8bbc-e639e391c6a5',
>  >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627', secondsToWait='0',
>>  >gracefully='false', reason='', ignoreNoVm='true'}), log id:
>7532a8c0
>    >    >2020-02-14 12:33:53,217Z INFO 
>    >    >[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>    >    >(ForkJoinPool-1-worker-8) [] Failed to destroy VM
>>    >'66b6d489-ceb8-486a-951a-355e21f13627' because VM does not exist,
>    >    >ignoring
>    >    >2020-02-14 12:33:53,217Z INFO 
>    >    >[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>> >(ForkJoinPool-1-worker-8) [] FINISH, DestroyVDSCommand, return: ,
>log
>    >    >id: 7532a8c0
>    >    >2020-02-14 12:33:53,217Z INFO 
>    >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>    >    >(ForkJoinPool-1-worker-8) [] VM
>> >'66b6d489-ceb8-486a-951a-355e21f13627'(HostedEngine) was
>unexpectedly
>    >    >detected as 'Down' on VDS
>>>'9808f434-5cd4-48b5-8bbc-e639e391c6a5'(ovirt-sj-01.ictv.com)
>(expected
>    >    >on 'f8d27efb-1527-45f0-97d6-d34a86abaaa2')
>    >    >2020-02-14 12:33:53,217Z ERROR
>    >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>>   >(ForkJoinPool-1-worker-8) [] Migration of VM 'HostedEngine' to
>host
>  >    >'ovirt-sj-01.ictv.com' failed: VM destroyed during the startup.
>    >    >2020-02-14 12:33:53,219Z INFO 
>    >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>    >    >(ForkJoinPool-1-worker-15) [] VM
>  >    >'66b6d489-ceb8-486a-951a-355e21f13627'(HostedEngine) moved from
>    >    >'MigratingFrom' --> 'Up'
>    >    >2020-02-14 12:33:53,219Z INFO 
>    >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>    >    >(ForkJoinPool-1-worker-15) [] Adding VM
>>   >'66b6d489-ceb8-486a-951a-355e21f13627'(HostedEngine) to re-run
>list
>    >    >2020-02-14 12:33:53,221Z ERROR
>    >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmsMonitoring]
>    >    >(ForkJoinPool-1-worker-15) [] Rerun VM
>    >    >'66b6d489-ceb8-486a-951a-355e21f13627'. Called from VDS
>    >    >'ovirt-sj-03.ictv.com'
>    >    >2020-02-14 12:33:53,259Z INFO 
>>  
>>[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand]
>    >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] START,
>    >    >MigrateStatusVDSCommand(HostName = ovirt-sj-03.ictv.com,
>>>MigrateStatusVDSCommandParameters:{hostId='f8d27efb-1527-45f0-97d6-d34a86abaaa2',
>  >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627'}), log id: 62bac076
>    >    >2020-02-14 12:33:53,265Z INFO 
>>  
>>[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand]
>    >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] FINISH,
>    >    >MigrateStatusVDSCommand, return: , log id: 62bac076
>    >    >2020-02-14 12:33:53,277Z WARN 
>>>[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>    >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] EVENT_ID:
>>  >VM_MIGRATION_TRYING_RERUN(128), Failed to migrate VM HostedEngine
>to
>   >    >Host ovirt-sj-01.ictv.com . Trying to migrate to another Host.
>    >    >2020-02-14 12:33:53,330Z INFO 
>    >    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
> >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] Candidate host
>>   >'ovirt-sj-04.ictv.com' ('d98843da-bd81-46c9-9425-065b196ac59d')
>was
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>    >    >id: null)
>    >    >2020-02-14 12:33:53,330Z INFO 
>    >    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
> >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] Candidate host
>>   >'ovirt-sj-05.ictv.com' ('e3176705-9fb0-41d6-8721-367dfa2e62bd')
>was
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>    >    >id: null)
>    >    >2020-02-14 12:33:53,345Z INFO 
>    >    >[org.ovirt.engine.core.bll.MigrateVmCommand]
>>    >(EE-ManagedThreadFactory-engine-Thread-377323) [] Running
>command:
>    >    >MigrateVmCommand internal: false. Entities affected :  ID:
>>  >66b6d489-ceb8-486a-951a-355e21f13627 Type: VMAction group
>MIGRATE_VM
>    >    >with role type USER
>    >    >2020-02-14 12:33:53,356Z INFO 
>    >    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
> >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] Candidate host
>>   >'ovirt-sj-04.ictv.com' ('d98843da-bd81-46c9-9425-065b196ac59d')
>was
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>    >    >id: 16f4559e-e262-4c9d-80b4-ec81c2cbf950)
>    >    >2020-02-14 12:33:53,356Z INFO 
>    >    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
> >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] Candidate host
>>   >'ovirt-sj-05.ictv.com' ('e3176705-9fb0-41d6-8721-367dfa2e62bd')
>was
>>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>(correlation
>    >    >id: 16f4559e-e262-4c9d-80b4-ec81c2cbf950)
>    >    >2020-02-14 12:33:53,380Z INFO 
>    >    >[org.ovirt.engine.core.vdsbroker.MigrateVDSCommand]
>    >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] START,
>    >    >MigrateVDSCommand(
>>>MigrateVDSCommandParameters:{hostId='f8d27efb-1527-45f0-97d6-d34a86abaaa2',
>    >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627',
>    >    >srcHost='ovirt-sj-03.ictv.com',
>    >    >dstVdsId='33e8ff78-e396-4f40-b43c-685bfaaee9af',
>  >    >dstHost='ovirt-sj-02.ictv.com:54321', migrationMethod='ONLINE',
>>  >tunnelMigration='false', migrationDowntime='0',
>autoConverge='true',
>>  >migrateCompressed='false', consoleAddress='null',
>maxBandwidth='40',
>    >    >enableGuestEvents='true', maxIncomingMigrations='2',
>    >    >maxOutgoingMigrations='2',
>   >    >convergenceSchedule='[init=[{name=setDowntime, params=[100]}],
>    >>stalling=[{limit=1, action={name=setDowntime, params=[150]}},
>    >{limit=2,
>    >    >action={name=setDowntime, params=[200]}}, {limit=3,
>    >    >action={name=setDowntime, params=[300]}}, {limit=4,
>    >    >action={name=setDowntime, params=[400]}}, {limit=6,
>    >    >action={name=setDowntime, params=[500]}}, {limit=-1,
>> >action={name=abort, params=[]}}]]', dstQemu='10.210.13.12'}), log
>id:
>    >    >d99059f
>    >    >2020-02-14 12:33:53,380Z INFO 
>>  
>>[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
>    >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] START,
>    >    >MigrateBrokerVDSCommand(HostName = ovirt-sj-03.ictv.com,
>>>MigrateVDSCommandParameters:{hostId='f8d27efb-1527-45f0-97d6-d34a86abaaa2',
>    >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627',
>    >    >srcHost='ovirt-sj-03.ictv.com',
>    >    >dstVdsId='33e8ff78-e396-4f40-b43c-685bfaaee9af',
>  >    >dstHost='ovirt-sj-02.ictv.com:54321', migrationMethod='ONLINE',
>>  >tunnelMigration='false', migrationDowntime='0',
>autoConverge='true',
>>  >migrateCompressed='false', consoleAddress='null',
>maxBandwidth='40',
>    >    >enableGuestEvents='true', maxIncomingMigrations='2',
>    >    >maxOutgoingMigrations='2',
>   >    >convergenceSchedule='[init=[{name=setDowntime, params=[100]}],
>    >>stalling=[{limit=1, action={name=setDowntime, params=[150]}},
>    >{limit=2,
>    >    >action={name=setDowntime, params=[200]}}, {limit=3,
>    >    >action={name=setDowntime, params=[300]}}, {limit=4,
>    >    >action={name=setDowntime, params=[400]}}, {limit=6,
>    >    >action={name=setDowntime, params=[500]}}, {limit=-1,
>> >action={name=abort, params=[]}}]]', dstQemu='10.210.13.12'}), log
>id:
>    >    >6f0483ac
>    >    >2020-02-14 12:33:53,386Z INFO 
>>  
>>[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
>    >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] FINISH,
>    >    >MigrateBrokerVDSCommand, return: , log id: 6f0483ac
>    >    >2020-02-14 12:33:53,388Z INFO 
>    >    >[org.ovirt.engine.core.vdsbroker.MigrateVDSCommand]
>    >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] FINISH,
>    >    >MigrateVDSCommand, return: MigratingFrom, log id: d99059f
>    >    >2020-02-14 12:33:53,391Z INFO 
>>>[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>    >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] EVENT_ID:
>>  >VM_MIGRATION_START(62), Migration started (VM: HostedEngine,
>Source:
>   >    >ovirt-sj-03.ictv.com, Destination: ovirt-sj-02.ictv.com, User:
>    >    >mvrgo...@ictv.com@ictv.com-authz).
>    >    >2020-02-14 12:33:55,108Z INFO 
>>    >[org.ovirt.engine.core.vdsbroker.monitoring.VmsStatisticsFetcher]
>> >(EE-ManagedThreadFactory-engineScheduled-Thread-96) [] Fetched 10
>VMs
>    >    >from VDS '33e8ff78-e396-4f40-b43c-685bfaaee9af'
>    >    >2020-02-14 12:33:55,110Z INFO 
>    >    >[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>    >    >

Hi Marko,

I  can't claim that this is 'The Only Solution'   as  I never hit your 
problem.Not exactly the same - mine was a broken selinux package :) .

My guess  is that either:
A)  Your Hypervisor CPU was  deprecated either  in 
qemu/vdsm/libvirt/whatever_component_it_is
B)  You hit a BUG


I agree that some warning should be provided to the cluster adminiatrator, yet 
I have never heard of such issue before.

The proposed steps are  actually a workaround - as you change your VMs CPU type 
to adjust to unknown change in oVirt/KVM/Qemu .

Still, don't forget to get a backup of the engine prior making any 
modifications.


P.S.: You have patched  the engine first,  right ?

Best Regards,
Strahil Nikolov
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/OHPWRLILQPVTQTYRY4V4EJ2H3BQR3B4A/

Reply via email to