I'm not sure, but I always thought that you need an agent for live migrations. You can always try installing either qemu-guest-agent or ovirt-guest-agent and check if live migration between hosts is possible.
Have you set the new cluster/dc version ? Best Regards Strahil NikolovOn Jul 9, 2019 17:42, Neil <nwilson...@gmail.com> wrote: > > I remember seeing the bug earlier but because it was closed thought it was > unrelated, this appears to be it.... > > https://bugzilla.redhat.com/show_bug.cgi?id=1670701 > > Perhaps I'm not understanding your question about the VM guest agent, but I > don't have any guest agent currently installed on the VM, not sure if the > output of my qemu-kvm process maybe answers this question?.... > > /usr/libexec/qemu-kvm -name guest=Headoffice.cbl-ho.local,debug-threads=on -S > -object > secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/domain-1-Headoffice.cbl-ho.lo/master-key.aes > -machine pc-i440fx-rhel7.3.0,accel=kvm,usb=off,dump-guest-core=off -cpu > Broadwell,vme=on,f16c=on,rdrand=on,hypervisor=on,arat=on,xsaveopt=on,abm=on,rtm=on,hle=on > -m 8192 -realtime mlock=off -smp 8,maxcpus=64,sockets=16,cores=4,threads=1 > -numa node,nodeid=0,cpus=0-7,mem=8192 -uuid > 9a6561b8-5702-43dc-9e92-1dc5dfed4eef -smbios > type=1,manufacturer=oVirt,product=oVirt > Node,version=7-3.1611.el7.centos,serial=4C4C4544-0034-5810-8033-C2C04F4E4B32,uuid=9a6561b8-5702-43dc-9e92-1dc5dfed4eef > -no-user-config -nodefaults -chardev > socket,id=charmonitor,fd=31,server,nowait -mon > chardev=charmonitor,id=monitor,mode=control -rtc > base=2019-07-09T10:26:53,driftfix=slew -global kvm-pit.lost_tick_policy=delay > -no-hpet -no-shutdown -boot strict=on -device > piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device > virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x4 -device > virtio-serial-pci,id=virtio-serial0,max_ports=16,bus=pci.0,addr=0x5 -drive > if=none,id=drive-ide0-1-0,readonly=on -device > ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -drive > file=/rhev/data-center/59831b91-00a5-01e4-0294-000000000018/8a607f8a-542a-473c-bb18-25c05fe2a3d4/images/56e8240c-a172-4f52-b0c1-2bddc4f34f93/9f245467-d31d-4f5a-8037-7c5012a4aa84,format=qcow2,if=none,id=drive-virtio-disk0,serial=56e8240c-a172-4f52-b0c1-2bddc4f34f93,werror=stop,rerror=stop,cache=none,aio=native > -device > virtio-blk-pci,scsi=off,bus=pci.0,addr=0x7,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1,write-cache=on > -netdev tap,fd=33,id=hostnet0,vhost=on,vhostfd=34 -device > virtio-net-pci,netdev=hostnet0,id=net0,mac=00:1a:4a:16:01:5b,bus=pci.0,addr=0x3 > -chardev socket,id=charchannel0,fd=35,server,nowait -device > virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=com.redhat.rhevm.vdsm > -chardev socket,id=charchannel1,fd=36,server,nowait -device > virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=channel1,name=org.qemu.guest_agent.0 > -chardev spicevmc,id=charchannel2,name=vdagent -device > virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel2,id=channel2,name=com.redhat.spice.0 > -spice > tls-port=5900,addr=10.0.1.11,x509-dir=/etc/pki/vdsm/libvirt-spice,tls-channel=default,tls-channel=main,tls-channel=display,tls-channel=inputs,tls-channel=cursor,tls-channel=playback,tls-channel=record,tls-channel=smartcard,tls-channel=usbredir,seamless-migration=on > -device > qxl-vga,id=video0,ram_size=67108864,vram_size=8388608,vram64_size_mb=0,vgamem_mb=16,max_outputs=1,bus=pci.0,addr=0x2 > -incoming defer -device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x6 > -object rng-random,id=objrng0,filename=/dev/urandom -device > virtio-rng-pci,rng=objrng0,id=rng0,bus=pci.0,addr=0x8 -sandbox > on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny -msg > timestamp=on > > Please shout if you need further info. > > Thanks. > > > > > > > On Tue, Jul 9, 2019 at 4:17 PM Strahil Nikolov <hunter86...@yahoo.com> wrote: >> >> Shouldn't cause that problem. >> >> You have to find the bug in bugzilla and report a regression (if it's not >> closed) , or open a new one and report the regression. >> As far as I remember , only the dashboard was affected due to new features >> about vdo disk savings. >>
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/IQCHU3VAIQQCG7NSBYK5UMZYFRTJ7B2E/