On 01/03/2013 05:07 PM, Tom Brown wrote:

interesting, please search for migrationCreate command on desination host and 
search for ERROR afterwords, what do you see?

----- Original Message -----
From: "Tom Brown" <t...@ng23.net>
To: users@ovirt.org
Sent: Thursday, January 3, 2013 4:12:05 PM
Subject: [Users] oVirt 3.1 - VM Migration Issue


Hi

I seem to have an issue with a single VM and migration, other VM's
can migrate OK - When migrating from the GUI it appears to just hang
but in the engine.log i see the following

2013-01-03 14:03:10,359 INFO  [org.ovirt.engine.core.bll.VdsSelector]
(ajp--0.0.0.0-8009-59) Checking for a specific VDS only -
id:a2d84a1e-3e18-11e2-8851-3cd92b4c8e89,
name:ovirt-node.domain-name, host_name(ip):10.192.42.165
2013-01-03 14:03:10,411 INFO
[org.ovirt.engine.core.bll.MigrateVmToServerCommand]
(pool-3-thread-48) [4d32917d] Running command:
MigrateVmToServerCommand internal: false. Entities affected :  ID:
9dc63ce4-0f76-4963-adfe-6f8eb1a44806 Type: VM
2013-01-03 14:03:10,413 INFO  [org.ovirt.engine.core.bll.VdsSelector]
(pool-3-thread-48) [4d32917d] Checking for a specific VDS only -
id:a2d84a1e-3e18-11e2-8851-3cd92b4c8e89,
name:ovirt-node.domain-name, host_name(ip):10.192.42.165
2013-01-03 14:03:11,028 INFO
[org.ovirt.engine.core.vdsbroker.MigrateVDSCommand]
(pool-3-thread-48) [4d32917d] START, MigrateVDSCommand(vdsId =
1a52b722-43a1-11e2-af96-3cd92b4c8e89,
vmId=9dc63ce4-0f76-4963-adfe-6f8eb1a44806, srcHost=10.192.42.196,
dstVdsId=a2d84a1e-3e18-11e2-8851-3cd92b4c8e89,
dstHost=10.192.42.165:54321, migrationMethod=ONLINE), log id:
5011789b
2013-01-03 14:03:11,030 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
(pool-3-thread-48) [4d32917d] VdsBroker::migrate::Entered
(vm_guid=9dc63ce4-0f76-4963-adfe-6f8eb1a44806,
srcHost=10.192.42.196, dstHost=10.192.42.165:54321,  method=online
2013-01-03 14:03:11,031 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
(pool-3-thread-48) [4d32917d] START, MigrateBrokerVDSCommand(vdsId =
1a52b722-43a1-11e2-af96-3cd92b4c8e89,
vmId=9dc63ce4-0f76-4963-adfe-6f8eb1a44806, srcHost=10.192.42.196,
dstVdsId=a2d84a1e-3e18-11e2-8851-3cd92b4c8e89,
dstHost=10.192.42.165:54321, migrationMethod=ONLINE), log id:
7cd53864
2013-01-03 14:03:11,041 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
(pool-3-thread-48) [4d32917d] FINISH, MigrateBrokerVDSCommand, log
id: 7cd53864
2013-01-03 14:03:11,086 INFO
[org.ovirt.engine.core.vdsbroker.MigrateVDSCommand]
(pool-3-thread-48) [4d32917d] FINISH, MigrateVDSCommand, return:
MigratingFrom, log id: 5011789b
2013-01-03 14:03:11,606 INFO
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
(QuartzScheduler_Worker-29) vds::refreshVmList vm id
9dc63ce4-0f76-4963-adfe-6f8eb1a44806 is migrating to vds
ovirt-node.domain-name ignoring it in the refresh till migration is
done
2013-01-03 14:03:12,836 INFO
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
(QuartzScheduler_Worker-36) VM test002.domain-name
9dc63ce4-0f76-4963-adfe-6f8eb1a44806 moved from MigratingFrom --> Up
2013-01-03 14:03:12,837 INFO
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
(QuartzScheduler_Worker-36) adding VM
9dc63ce4-0f76-4963-adfe-6f8eb1a44806 to re-run list
2013-01-03 14:03:12,852 ERROR
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
(QuartzScheduler_Worker-36) Rerun vm
9dc63ce4-0f76-4963-adfe-6f8eb1a44806. Called from vds
ovirt-node002.domain-name
2013-01-03 14:03:12,855 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand]
(pool-3-thread-48) START, MigrateStatusVDSCommand(vdsId =
1a52b722-43a1-11e2-af96-3cd92b4c8e89,
vmId=9dc63ce4-0f76-4963-adfe-6f8eb1a44806), log id: 4721a1f3
2013-01-03 14:03:12,864 ERROR
[org.ovirt.engine.core.vdsbroker.vdsbroker.BrokerCommandBase]
(pool-3-thread-48) Failed in MigrateStatusVDS method
2013-01-03 14:03:12,865 ERROR
[org.ovirt.engine.core.vdsbroker.vdsbroker.BrokerCommandBase]
(pool-3-thread-48) Error code migrateErr and error message
VDSGenericException: VDSErrorException: Failed to MigrateStatusVDS,
error = Fatal error during migration
2013-01-03 14:03:12,865 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.BrokerCommandBase]
(pool-3-thread-48) Command
org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand
return value
Class Name:
org.ovirt.engine.core.vdsbroker.vdsbroker.StatusOnlyReturnForXmlRpc
mStatus                       Class Name:
org.ovirt.engine.core.vdsbroker.vdsbroker.StatusForXmlRpc
mCode                         12
mMessage                      Fatal error during migration


2013-01-03 14:03:12,866 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.BrokerCommandBase]
(pool-3-thread-48) Vds: ovirt-node002.itvonline.ads
2013-01-03 14:03:12,867 ERROR
[org.ovirt.engine.core.vdsbroker.VDSCommandBase] (pool-3-thread-48)
Command MigrateStatusVDS execution failed. Exception:
VDSErrorException: VDSGenericException: VDSErrorException: Failed to
MigrateStatusVDS, error = Fatal error during migration
2013-01-03 14:03:12,867 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand]
(pool-3-thread-48) FINISH, MigrateStatusVDSCommand, log id: 4721a1f3

Does anyone have any idea what this might be? I am using 3.1 from
dreyou as these are CentOS 6 nodes

any clue on which log on the new host ? I see the following in messages
VDSM is the virtualization agent. look at /var/log/vdsm/vdsm.log

Jan  3 16:03:20 ovirt-node vdsm Storage.LVM WARNING lvm vgs failed: 5 [] ['  Volume group 
"ab686999-f320-4a61-ae07-e99c2f858996" not found']
Jan  3 16:03:20 ovirt-node vdsm Storage.StorageDomain WARNING Resource 
namespace ab686999-f320-4a61-ae07-e99c2f858996_imageNS already registered
Jan  3 16:03:20 ovirt-node vdsm Storage.StorageDomain WARNING Resource 
namespace ab686999-f320-4a61-ae07-e99c2f858996_volumeNS already registered
Jan  3 16:03:58 ovirt-node vdsm vm.Vm WARNING 
vmId=`9dc63ce4-0f76-4963-adfe-6f8eb1a44806`::Unknown type found, device: 
'{'device': 'unix', 'alias': 'channel0', 'type': 'channel', 'address': {'bus': 
'0', 'controller': '0', 'type': 'virtio-serial', 'port': '1'}}' found
Jan  3 16:03:58 ovirt-node vdsm vm.Vm WARNING 
vmId=`9dc63ce4-0f76-4963-adfe-6f8eb1a44806`::Unknown type found, device: 
'{'device': 'unix', 'alias': 'channel1', 'type': 'channel', 'address': {'bus': 
'0', 'controller': '0', 'type': 'virtio-serial', 'port': '2'}}' found
Jan  3 16:03:59 ovirt-node kernel: device vnet2 entered promiscuous mode
Jan  3 16:03:59 ovirt-node kernel: ovirtmgmt: port 4(vnet2) entering forwarding 
state
Jan  3 16:03:59 ovirt-node kernel: ovirtmgmt: port 4(vnet2) entering disabled 
state
Jan  3 16:03:59 ovirt-node kernel: device vnet2 left promiscuous mode
Jan  3 16:03:59 ovirt-node kernel: ovirtmgmt: port 4(vnet2) entering disabled 
state

and the following in the qemu log for that VM on the new node

2013-01-03 16:03:59.706+0000: starting up
LC_ALL=C PATH=/sbin:/usr/sbin:/bin:/usr/bin QEMU_AUDIO_DRV=none 
/usr/libexec/qemu-kvm -S -M rhel6.3.0 -cpu Nehalem -enable-kvm -m 1024 -smp 
1,sockets=1,cores=1,threads=1 -name test002.itvonline.ads -uuid 
9dc63ce4-0f76-4963-adfe-6f8eb1a44806 -smbios 
type=1,manufacturer=oVirt,product=oVirt 
Node,version=6-3.el6.centos.9,serial=55414E03-C241-11DF-BBDA-64093408D485_d4:85:64:09:34:08,uuid=9dc63ce4-0f76-4963-adfe-6f8eb1a44806
 -nodefconfig -nodefaults -chardev 
socket,id=charmonitor,path=/var/lib/libvirt/qemu/test002.itvonline.ads.monitor,server,nowait
 -mon chardev=charmonitor,id=monitor,mode=control -rtc 
base=2013-01-03T16:03:58,driftfix=slew -no-shutdown -device 
virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -device 
piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -drive 
if=none,media=cdrom,id=drive-ide0-1-0,readonly=on,format=raw,serial= -device 
ide-drive,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0,bootindex=1 -drive 
file=/rhev/data-center/bb0beebf-edab-41e2-83b8-16bdbb
c5
dda7/2a1939bd-9fa3-4896-b8a9-46234172aae7/images/e8711e5d-2f06-4c0f-b5c6-fa0806d7448f/0d93c51f-f838-4143-815c-9b3457d1a934,if=none,id=drive-virtio-disk0,format=raw,serial=e8711e5d-2f06-4c0f-b5c6-fa0806d7448f,cache=none,werror=stop,rerror=stop,aio=threads -device virtio-blk-pci,scsi=off,bus=pci.0,addr=0x5,drive=drive-virtio-disk0,id=virtio-disk0 -netdev tap,fd=32,id=hostnet0,vhost=on,vhostfd=33 -device virtio-net-pci,netdev=hostnet0,id=net0,mac=00:1a:4a:c0:2a:00,bus=pci.0,addr=0x3 -chardev socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/test002.itvonline.ads.com.redhat.rhevm.vdsm,server,nowait -device virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=com.redhat.rhevm.vdsm -chardev socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/test002.itvonline.ads.org.qemu.guest_agent.0,server,nowait -device virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=channel1,name=org.qemu.guest_agent.0 -chardev pty,id=charconsole0
-de
  vice virtconsole,chardev=charconsole0,id=console0 -device 
usb-tablet,id=input0 -vnc 10.192.42.165:4,password -k en-us -vga qxl -global 
qxl-vga.vram_size=67108864 -incoming tcp:0.0.0.0:49160 -device 
virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x6
2013-01-03 16:03:59.955+0000: shutting down

but thats about it?

thanks

_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to