My un-authoritative answer: I've been tracking this bug (or several with the
same symptoms) for going on a couple years. It's ridiculously common,
apparently well known to the Xen/Xensource guys judging by the number of
reports/bugs posted, but I haven't seen mention of it actually being addressed
and resolved. Unfortunately I see the same issue with it cropping up after VM
moves, though it occurs /every/ time there is a VM migration, once per
processor in the VM; doesn't matter if there is any IO on the Dom0 or DomU.
Occasionally VM's die during a migration and have to be manually
destroyed/restarted.
I do see evidence of significant instability (not implying it is related to the
above softlockup issues) however, in either VM moves migrating from a Xeon
(5345) to Opteron Dom0, and in high-utilization DomU's which are just plain
flaky and reboot/die semi-frequently even when never altered from their start Dom0.
For me, it currently means running only low-priority non-production services in
a VM, and not shelling out for RHEL5 support for the project (contrary to what
I planned) since it's not being addressed. I'd be curious if this is being
addressed in the Xen 3.2 release for RHEL5*...
Cheers,
/eli
Brett Worth wrote:
Hello All.
I've just started looking into Xen and have a test environment in
place. I'm seeing an
annoying problem that I thought worthy of a post.
Config:
I have 2 x HP DL585 servers each with 4 Dual core Opterons (non-vmx) and
16GB RAM
configured as Xen servers. These run CentOS 5.1 with the latest updates
applied. These
system both attach to an iSCSI target which is an HP DL385 running ietd
and serving SAN
based storage.
I have a test VM running CentOS 5.1 also updated.
Problem:
If I run the VM on a single server everything is OK. If I do a migrate
of the VM to the
other server I start getting random "BUG: soft lockup detected on CPU#?"
messages on the
VM console. The messages seem to happen with IO but not every time. A
reboot of the VM
on the new server will stop these messages.
I've also left the VM running overnight a couple of times and when I do
I find that any
external sessions (ssh) are hung in the morning but the console session
is not. New ssh
sessions can be started and seem to work.
After much googling it looks like the kernel messages can occur if dom0
is very busy but
mine is not.
Any suggestions?
Regards
Brett Worth
_______________________________________________
CentOS-virt mailing list
CentOS-virt@centos.org
http://lists.centos.org/mailman/listinfo/centos-virt
_______________________________________________
CentOS-virt mailing list
CentOS-virt@centos.org
http://lists.centos.org/mailman/listinfo/centos-virt