I accidentally first sent this reply from a different email address, and am 
resending with my correct email address.  I apologize if the reply shows up 
twice in the list.


Hello,

Thanks for this information.  I think I may have stumbled on a solution 
(stumble really is the operative word here).  To follow up on the responses 
though:

I've tried both with passthrough and without it, and the results were the same.

In my searching for a solution I found out about centos-release-qemu-ev and 
installed qemu-ev.  It didn't seem to change this behaviour, but it did appear 
to reduce the host cpu with idle guests (although not as much as I'd hoped).

I used the latest virtio.iso from Fedora to install what drivers I could.  Some 
drivers wouldn't install though (eg. network and graphics)... I got  messages 
about version incompatibility, or configuration errors.

I decided to do a completely new install (O/S and CloudStack) on a different, much older machine, 
and found that I had different behaviour.  The VM's ran without any problems.  Aside from the 
hardware differences, I also used just the default "Medium" Compute Offering for the VM's 
(which I think is 1GB RAM and 1GHz CPU).  And this is where I stumbled on the solution.  On the 
problematic system, I had created a Compute Offering that I thought was suited to the hardware 
configuration and the number of concurrent VM's that would run on it.  So I switched the VM's to 
the default "Medium" Compute Offering on this new computer.  And the VM's didn't freeze.  
I experimented with different Compute Offerings and found that this was what made the difference.  
In this case, it looks like the Compute Offering's CPU speed has to be set somewhat lower than the 
actual CPU speed.  Adding cores to the Compute Offering with the lower CPU speed also worked.

I think this is just my lack of knowledge... I seemed to recall that QEMU 
wouldn't start a VM with an incompatible configuration.  But that may not be 
the case, as this experience indicates.  It's possible too that the 
configuration wasn't incompatible but that it happened to be something that 
caused Windows to not play nice because I had no issues with an Ubuntu Desktop 
VM using the same Compute Offering.

Thanks for your help,
Dave

On 08-09-20 6:04 p.m., Eric Lee Green wrote:
I will note that we have several Windows 10 / Windows 2018 genre VM's on Centos 
7.8.2003 and are not seeing any guest freezes. But:

1) We are not using CPU host-passthrough.

2) We do have the QEMU Windows guest drivers installed in the VM's.

3) I am running Cloudstack 4.11.3. I tried upgrading to Cloudstack 4.14.0 and 
the upgrade failed, the agent was unable to ssh into virtual routers to 
configure them.

That said, the version of Cloudstack should not matter, since all Cloudstack 
does is tell libvirtd to launch the VM. libvirtd handles everything from there. 
We are literally using the stock Centos 7.8 libvirtd.conf as modified by the 
agent configurator.  I wonder if you are having a hardware problem on your new 
server, or if there is a Linux OS problem handling your new server. Use 
journalctl and see if you're seeing anything at the Linux OS level when a 
freeze happens, and check the qemu log files for the instance also, because 
normally libvirtd is quite reliable and Centos 7.8.2003 supports Windows 10 
VM's just fine.

On 8/9/2020 8:48 AM, Rohit Yadav wrote:
Hi Dave,

I've not specifically worked with Windows guest VMs on KVM, but what you've 
described is largely subject to a guest OS support by the hypervisor, in this 
case the specific libvirt/qemu/linux-kernel version and any guest tools/drivers 
that must be installed inside the Windows guest VM. You've yourself 
acknowledged that the issue is not seen in your older CentOS6 based environment.

To rule out CloudStack (a) you may add or upgrade hosts in a cluster to use 
qemu-ev (enterprise qemu release by CentOS SIG 
https://wiki.centos.org/SpecialInterestGroup/Virtualization, i.e. install the 
centos-release-qemu-ev pkg) or (b) you may add a new cluster with Ubuntu 18.04 
KVM hosts and recreate your Windows VM setup. Or, it could be a specific 
Windows build or requires additional drivers (such as 
https://www.linux-kvm.org/page/WindowsGuestDrivers/Download_Drivers).


Hope this helps.


Regards.

________________________________
From: Dave Lapointe <[email protected]>
Sent: Saturday, August 8, 2020 00:48
To: [email protected] <[email protected]>
Subject: Windows 10 KVM becoming inaccessible

Hi,

I've been lurking on this group for a few years now and have found the 
information that people have posted here to be quite helpful for me to 
understand CloudStack better.  I've never replied here because there are always 
far more knowledgeable people on this list who can offer much better insight 
than I ever could.

An issue has arisen recently that I can't find any solution for.  I apologize 
ahead of time if this is the wrong list to post to.

I recently configured a new server to run CloudStack using Centos 7.8.2003 and CloudStack 
4.14, and configured some Windows 10 LTSB KVM guests.  This is a fairly specialized 
server, so the configuration is a little unusual.  It's configured to use the 
"cloudbr0" software bridge for the guest network which is then routed 
externally through a single NIC.  Also, because the VM's will never be migrated, I've set 
guest.cpu.mode=host-passthrough.

What's been happening though is that the VM's will just freeze sometimes, 
apparently randomly.  Sometimes it will happen during boot, or a couple minutes 
after connecting by RDP.  And sometimes the VM won't freeze at all.  I haven't 
been able to determine a pattern as to when this will happen.  And I haven't 
found anything in the logs that might help me understand what's happening 
(/var/log/messages and /var/log/cloudstack/management/management-server.log).  
I've checked on the QEMU and Linux forums, but have only found a bit of 
information about VM's freezing for people using specific graphics drivers with 
passthrough for their graphics cards.  I tried removing 
guest.cpu.mode=host-passthrough but that made no difference.

What's especially odd to me is that this didn't happen with older systems I've 
created (eg. CentOS 6 and CloudStack 4.9).  I've setup half a dozen or so using 
the same configuration as this system, just older software.

I can't tell if this is related to CloudStack (maybe there is something in the 
guest parameters that is causing this), or if this is strictly a KVM issue.  
And since I can't find anything in the logs I don't know where else to look.  
I'm hoping to get some suggestions from this list so that I can do some more 
digging.

Thanks,
Dave

[email protected]
www.shapeblue.com
3 London Bridge Street,  3rd floor, News Building, London  SE1 9SGUK
@shapeblue

Reply via email to