On Friday 19 September 2008 12:41:46 pm you wrote:
> Are you using filesystem backed storage for the guest images or direct
> block device storage? I assume there's heavy write activity on the
> guests when these hangs happen?

Yes, they happen when one VM is doing heavy writes.  I'm actually using a 
whole stack of things:

OCFS2 on DRBD (Primary-Primary) on LVM Volume (continuous) on LUKS-encrypted 
partition.  Fun debugging that, heh?

In trying to figure out the problem, I tried to reconfigure DRBD to use 
Protocol B instead of C.  However, it failed to make the switch and both 
nodes disconnected so now I have a split-brain.  In try to fix the split 
brain I'm taking down on one node all the VM's one by one, copying the VM 
drives from one node to the next, and starting up on the other node (old-
fashioned migration).  Yes, I'm having *lots* of fun!  Perfect way to end 
the week!

So, any ideas on how to solve the bottleneck?  Isn't the CFQ scheduler 
supposed to grant every processes the same amount of I/O?  Is there a way to 
change something in proc to avoid this situation?

-- 
Alberto Treviño
BYU Testing Center
Brigham Young University

--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to