Hello

If on the VM on top of ESXI full VT / Virtualization is enabled, then for
the VM should be transparent create a nested setup.

So I would expend bit more time on the VM in ESXi and see if you are
missing some flag.

Alvaro


On Fri, Oct 6, 2017 at 3:00 PM, Mihai <skyra...@gmail.com> wrote:

> Hi there,
>
> I'm hopefully in the right place with my question and if yes, then it's
> going to be quite a headache one :)
> I have a server (HP DL380G8p)  with 2 x 12 core CPUs (Intel E5-2695) + 768
> GB Ram.
> On top of it I run an ESX 6.5 and then my lab experiments.
> So: ESX --> Ubuntu/Debian --> vagrant libvirt --> Cumulus VX topology/VMs
> / or just a simple Ubuntu vagrant box
> Recently I wanted to give Vagrant a try so:
> - I installed an Ubuntu 16.04.3 LTS, then an 17.04 and also a Debian 9.1
> (trying to see if kernel version, other distro packages are having an
> influence on this)
> - From my notes so far:
>    - libvirt -3.0
>    - vagrant 2.0
>    - kernel 4.90.3-amd64
> - Virtual machine with Linux/Vagrant/Qemu with 64GB Ram reserved for (no
> memory ballooning, etc)
>  - I cloned these repositories:  https://github.
> com/CumulusNetworks/topology_converter.git
>    and https://github.com/CumulusNetworks/topology_converter.git
> - They start with vagrant Linux VMs + I also tried with a simple Vagrant
> Ubuntu box (no special settings)
> - All the time after "vagrant up", if I do a "virsh console <id>", I get:
> Vagrant:
> ==> leaf-r01n01: Creating shared folders metadata...
> ==> leaf-r01n01: Starting domain.
> ==> leaf-r01n01: Waiting for domain to get an IP address...
>
> Virsh:
> Connected to domain topology_converter_leaf-r01n01
> Escape character is ^]
> [   74.380034] INFO: rcu_sched detected stalls on CPUs/tasks:
> [   74.380034] (detected by 0, t=60002 jiffies, g=3030, c=3029, q=23)
> [   74.380034] All QSes seen, last rcu_sched kthread activity 60002
> (4294741676-4294681674), jiffies_till_next_fqs=3, root ->qsmask 0x0
> [   74.380034] rcu_sched kthread starved for 60002 jiffies!
> [  100.132032] NMI watchdog: BUG: soft lockup - CPU#0 stuck for 23s!
> [dhclient:2179]
> [  128.132031] NMI watchdog: BUG: soft lockup - CPU#0 stuck for 23s!
> [dhclient:2179]
> [  156.132037] NMI watchdog: BUG: soft lockup - CPU#0 stuck for 22s!
> [dhclient:2179]
>
> I suspected this to be just another qemu nested setup problem BUT I do
> have another 2 VMs where I play with qemu and all works fine.
> This might mean that it is related to the parameters that vagrant ->
> libvirt -> uses to start qemu/kvm machines.
> If it is, then the question is what exactly is causing this behavior?
>
> On top of this, I tried also:
> - to disable acpi on boot
> - reserve also CPU resources on VMWare
> - I checked to be sure that VMWare exported the hardware virtualization
> flags (for nested virtualization)
> flags  : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov
> pat pse36 clflush dts mmx fxsr sse sse2 ss ht syscall nx rdtscp lm
> constant_tsc arch_perfmon pebs bts nopl xtopology tsc_reliable nonstop_tsc
> aperfmperf pni pclmulqdq vmx ssse3 cx16 pcid sse4_1 sse4_2 x2apic popcnt
> tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm epb
> tpr_shadow vnmi ept vpid fsgsbase tsc_adjust smep dtherm ida arat pln pts
>
> Vagrant starts the qemu VM with these parameters:
> /usr/bin/qemu-system-x86_64 -name 
> guest=topology_converter_leaf-r01n01,debug-threads=on
> -S -object secret,id=masterKey0,format=raw,file=/var/lib/libvirt/
> qemu/domain-3-topology_converter_l/master-key.aes -machine
> pc-i440fx-zesty,accel=kvm,usb=off,dump-guest-core=off -cpu
> IvyBridge,+ds,+ss,+ht,+vmx,+pcid,+osxsave,+hypervisor,+arat,+tsc_adjust
> -m 512 -realtime mlock=off -smp 1,sockets=1,cores=1,threads=1 -uuid
> 7e4de018-2e63-4686-a110-4cec6d587758 -no-user-config -nodefaults -chardev
> socket,id=charmonitor,path=/var/lib/libvirt/qemu/domain-3-
> topology_converter_l/monitor.sock,server,nowait -mon
> chardev=charmonitor,id=monitor,mode=control -rtc base=utc -no-shutdown
> -boot strict=on -device piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2
> -device ahci,id=sata0,bus=pci.0,addr=0x3 -drive
> file=/var/lib/libvirt/images/topology_converter_leaf-
> r01n01.img,format=qcow2,if=none,id=drive-sata0-0-0 -device
> ide-hd,bus=sata0.0,drive=drive-sata0-0-0,id=sata0-0-0,bootindex=1 -netdev
> tap,fd=26,id=hostnet0,vhost=on,vhostfd=28 -device virtio-net-pci,netdev=
> hostnet0,id=net0,mac=52:54:00:86:cd:fe,bus=pci.0,addr=0x5 -netdev
> socket,udp=127.0.0.1:9008,localaddr=127.0.0.1:8008,id=hostnet1 -device
> virtio-net-pci,netdev=hostnet1,id=net1,mac=44:38:39:00:00:0d,bus=pci.0,addr=0x6
> -netdev socket,udp=127.0.0.1:8002,localaddr=127.0.0.1:9002,id=hostnet2
> -device 
> virtio-net-pci,netdev=hostnet2,id=net2,mac=44:38:39:00:00:04,bus=pci.0,addr=0x7
> -netdev socket,udp=127.0.0.1:8005,localaddr=127.0.0.1:9005,id=hostnet3
> -device 
> virtio-net-pci,netdev=hostnet3,id=net3,mac=44:38:39:00:00:0a,bus=pci.0,addr=0x8
> -netdev socket,udp=127.0.0.1:8004,localaddr=127.0.0.1:9004,id=hostnet4
> -device 
> virtio-net-pci,netdev=hostnet4,id=net4,mac=44:38:39:00:00:08,bus=pci.0,addr=0x9
> -chardev pty,id=charserial0 -device isa-serial,chardev=charserial0,id=serial0
> -vnc 127.0.0.1:0 -k en-us -device cirrus-vga,id=video0,bus=pci.0,addr=0x2
> -device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x4 -msg
> timestamp=on
>
> Now I ran out of ideas and this is driving me crazy already.
> Does anyone have any clue as to what else I should try?
> I'm afraid I reached a deadend.
>
> Thanks in advance and hoping for a life saving solution to get me unstuck
> :)
>
>
>
> --
> This mailing list is governed under the HashiCorp Community Guidelines -
> https://www.hashicorp.com/community-guidelines.html. Behavior in
> violation of those guidelines may result in your removal from this mailing
> list.
>
> GitHub Issues: https://github.com/mitchellh/vagrant/issues
> IRC: #vagrant on Freenode
> ---
> You received this message because you are subscribed to the Google Groups
> "Vagrant" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to vagrant-up+unsubscr...@googlegroups.com.
> To view this discussion on the web visit https://groups.google.com/d/
> msgid/vagrant-up/91da599d-0804-4e48-9b79-01545c33ea0b%40googlegroups.com
> <https://groups.google.com/d/msgid/vagrant-up/91da599d-0804-4e48-9b79-01545c33ea0b%40googlegroups.com?utm_medium=email&utm_source=footer>
> .
> For more options, visit https://groups.google.com/d/optout.
>



-- 
Alvaro

-- 
This mailing list is governed under the HashiCorp Community Guidelines - 
https://www.hashicorp.com/community-guidelines.html. Behavior in violation of 
those guidelines may result in your removal from this mailing list.

GitHub Issues: https://github.com/mitchellh/vagrant/issues
IRC: #vagrant on Freenode
--- 
You received this message because you are subscribed to the Google Groups 
"Vagrant" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to vagrant-up+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/vagrant-up/CAHqq0eybxHwgqACMe%2BFQSwaUkB9njZRsJ1C8FVBg8%2BQWBARrxA%40mail.gmail.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to