Happy to say that I just passed this "Get local VM IP" step There were a lot of leftover from previous failed attempts (cf log I sent earlier : "internal error: Failed to autostart storage pool..." ) Those were not cleaned up by ovirt-hosted-engine-cleanup
I had to do the followinf so libvirt would be happy again : rm -rf /etc/libvirt/storage/*.xml rm -rf /etc/libvirt/storage/autostart/* rm -rf /var/tmp/local* ovirt-hosted-engine-cleanup is not doing a really good job Guillaume Pavese Ingénieur Système et Réseau Interactiv-Group On Tue, Feb 26, 2019 at 3:49 AM Guillaume Pavese < guillaume.pav...@interactiv-group.com> wrote: > journalctl -u libvirtd.service : > > févr. 25 18:47:24 vs-inf-int-kvm-fr-301-210.hostics.fr systemd[1]: > Stopping Virtualization daemon... > févr. 25 18:47:24 vs-inf-int-kvm-fr-301-210.hostics.fr systemd[1]: > Stopped Virtualization daemon. > févr. 25 18:47:34 vs-inf-int-kvm-fr-301-210.hostics.fr systemd[1]: > Starting Virtualization daemon... > févr. 25 18:47:34 vs-inf-int-kvm-fr-301-210.hostics.fr systemd[1]: > Started Virtualization daemon. > févr. 25 18:47:34 vs-inf-int-kvm-fr-301-210.hostics.fr dnsmasq[6310]: > read /etc/hosts - 4 addresses > févr. 25 18:47:34 vs-inf-int-kvm-fr-301-210.hostics.fr dnsmasq[6310]: > read /var/lib/libvirt/dnsmasq/default.addnhosts - 0 addresses > févr. 25 18:47:34 vs-inf-int-kvm-fr-301-210.hostics.fr > dnsmasq-dhcp[6310]: read /var/lib/libvirt/dnsmasq/default.hostsfile > févr. 25 18:47:34 vs-inf-int-kvm-fr-301-210.hostics.fr libvirtd[13535]: > 2019-02-25 17:47:34.739+0000: 13551: info : libvirt version: 4.5.0, > package: 10.el7_6.4 (CentOS BuildSystem <http://bugs.centos.org>, > 2019-01-29-17:31:22, x86-01.bsys.centos.org) > févr. 25 18:47:34 vs-inf-int-kvm-fr-301-210.hostics.fr libvirtd[13535]: > 2019-02-25 17:47:34.739+0000: 13551: info : hostname: > vs-inf-int-kvm-fr-301-210.hostics.fr > févr. 25 18:47:34 vs-inf-int-kvm-fr-301-210.hostics.fr libvirtd[13535]: > 2019-02-25 17:47:34.739+0000: 13551: error : virDirOpenInternal:2936 : > cannot open directory > '/var/tmp/localvmgmyYik/images/15023c8a-e3a7-4851-a97d-3b90996b423b': No > such file or directory > févr. 25 18:47:34 vs-inf-int-kvm-fr-301-210.hostics.fr libvirtd[13535]: > 2019-02-25 17:47:34.740+0000: 13551: error : > storageDriverAutostartCallback:209 : internal error: Failed to autostart > storage pool '15023c8a-e3a7-4851-a97d-3b90996b423b': cannot open directory > '/var/tmp/localvmgmyYik/images/15023c8a-e3a7-4851-a97d-3b90996b423b': No > such file or directory > févr. 25 18:47:34 vs-inf-int-kvm-fr-301-210.hostics.fr libvirtd[13535]: > 2019-02-25 17:47:34.740+0000: 13551: error : virDirOpenInternal:2936 : > cannot open directory '/var/tmp/localvmdRIozH': No such file or directory > févr. 25 18:47:34 vs-inf-int-kvm-fr-301-210.hostics.fr libvirtd[13535]: > 2019-02-25 17:47:34.740+0000: 13551: error : > storageDriverAutostartCallback:209 : internal error: Failed to autostart > storage pool 'localvmdRIozH': cannot open directory > '/var/tmp/localvmdRIozH': No such file or directory > févr. 25 18:47:34 vs-inf-int-kvm-fr-301-210.hostics.fr libvirtd[13535]: > 2019-02-25 17:47:34.740+0000: 13551: error : virDirOpenInternal:2936 : > cannot open directory > '/var/tmp/localvmdRIozH/images/15023c8a-e3a7-4851-a97d-3b90996b423b': No > such file or directory > févr. 25 18:47:34 vs-inf-int-kvm-fr-301-210.hostics.fr libvirtd[13535]: > 2019-02-25 17:47:34.740+0000: 13551: error : > storageDriverAutostartCallback:209 : internal error: Failed to autostart > storage pool '15023c8a-e3a7-4851-a97d-3b90996b423b-1': cannot open > directory > '/var/tmp/localvmdRIozH/images/15023c8a-e3a7-4851-a97d-3b90996b423b': No > such file or directory > févr. 25 18:47:34 vs-inf-int-kvm-fr-301-210.hostics.fr libvirtd[13535]: > 2019-02-25 17:47:34.740+0000: 13551: error : virDirOpenInternal:2936 : > cannot open directory '/var/tmp/localvmgmyYik': No such file or directory > févr. 25 18:47:34 vs-inf-int-kvm-fr-301-210.hostics.fr libvirtd[13535]: > 2019-02-25 17:47:34.740+0000: 13551: error : > storageDriverAutostartCallback:209 : internal error: Failed to autostart > storage pool 'localvmgmyYik': cannot open directory > '/var/tmp/localvmgmyYik': No such file or directory > > > /var/log/libvirt/qemu/HostedEngineLocal.log : > > 2019-02-25 17:50:08.694+0000: starting up libvirt version: 4.5.0, package: > 10.el7_6.4 (CentOS BuildSystem <http://bugs.centos.org>, > 2019-01-29-17:31:22, x86-01.bsys.centos.org), qemu version: > 2.12.0qemu-kvm-ev-2.12.0-18.el7_6.3.1, kernel: 3.10.0-957.5.1.el7.x86_64, > hostname: vs-inf-int-kvm-fr-301-210.hostics.fr > LC_ALL=C PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin > QEMU_AUDIO_DRV=none /usr/libexec/qemu-kvm -name > guest=HostedEngineLocal,debug-threads=on -S -object > secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/domain-1-HostedEngineLocal/master-key.aes > -machine pc-i440fx-rhel7.6.0,accel=kvm,usb=off,dump-guest-core=off -cpu > Haswell-noTSX,+kvmclock -m 16384 -realtime mlock=off -smp > 4,sockets=4,cores=1,threads=1 -uuid 8ba608c8-b721-4b5b-b839-b62f5e919814 > -no-user-config -nodefaults -chardev > socket,id=charmonitor,fd=27,server,nowait -mon > chardev=charmonitor,id=monitor,mode=control -rtc base=utc -no-shutdown > -global PIIX4_PM.disable_s3=1 -global PIIX4_PM.disable_s4=1 -boot > menu=off,strict=on -device > virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -drive > file=/var/tmp/localvmlF5yTM/images/15023c8a-e3a7-4851-a97d-3b90996b423b/07fdcff3-11ce-4f7c-af05-0a878593e78e,format=qcow2,if=none,id=drive-virtio-disk0 > -device > virtio-blk-pci,scsi=off,bus=pci.0,addr=0x5,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1 > -drive > file=/var/tmp/localvmlF5yTM/seed.iso,format=raw,if=none,id=drive-ide0-0-0,readonly=on > -device ide-cd,bus=ide.0,unit=0,drive=drive-ide0-0-0,id=ide0-0-0 -netdev > tap,fd=29,id=hostnet0,vhost=on,vhostfd=30 -device > virtio-net-pci,netdev=hostnet0,id=net0,mac=00:16:3e:1d:4b:b6,bus=pci.0,addr=0x3 > -chardev pty,id=charserial0 -device > isa-serial,chardev=charserial0,id=serial0 -chardev > socket,id=charchannel0,fd=31,server,nowait -device > virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=org.qemu.guest_agent.0 > -vnc 127.0.0.1:0 -device VGA,id=video0,vgamem_mb=16,bus=pci.0,addr=0x2 > -object rng-random,id=objrng0,filename=/dev/random -device > virtio-rng-pci,rng=objrng0,id=rng0,bus=pci.0,addr=0x6 -sandbox > on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny > -msg timestamp=on > 2019-02-25T17:50:08.904663Z qemu-kvm: -chardev pty,id=charserial0: char > device redirected to /dev/pts/4 (label charserial0) > 2019-02-25T17:50:08.911239Z qemu-kvm: warning: host doesn't support > requested feature: CPUID.07H:EBX.invpcid [bit 10] > 2019-02-25T17:50:08.917723Z qemu-kvm: warning: host doesn't support > requested feature: CPUID.07H:EBX.invpcid [bit 10] > 2019-02-25T17:50:08.918494Z qemu-kvm: warning: host doesn't support > requested feature: CPUID.07H:EBX.invpcid [bit 10] > 2019-02-25T17:50:08.919217Z qemu-kvm: warning: host doesn't support > requested feature: CPUID.07H:EBX.invpcid [bit 10] > > I guess there is something about those last warnings? > It should be noted that I previously successfully deployed oVirt 4.2 in > the same Nested environment > > Running libvirt in debug mode will need to wait tomorrow ; my night is > already cut to nothing much anymore XD > > Guillaume Pavese > Ingénieur Système et Réseau > Interactiv-Group > > > On Tue, Feb 26, 2019 at 3:33 AM Simone Tiraboschi <stira...@redhat.com> > wrote: > >> >> >> On Mon, Feb 25, 2019 at 7:15 PM Guillaume Pavese < >> guillaume.pav...@interactiv-group.com> wrote: >> >>> No, as indicated previously, still : >>> >>> [root@vs-inf-int-kvm-fr-301-210 ~]# virsh -r net-dhcp-leases default >>> Expiry Time MAC address Protocol IP address >>> Hostname Client ID or DUID >>> >>> ------------------------------------------------------------------------------------------------------------------- >>> >>> [root@vs-inf-int-kvm-fr-301-210 ~]# >>> >>> >>> I did not see any relevant log on the HE vm. Is there something I should >>> look for there? >>> >> >> This smells really bad: I'd suggest to check /var/log/messages >> and /var/log/libvirt/qemu/HostedEngineLocal.log for libvirt errors; >> if nothing is there can I ask you to try reexecuting with libvirt debug >> logs (edit /etc/libvirt/libvirtd.conf). >> >> Honestly I'm not able to reproduce it on my side. >> >> >>> >>> >>> Guillaume Pavese >>> Ingénieur Système et Réseau >>> Interactiv-Group >>> >>> >>> On Tue, Feb 26, 2019 at 3:12 AM Simone Tiraboschi <stira...@redhat.com> >>> wrote: >>> >>>> >>>> >>>> On Mon, Feb 25, 2019 at 7:04 PM Guillaume Pavese < >>>> guillaume.pav...@interactiv-group.com> wrote: >>>> >>>>> I still can't connect with VNC remotely but locally with X forwarding >>>>> it works. >>>>> However my connection has too high latency for that to be usable (I'm >>>>> in Japan, my hosts in France, ~250 ms ping) >>>>> >>>>> But I could see that the VM is booted! >>>>> >>>>> and in Hosts logs there is : >>>>> >>>>> févr. 25 18:51:12 vs-inf-int-kvm-fr-301-210.hostics.fr python[14719]: >>>>> ansible-command Invoked with warn=True executable=None _uses_shell=True >>>>> _raw_params=virsh -r net-dhcp-leases default | grep -i 00:16:3e:1d:4b:b6 | >>>>> awk '{ print $5 }' | cut -f1 -d'/' removes=None argv=None creates=None >>>>> chdir=None stdin=None >>>>> févr. 25 18:51:30 vs-inf-int-kvm-fr-301-210.hostics.fr >>>>> dnsmasq-dhcp[6310]: DHCPDISCOVER(virbr0) 00:16:3e:1d:4b:b6 >>>>> févr. 25 18:51:30 vs-inf-int-kvm-fr-301-210.hostics.fr >>>>> dnsmasq-dhcp[6310]: DHCPOFFER(virbr0) 192.168.122.14 00:16:3e:1d:4b:b6 >>>>> févr. 25 18:51:30 vs-inf-int-kvm-fr-301-210.hostics.fr >>>>> dnsmasq-dhcp[6310]: DHCPREQUEST(virbr0) 192.168.122.14 00:16:3e:1d:4b:b6 >>>>> févr. 25 18:51:30 vs-inf-int-kvm-fr-301-210.hostics.fr >>>>> dnsmasq-dhcp[6310]: DHCPACK(virbr0) 192.168.122.14 00:16:3e:1d:4b:b6 >>>>> vs-inf-int-ovt-fr-301-210 >>>>> févr. 25 18:51:42 vs-inf-int-kvm-fr-301-210.hostics.fr python[14757]: >>>>> ansible-command Invoked with warn=True executable=None _uses_shell=True >>>>> _raw_params=virsh -r net-dhcp-leases default | grep -i 00:16:3e:1d:4b:b6 | >>>>> awk '{ print $5 }' | cut -f1 -d'/' removes=None argv=None creates=None >>>>> chdir=None stdin=None >>>>> févr. 25 18:52:12 vs-inf-int-kvm-fr-301-210.hostics.fr python[14789]: >>>>> ansible-command Invoked with warn=True executable=None _uses_shell=True >>>>> _raw_params=virsh -r net-dhcp-leases default | grep -i 00:16:3e:1d:4b:b6 | >>>>> awk '{ print $5 }' | cut -f1 -d'/' removes=None argv=None creates=None >>>>> chdir=None stdin=None >>>>> févr. 25 18:52:43 vs-inf-int-kvm-fr-301-210.hostics.fr python[14818]: >>>>> ansible-command Invoked with warn=True executable=None _uses_shell=True >>>>> _raw_params=virsh -r net-dhcp-leases default | grep -i 00:16:3e:1d:4b:b6 | >>>>> awk '{ print $5 }' | cut -f1 -d'/' removes=None argv=None creates=None >>>>> chdir=None stdin=None >>>>> .... >>>>> >>>>> ssh to the vm works too : >>>>> >>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# ssh root@192.168.122.14 >>>>> The authenticity of host '192.168.122.14 (192.168.122.14)' can't be >>>>> established. >>>>> ECDSA key fingerprint is >>>>> SHA256:+/pUzTGVA4kCyICb7XgqrxWYYkqzmDjVmdAahiBFgOQ. >>>>> ECDSA key fingerprint is >>>>> MD5:4b:ef:ff:4a:7c:1a:af:c2:af:4a:0f:14:a3:c5:31:fb. >>>>> Are you sure you want to continue connecting (yes/no)? yes >>>>> Warning: Permanently added '192.168.122.14' (ECDSA) to the list of >>>>> known hosts. >>>>> root@192.168.122.14's password: >>>>> [root@vs-inf-int-ovt-fr-301-210 ~]# >>>>> >>>>> >>>>> But the test that the playbook tries still fails with empty result : >>>>> >>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# virsh -r net-dhcp-leases default >>>>> Expiry Time MAC address Protocol IP address >>>>> Hostname Client ID or DUID >>>>> >>>>> ------------------------------------------------------------------------------------------------------------------- >>>>> >>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# >>>>> >>>>> >>>> This smells like a bug to me: >>>> and nothing at all in the output of >>>> virsh -r net-dhcp-leases default >>>> >>>> ? >>>> >>>> >>>>> >>>>> >>>>> >>>>> Guillaume Pavese >>>>> Ingénieur Système et Réseau >>>>> Interactiv-Group >>>>> >>>>> >>>>> On Tue, Feb 26, 2019 at 1:54 AM Simone Tiraboschi <stira...@redhat.com> >>>>> wrote: >>>>> >>>>>> >>>>>> >>>>>> On Mon, Feb 25, 2019 at 5:50 PM Guillaume Pavese < >>>>>> guillaume.pav...@interactiv-group.com> wrote: >>>>>> >>>>>>> I did that but no success yet. >>>>>>> >>>>>>> I see that "Get local VM IP" task tries the following : >>>>>>> >>>>>>> virsh -r net-dhcp-leases default | grep -i {{ he_vm_mac_addr }} | >>>>>>> awk '{ print $5 }' | cut -f1 -d'/' >>>>>>> >>>>>>> >>>>>>> However while the task is running, and vm running in qemu, "virsh -r >>>>>>> net-dhcp-leases default" never returns anything : >>>>>>> >>>>>> >>>>>> Yes, I think that libvirt will never provide a DHCP lease since the >>>>>> appliance OS never correctly complete the boot process. >>>>>> I'd suggest to connect to the running VM via vnc DURING the boot >>>>>> process and check what's wrong. >>>>>> >>>>>> >>>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# virsh -r net-dhcp-leases default >>>>>>> Expiry Time MAC address Protocol IP address >>>>>>> Hostname Client ID or DUID >>>>>>> >>>>>>> ------------------------------------------------------------------------------------------------------------------- >>>>>>> >>>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> Guillaume Pavese >>>>>>> Ingénieur Système et Réseau >>>>>>> Interactiv-Group >>>>>>> >>>>>>> >>>>>>> On Tue, Feb 26, 2019 at 12:33 AM Simone Tiraboschi < >>>>>>> stira...@redhat.com> wrote: >>>>>>> >>>>>>>> OK, try this: >>>>>>>> temporary >>>>>>>> edit >>>>>>>> /usr/share/ansible/roles/ovirt.hosted_engine_setup/tasks/bootstrap_local_vm/02_create_local_vm.yml >>>>>>>> around line 120 >>>>>>>> and edit tasks "Get local VM IP" >>>>>>>> changing from "retries: 50" to "retries: 500" so that you have >>>>>>>> more time to debug it >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Mon, Feb 25, 2019 at 4:20 PM Guillaume Pavese < >>>>>>>> guillaume.pav...@interactiv-group.com> wrote: >>>>>>>> >>>>>>>>> I retried after killing the remaining qemu process and >>>>>>>>> doing ovirt-hosted-engine-cleanup >>>>>>>>> The new attempt failed again at the same step. Then after it >>>>>>>>> fails, it cleans the temporary files (and vm disk) but *qemu >>>>>>>>> still runs!* : >>>>>>>>> >>>>>>>>> [ INFO ] TASK [ovirt.hosted_engine_setup : Get local VM IP] >>>>>>>>> >>>>>>>>> [ ERROR ] fatal: [localhost]: FAILED! => {"attempts": 50, >>>>>>>>> "changed": true, "cmd": "virsh -r net-dhcp-leases default | grep -i >>>>>>>>> 00:16:3e:6c:e8:f9 | awk '{ print $5 }' | cut -f1 -d'/'", "delta": >>>>>>>>> "0:00:00.092436", "end": "2019-02-25 16:09:38.863263", "rc": 0, >>>>>>>>> "start": >>>>>>>>> "2019-02-25 16:09:38.770827", "stderr": "", "stderr_lines": [], >>>>>>>>> "stdout": >>>>>>>>> "", "stdout_lines": []} >>>>>>>>> [ INFO ] TASK [ovirt.hosted_engine_setup : include_tasks] >>>>>>>>> [ INFO ] ok: [localhost] >>>>>>>>> [ INFO ] TASK [ovirt.hosted_engine_setup : Remove local vm dir] >>>>>>>>> [ INFO ] changed: [localhost] >>>>>>>>> [ INFO ] TASK [ovirt.hosted_engine_setup : Remove temporary entry >>>>>>>>> in /etc/hosts for the local VM] >>>>>>>>> [ INFO ] ok: [localhost] >>>>>>>>> [ INFO ] TASK [ovirt.hosted_engine_setup : Notify the user about >>>>>>>>> a failure] >>>>>>>>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": >>>>>>>>> "The system may not be provisioned according to the playbook results: >>>>>>>>> please check the logs for the issue, fix accordingly or re-deploy from >>>>>>>>> scratch.\n"} >>>>>>>>> [ ERROR ] Failed to execute stage 'Closing up': Failed executing >>>>>>>>> ansible-playbook >>>>>>>>> [ INFO ] Stage: Clean up >>>>>>>>> [ INFO ] Cleaning temporary resources >>>>>>>>> ... >>>>>>>>> >>>>>>>>> [ INFO ] TASK [ovirt.hosted_engine_setup : Remove local vm dir] >>>>>>>>> [ INFO ] ok: [localhost] >>>>>>>>> [ INFO ] TASK [ovirt.hosted_engine_setup : Remove temporary entry >>>>>>>>> in /etc/hosts for the local VM] >>>>>>>>> [ INFO ] ok: [localhost] >>>>>>>>> [ INFO ] Generating answer file >>>>>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20190225161011.conf' >>>>>>>>> [ INFO ] Stage: Pre-termination >>>>>>>>> [ INFO ] Stage: Termination >>>>>>>>> [ ERROR ] Hosted Engine deployment failed: please check the logs >>>>>>>>> for the issue, fix accordingly or re-deploy from scratch. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# ps aux | grep qemu >>>>>>>>> root 4021 0.0 0.0 24844 1788 ? Ss févr.22 0:00 >>>>>>>>> /usr/bin/qemu-ga --method=virtio-serial >>>>>>>>> --path=/dev/virtio-ports/org.qemu.guest_agent.0 >>>>>>>>> --blacklist=guest-file-open,guest-file-close,guest-file-read,guest-file-write,guest-file-seek,guest-file-flush,guest-exec,guest-exec-status >>>>>>>>> -F/etc/qemu-ga/fsfreeze-hook >>>>>>>>> qemu 26463 22.9 4.8 17684512 1088844 ? Sl 16:01 3:09 >>>>>>>>> /usr/libexec/qemu-kvm -name guest=HostedEngineLocal,debug-threads=on >>>>>>>>> -S >>>>>>>>> -object >>>>>>>>> secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/domain-1-HostedEngineLocal/master-key.aes >>>>>>>>> -machine pc-i440fx-rhel7.6.0,accel=kvm,usb=off,dump-guest-core=off >>>>>>>>> -cpu >>>>>>>>> Haswell-noTSX,+kvmclock -m 16384 -realtime mlock=off -smp >>>>>>>>> 4,sockets=4,cores=1,threads=1 -uuid >>>>>>>>> 316eca5f-81de-4a0b-af1f-58f910402a8e >>>>>>>>> -no-user-config -nodefaults -chardev >>>>>>>>> socket,id=charmonitor,fd=27,server,nowait -mon >>>>>>>>> chardev=charmonitor,id=monitor,mode=control -rtc base=utc -no-shutdown >>>>>>>>> -global PIIX4_PM.disable_s3=1 -global PIIX4_PM.disable_s4=1 -boot >>>>>>>>> menu=off,strict=on -device >>>>>>>>> virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -drive >>>>>>>>> file=/var/tmp/localvmdRIozH/images/15023c8a-e3a7-4851-a97d-3b90996b423b/07fdcff3-11ce-4f7c-af05-0a878593e78e,format=qcow2,if=none,id=drive-virtio-disk0 >>>>>>>>> -device >>>>>>>>> virtio-blk-pci,scsi=off,bus=pci.0,addr=0x5,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1 >>>>>>>>> -drive >>>>>>>>> file=/var/tmp/localvmdRIozH/seed.iso,format=raw,if=none,id=drive-ide0-0-0,readonly=on >>>>>>>>> -device ide-cd,bus=ide.0,unit=0,drive=drive-ide0-0-0,id=ide0-0-0 >>>>>>>>> -netdev >>>>>>>>> tap,fd=29,id=hostnet0,vhost=on,vhostfd=30 -device >>>>>>>>> virtio-net-pci,netdev=hostnet0,id=net0,mac=00:16:3e:6c:e8:f9,bus=pci.0,addr=0x3 >>>>>>>>> -chardev pty,id=charserial0 -device >>>>>>>>> isa-serial,chardev=charserial0,id=serial0 -chardev >>>>>>>>> socket,id=charchannel0,fd=31,server,nowait -device >>>>>>>>> virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=org.qemu.guest_agent.0 >>>>>>>>> -vnc 127.0.0.1:0 -device >>>>>>>>> VGA,id=video0,vgamem_mb=16,bus=pci.0,addr=0x2 -object >>>>>>>>> rng-random,id=objrng0,filename=/dev/random -device >>>>>>>>> virtio-rng-pci,rng=objrng0,id=rng0,bus=pci.0,addr=0x6 -sandbox >>>>>>>>> on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny >>>>>>>>> -msg timestamp=on >>>>>>>>> root 28416 0.0 0.0 112712 980 pts/3 S+ 16:14 0:00 >>>>>>>>> grep --color=auto qemu >>>>>>>>> >>>>>>>>> >>>>>>>>> Before the first Error, while the vm was running for sure and the >>>>>>>>> disk was there, I also unsuccessfuly tried to connect to it with VNC >>>>>>>>> and >>>>>>>>> got the same error I got before : >>>>>>>>> >>>>>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# debug1: Connection to port >>>>>>>>> 5900 forwarding to vs-inf-int-kvm-fr-301-210.hostics.fr port 5900 >>>>>>>>> requested. >>>>>>>>> debug1: channel 3: new [direct-tcpip] >>>>>>>>> channel 3: open failed: connect failed: Connection refused >>>>>>>>> debug1: channel 3: free: direct-tcpip: listening port 5900 for >>>>>>>>> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900, connect from >>>>>>>>> 127.0.0.1 port 37002 to 127.0.0.1 port 5900, nchannels 4 >>>>>>>>> >>>>>>>>> >>>>>>>>> Guillaume Pavese >>>>>>>>> Ingénieur Système et Réseau >>>>>>>>> Interactiv-Group >>>>>>>>> >>>>>>>>> >>>>>>>>> On Mon, Feb 25, 2019 at 11:57 PM Guillaume Pavese < >>>>>>>>> guillaume.pav...@interactiv-group.com> wrote: >>>>>>>>> >>>>>>>>>> Something was definitely wrong ; as indicated, qemu process >>>>>>>>>> for guest=HostedEngineLocal was running but the disk file did not >>>>>>>>>> exist >>>>>>>>>> anymore... >>>>>>>>>> No surprise I could not connect >>>>>>>>>> >>>>>>>>>> I am retrying >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Guillaume Pavese >>>>>>>>>> Ingénieur Système et Réseau >>>>>>>>>> Interactiv-Group >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Mon, Feb 25, 2019 at 11:15 PM Guillaume Pavese < >>>>>>>>>> guillaume.pav...@interactiv-group.com> wrote: >>>>>>>>>> >>>>>>>>>>> It fails too : >>>>>>>>>>> I made sure PermitTunnel=yes in sshd config but when I try to >>>>>>>>>>> connect to the forwarded port I get the following error on the >>>>>>>>>>> openened >>>>>>>>>>> host ssh session : >>>>>>>>>>> >>>>>>>>>>> [gpavese@sheepora-X230 ~]$ ssh -v -L 5900: >>>>>>>>>>> vs-inf-int-kvm-fr-301-210.hostics.fr:5900 >>>>>>>>>>> r...@vs-inf-int-kvm-fr-301-210.hostics.fr >>>>>>>>>>> ... >>>>>>>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# >>>>>>>>>>> debug1: channel 3: free: direct-tcpip: listening port 5900 for >>>>>>>>>>> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900, connect from >>>>>>>>>>> ::1 port 42144 to ::1 port 5900, nchannels 4 >>>>>>>>>>> debug1: Connection to port 5900 forwarding to >>>>>>>>>>> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900 requested. >>>>>>>>>>> debug1: channel 3: new [direct-tcpip] >>>>>>>>>>> channel 3: open failed: connect failed: Connection refused >>>>>>>>>>> debug1: channel 3: free: direct-tcpip: listening port 5900 for >>>>>>>>>>> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900, connect from >>>>>>>>>>> 127.0.0.1 port 32778 to 127.0.0.1 port 5900, nchannels 4 >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> and in journalctl : >>>>>>>>>>> >>>>>>>>>>> févr. 25 14:55:38 vs-inf-int-kvm-fr-301-210.hostics.fr >>>>>>>>>>> sshd[19595]: error: connect_to >>>>>>>>>>> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900: failed. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Guillaume Pavese >>>>>>>>>>> Ingénieur Système et Réseau >>>>>>>>>>> Interactiv-Group >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Mon, Feb 25, 2019 at 10:44 PM Simone Tiraboschi < >>>>>>>>>>> stira...@redhat.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Mon, Feb 25, 2019 at 2:35 PM Guillaume Pavese < >>>>>>>>>>>> guillaume.pav...@interactiv-group.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> I made sure of everything and even stopped firewalld but still >>>>>>>>>>>>> can't connect : >>>>>>>>>>>>> >>>>>>>>>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# cat >>>>>>>>>>>>> /var/run/libvirt/qemu/HostedEngineLocal.xml >>>>>>>>>>>>> <graphics type='vnc' port='*5900*' autoport='yes' >>>>>>>>>>>>> *listen='127.0.0.1*'> >>>>>>>>>>>>> <listen type='address' address='*127.0.0.1*' >>>>>>>>>>>>> fromConfig='1' autoGenerated='no'/> >>>>>>>>>>>>> >>>>>>>>>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# netstat -pan | grep 59 >>>>>>>>>>>>> tcp 0 0 127.0.0.1:5900 0.0.0.0:* >>>>>>>>>>>>> LISTEN 13376/qemu-kvm >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> I suggest to try ssh tunneling, run >>>>>>>>>>>> ssh -L 5900:vs-inf-int-kvm-fr-301-210.hostics.fr:5900 >>>>>>>>>>>> r...@vs-inf-int-kvm-fr-301-210.hostics.fr >>>>>>>>>>>> >>>>>>>>>>>> and then >>>>>>>>>>>> remote-viewer vnc://localhost:5900 >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# systemctl status >>>>>>>>>>>>> firewalld.service >>>>>>>>>>>>> ● firewalld.service - firewalld - dynamic firewall daemon >>>>>>>>>>>>> Loaded: loaded (/usr/lib/systemd/system/firewalld.service; >>>>>>>>>>>>> enabled; vendor preset: enabled) >>>>>>>>>>>>> *Active: inactive (dead)* >>>>>>>>>>>>> *févr. 25 14:24:03 vs-inf-int-kvm-fr-301-210.hostics.fr >>>>>>>>>>>>> <http://vs-inf-int-kvm-fr-301-210.hostics.fr> systemd[1]: Stopped >>>>>>>>>>>>> firewalld >>>>>>>>>>>>> - dynamic firewall daemon.* >>>>>>>>>>>>> >>>>>>>>>>>>> From my laptop : >>>>>>>>>>>>> [gpavese@sheepora-X230 ~]$ telnet >>>>>>>>>>>>> vs-inf-int-kvm-fr-301-210.hostics.fr *5900* >>>>>>>>>>>>> Trying 10.199.210.11... >>>>>>>>>>>>> [*nothing gets through...*] >>>>>>>>>>>>> ^C >>>>>>>>>>>>> >>>>>>>>>>>>> For making sure : >>>>>>>>>>>>> [gpavese@sheepora-X230 ~]$ telnet >>>>>>>>>>>>> vs-inf-int-kvm-fr-301-210.hostics.fr *9090* >>>>>>>>>>>>> Trying 10.199.210.11... >>>>>>>>>>>>> *Connected* to vs-inf-int-kvm-fr-301-210.hostics.fr. >>>>>>>>>>>>> Escape character is '^]'. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Guillaume Pavese >>>>>>>>>>>>> Ingénieur Système et Réseau >>>>>>>>>>>>> Interactiv-Group >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Mon, Feb 25, 2019 at 10:24 PM Parth Dhanjal < >>>>>>>>>>>>> dpa...@redhat.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hey! >>>>>>>>>>>>>> >>>>>>>>>>>>>> You can check under /var/run/libvirt/qemu/HostedEngine.xml >>>>>>>>>>>>>> Search for 'vnc' >>>>>>>>>>>>>> From there you can look up the port on which the HE VM is >>>>>>>>>>>>>> available and connect to the same. >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Mon, Feb 25, 2019 at 6:47 PM Guillaume Pavese < >>>>>>>>>>>>>> guillaume.pav...@interactiv-group.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> 1) I am running in a Nested env, but under libvirt/kvm on >>>>>>>>>>>>>>> remote Centos 7.4 Hosts >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Please advise how to connect with VNC to the local HE vm. I >>>>>>>>>>>>>>> see it's running, but this is on a remote host, not my local >>>>>>>>>>>>>>> machine : >>>>>>>>>>>>>>> qemu 13376 100 3.7 17679424 845216 ? Sl 12:46 >>>>>>>>>>>>>>> 85:08 /usr/libexec/qemu-kvm -name >>>>>>>>>>>>>>> guest=HostedEngineLocal,debug-threads=on >>>>>>>>>>>>>>> -S -object >>>>>>>>>>>>>>> secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/domain-1-HostedEngineLocal/master-key.aes >>>>>>>>>>>>>>> -machine >>>>>>>>>>>>>>> pc-i440fx-rhel7.6.0,accel=kvm,usb=off,dump-guest-core=off -cpu >>>>>>>>>>>>>>> Haswell-noTSX,+kvmclock -m 16384 -realtime mlock=off -smp >>>>>>>>>>>>>>> 4,sockets=4,cores=1,threads=1 -uuid >>>>>>>>>>>>>>> 6fe7c1c3-ea93-4343-a385-0d9e14bb563a >>>>>>>>>>>>>>> -no-user-config -nodefaults -chardev >>>>>>>>>>>>>>> socket,id=charmonitor,fd=27,server,nowait -mon >>>>>>>>>>>>>>> chardev=charmonitor,id=monitor,mode=control -rtc base=utc >>>>>>>>>>>>>>> -no-shutdown >>>>>>>>>>>>>>> -global PIIX4_PM.disable_s3=1 -global PIIX4_PM.disable_s4=1 >>>>>>>>>>>>>>> -boot >>>>>>>>>>>>>>> menu=off,strict=on -device >>>>>>>>>>>>>>> virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -drive >>>>>>>>>>>>>>> file=/var/tmp/localvmgmyYik/images/15023c8a-e3a7-4851-a97d-3b90996b423b/07fdcff3-11ce-4f7c-af05-0a878593e78e,format=qcow2,if=none,id=drive-virtio-disk0 >>>>>>>>>>>>>>> -device >>>>>>>>>>>>>>> virtio-blk-pci,scsi=off,bus=pci.0,addr=0x5,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1 >>>>>>>>>>>>>>> -drive >>>>>>>>>>>>>>> file=/var/tmp/localvmgmyYik/seed.iso,format=raw,if=none,id=drive-ide0-0-0,readonly=on >>>>>>>>>>>>>>> -device >>>>>>>>>>>>>>> ide-cd,bus=ide.0,unit=0,drive=drive-ide0-0-0,id=ide0-0-0 -netdev >>>>>>>>>>>>>>> tap,fd=29,id=hostnet0,vhost=on,vhostfd=30 -device >>>>>>>>>>>>>>> virtio-net-pci,netdev=hostnet0,id=net0,mac=00:16:3e:3e:fe:28,bus=pci.0,addr=0x3 >>>>>>>>>>>>>>> -chardev pty,id=charserial0 -device >>>>>>>>>>>>>>> isa-serial,chardev=charserial0,id=serial0 -chardev >>>>>>>>>>>>>>> socket,id=charchannel0,fd=31,server,nowait -device >>>>>>>>>>>>>>> virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=org.qemu.guest_agent.0 >>>>>>>>>>>>>>> *-vnc 127.0.0.1:0 <http://127.0.0.1:0> -device >>>>>>>>>>>>>>> VGA*,id=video0,vgamem_mb=16,bus=pci.0,addr=0x2 >>>>>>>>>>>>>>> -object rng-random,id=objrng0,filename=/dev/random -device >>>>>>>>>>>>>>> virtio-rng-pci,rng=objrng0,id=rng0,bus=pci.0,addr=0x6 -sandbox >>>>>>>>>>>>>>> on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny >>>>>>>>>>>>>>> -msg timestamp=on >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> 2) [root@vs-inf-int-kvm-fr-301-210 ~]# cat >>>>>>>>>>>>>>> /etc/libvirt/qemu/networks/default.xml >>>>>>>>>>>>>>> <!-- >>>>>>>>>>>>>>> WARNING: THIS IS AN AUTO-GENERATED FILE. CHANGES TO IT ARE >>>>>>>>>>>>>>> LIKELY TO BE >>>>>>>>>>>>>>> OVERWRITTEN AND LOST. Changes to this xml configuration >>>>>>>>>>>>>>> should be made using: >>>>>>>>>>>>>>> virsh net-edit default >>>>>>>>>>>>>>> or other application using the libvirt API. >>>>>>>>>>>>>>> --> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> <network> >>>>>>>>>>>>>>> <name>default</name> >>>>>>>>>>>>>>> <uuid>ba7bbfc8-28b8-459e-a42d-c2d6218e2cb6</uuid> >>>>>>>>>>>>>>> <forward mode='nat'/> >>>>>>>>>>>>>>> <bridge name='virbr0' stp='on' delay='0'/> >>>>>>>>>>>>>>> <mac address='52:54:00:e5:fe:3b'/> >>>>>>>>>>>>>>> <ip address='192.168.122.1' netmask='255.255.255.0'> >>>>>>>>>>>>>>> <dhcp> >>>>>>>>>>>>>>> <range start='192.168.122.2' end='192.168.122.254'/> >>>>>>>>>>>>>>> </dhcp> >>>>>>>>>>>>>>> </ip> >>>>>>>>>>>>>>> </network> >>>>>>>>>>>>>>> You have new mail in /var/spool/mail/root >>>>>>>>>>>>>>> [root@vs-inf-int-kvm-fr-301-210 ~] >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Guillaume Pavese >>>>>>>>>>>>>>> Ingénieur Système et Réseau >>>>>>>>>>>>>>> Interactiv-Group >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Mon, Feb 25, 2019 at 9:57 PM Simone Tiraboschi < >>>>>>>>>>>>>>> stira...@redhat.com> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Mon, Feb 25, 2019 at 1:14 PM Guillaume Pavese < >>>>>>>>>>>>>>>> guillaume.pav...@interactiv-group.com> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> He deployment with "hosted-engine --deploy" fails at TASK >>>>>>>>>>>>>>>>> [ovirt.hosted_engine_setup : Get local VM IP] >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> See following Error : >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> 2019-02-25 12:46:50,154+0100 INFO >>>>>>>>>>>>>>>>> otopi.ovirt_hosted_engine_setup.ansible_utils >>>>>>>>>>>>>>>>> ansible_utils._process_output:109 TASK >>>>>>>>>>>>>>>>> [ovirt.hosted_engine_setup : Get >>>>>>>>>>>>>>>>> local VM IP] >>>>>>>>>>>>>>>>> 2019-02-25 12:55:26,823+0100 DEBUG >>>>>>>>>>>>>>>>> otopi.ovirt_hosted_engine_setup.ansible_utils >>>>>>>>>>>>>>>>> ansible_utils._process_output:103 {u'_ansible_parsed': True, >>>>>>>>>>>>>>>>> u'stderr_lines': [], u'cmd': u"virsh -r net-dhcp-leases >>>>>>>>>>>>>>>>> default | grep -i 00 >>>>>>>>>>>>>>>>> :16:3e:3e:fe:28 | awk '{ print $5 }' | cut -f1 -d'/'", >>>>>>>>>>>>>>>>> u'end': u'2019-02-25 12:55:26.666925', u'_ansible_no_log': >>>>>>>>>>>>>>>>> False, >>>>>>>>>>>>>>>>> u'stdout': u'', u'changed': True, u'invocation': >>>>>>>>>>>>>>>>> {u'module_args': {u'warn': >>>>>>>>>>>>>>>>> True, u'executable': >>>>>>>>>>>>>>>>> None, u'_uses_shell': True, u'_raw_params': u"virsh -r >>>>>>>>>>>>>>>>> net-dhcp-leases default | grep -i 00:16:3e:3e:fe:28 | awk '{ >>>>>>>>>>>>>>>>> print $5 }' | >>>>>>>>>>>>>>>>> cut -f1 -d'/'", u'removes': None, u'argv': None, u'creates': >>>>>>>>>>>>>>>>> None, >>>>>>>>>>>>>>>>> u'chdir': None, u'std >>>>>>>>>>>>>>>>> in': None}}, u'start': u'2019-02-25 12:55:26.584686', >>>>>>>>>>>>>>>>> u'attempts': 50, u'stderr': u'', u'rc': 0, u'delta': >>>>>>>>>>>>>>>>> u'0:00:00.082239', >>>>>>>>>>>>>>>>> u'stdout_lines': []} >>>>>>>>>>>>>>>>> 2019-02-25 12:55:26,924+0100 ERROR >>>>>>>>>>>>>>>>> otopi.ovirt_hosted_engine_setup.ansible_utils >>>>>>>>>>>>>>>>> ansible_utils._process_output:107 fatal: [localhost]: FAILED! >>>>>>>>>>>>>>>>> => >>>>>>>>>>>>>>>>> {"attempts": 50, "changed": true, "cmd": "virsh -r >>>>>>>>>>>>>>>>> net-dhcp-leases default >>>>>>>>>>>>>>>>> | grep -i 00:16:3e:3e:fe:28 | awk '{ print $5 }' | cut -f1 >>>>>>>>>>>>>>>>> -d'/'", "delta": >>>>>>>>>>>>>>>>> "0:00:00.082239", "end": "2019-02-25 12:55:26.666925", "rc": >>>>>>>>>>>>>>>>> 0, "start": >>>>>>>>>>>>>>>>> "2019-02-25 12:55:26.584686", "stderr": "", "stderr_lines": >>>>>>>>>>>>>>>>> [], "stdout": >>>>>>>>>>>>>>>>> "", "stdout_lines": []} >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Here we are just waiting for the bootstrap engine VM to >>>>>>>>>>>>>>>> fetch an IP address from default libvirt network over DHCP but >>>>>>>>>>>>>>>> it your case >>>>>>>>>>>>>>>> it never happened. >>>>>>>>>>>>>>>> Possible issues: something went wrong in the bootstrap >>>>>>>>>>>>>>>> process for the engine VM or the default libvirt network is >>>>>>>>>>>>>>>> not correctly >>>>>>>>>>>>>>>> configured. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> 1. can you try to reach the engine VM via VNC and check >>>>>>>>>>>>>>>> what's happening there? (another question, are you running it >>>>>>>>>>>>>>>> nested? AFAIK >>>>>>>>>>>>>>>> it will not work if nested over ESXi) >>>>>>>>>>>>>>>> 2. can you please share the output of >>>>>>>>>>>>>>>> cat /etc/libvirt/qemu/networks/default.xml >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Guillaume Pavese >>>>>>>>>>>>>>>>> Ingénieur Système et Réseau >>>>>>>>>>>>>>>>> Interactiv-Group >>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>> Users mailing list -- users@ovirt.org >>>>>>>>>>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>>>>>>>>>>>> Privacy Statement: >>>>>>>>>>>>>>>>> https://www.ovirt.org/site/privacy-policy/ >>>>>>>>>>>>>>>>> oVirt Code of Conduct: >>>>>>>>>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>>>>>>>>>>> List Archives: >>>>>>>>>>>>>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/VXRMU3SQWTMB2YYNMOMD7I5NX7RZQ2IW/ >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>> Users mailing list -- users@ovirt.org >>>>>>>>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>>>>>>>>>> Privacy Statement: >>>>>>>>>>>>>>> https://www.ovirt.org/site/privacy-policy/ >>>>>>>>>>>>>>> oVirt Code of Conduct: >>>>>>>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>>>>>>>>> List Archives: >>>>>>>>>>>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/45UR44ITQTV7YVE6GLJWEV2AV22TIGXK/ >>>>>>>>>>>>>>> >>>>>>>>>>>>>>
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/QU5UCNIUPQTK6WTWIW6O5DN6ZBVQY6XW/