I still can't connect with VNC remotely but locally with X forwarding it
works.
However my connection has too high latency for that to be usable (I'm in
Japan, my hosts in France, ~250 ms ping)

But I could see that the VM is booted!

and in Hosts logs there is :

févr. 25 18:51:12 vs-inf-int-kvm-fr-301-210.hostics.fr python[14719]:
ansible-command Invoked with warn=True executable=None _uses_shell=True
_raw_params=virsh -r net-dhcp-leases default | grep -i 00:16:3e:1d:4b:b6 |
awk '{ print $5 }' | cut -f1 -d'/' removes=None argv=None creates=None
chdir=None stdin=None
févr. 25 18:51:30 vs-inf-int-kvm-fr-301-210.hostics.fr dnsmasq-dhcp[6310]:
DHCPDISCOVER(virbr0) 00:16:3e:1d:4b:b6
févr. 25 18:51:30 vs-inf-int-kvm-fr-301-210.hostics.fr dnsmasq-dhcp[6310]:
DHCPOFFER(virbr0) 192.168.122.14 00:16:3e:1d:4b:b6
févr. 25 18:51:30 vs-inf-int-kvm-fr-301-210.hostics.fr dnsmasq-dhcp[6310]:
DHCPREQUEST(virbr0) 192.168.122.14 00:16:3e:1d:4b:b6
févr. 25 18:51:30 vs-inf-int-kvm-fr-301-210.hostics.fr dnsmasq-dhcp[6310]:
DHCPACK(virbr0) 192.168.122.14 00:16:3e:1d:4b:b6 vs-inf-int-ovt-fr-301-210
févr. 25 18:51:42 vs-inf-int-kvm-fr-301-210.hostics.fr python[14757]:
ansible-command Invoked with warn=True executable=None _uses_shell=True
_raw_params=virsh -r net-dhcp-leases default | grep -i 00:16:3e:1d:4b:b6 |
awk '{ print $5 }' | cut -f1 -d'/' removes=None argv=None creates=None
chdir=None stdin=None
févr. 25 18:52:12 vs-inf-int-kvm-fr-301-210.hostics.fr python[14789]:
ansible-command Invoked with warn=True executable=None _uses_shell=True
_raw_params=virsh -r net-dhcp-leases default | grep -i 00:16:3e:1d:4b:b6 |
awk '{ print $5 }' | cut -f1 -d'/' removes=None argv=None creates=None
chdir=None stdin=None
févr. 25 18:52:43 vs-inf-int-kvm-fr-301-210.hostics.fr python[14818]:
ansible-command Invoked with warn=True executable=None _uses_shell=True
_raw_params=virsh -r net-dhcp-leases default | grep -i 00:16:3e:1d:4b:b6 |
awk '{ print $5 }' | cut -f1 -d'/' removes=None argv=None creates=None
chdir=None stdin=None
....

ssh to the vm works too :

[root@vs-inf-int-kvm-fr-301-210 ~]# ssh root@192.168.122.14
The authenticity of host '192.168.122.14 (192.168.122.14)' can't be
established.
ECDSA key fingerprint is SHA256:+/pUzTGVA4kCyICb7XgqrxWYYkqzmDjVmdAahiBFgOQ.
ECDSA key fingerprint is
MD5:4b:ef:ff:4a:7c:1a:af:c2:af:4a:0f:14:a3:c5:31:fb.
Are you sure you want to continue connecting (yes/no)? yes
Warning: Permanently added '192.168.122.14' (ECDSA) to the list of known
hosts.
root@192.168.122.14's password:
[root@vs-inf-int-ovt-fr-301-210 ~]#


But the test that the playbook tries still fails with empty result :

[root@vs-inf-int-kvm-fr-301-210 ~]# virsh -r net-dhcp-leases default
 Expiry Time          MAC address        Protocol  IP address
  Hostname        Client ID or DUID
-------------------------------------------------------------------------------------------------------------------

[root@vs-inf-int-kvm-fr-301-210 ~]#




Guillaume Pavese
Ingénieur Système et Réseau
Interactiv-Group


On Tue, Feb 26, 2019 at 1:54 AM Simone Tiraboschi <stira...@redhat.com>
wrote:

>
>
> On Mon, Feb 25, 2019 at 5:50 PM Guillaume Pavese <
> guillaume.pav...@interactiv-group.com> wrote:
>
>> I did that but no success yet.
>>
>> I see that "Get local VM IP" task tries the following :
>>
>> virsh -r net-dhcp-leases default | grep -i {{ he_vm_mac_addr }} | awk '{
>> print $5 }' | cut -f1 -d'/'
>>
>>
>> However while the task is running, and vm running in qemu, "virsh -r
>> net-dhcp-leases default" never returns anything :
>>
>
> Yes, I think that libvirt will never provide a DHCP lease since the
> appliance OS never correctly complete the boot process.
> I'd suggest to connect to the running VM via vnc DURING the boot process
> and check what's wrong.
>
>
>> [root@vs-inf-int-kvm-fr-301-210 ~]# virsh -r net-dhcp-leases default
>>  Expiry Time          MAC address        Protocol  IP address
>>     Hostname        Client ID or DUID
>>
>> -------------------------------------------------------------------------------------------------------------------
>>
>> [root@vs-inf-int-kvm-fr-301-210 ~]#
>>
>>
>>
>>
>> Guillaume Pavese
>> Ingénieur Système et Réseau
>> Interactiv-Group
>>
>>
>> On Tue, Feb 26, 2019 at 12:33 AM Simone Tiraboschi <stira...@redhat.com>
>> wrote:
>>
>>> OK, try this:
>>> temporary
>>> edit 
>>> /usr/share/ansible/roles/ovirt.hosted_engine_setup/tasks/bootstrap_local_vm/02_create_local_vm.yml
>>> around line 120
>>> and edit tasks "Get local VM IP"
>>> changing from "retries: 50" to  "retries: 500" so that you have more
>>> time to debug it
>>>
>>>
>>>
>>> On Mon, Feb 25, 2019 at 4:20 PM Guillaume Pavese <
>>> guillaume.pav...@interactiv-group.com> wrote:
>>>
>>>> I retried after killing the remaining qemu process and
>>>> doing ovirt-hosted-engine-cleanup
>>>> The new attempt failed again at the same step. Then after it fails, it
>>>> cleans the temporary files (and vm disk) but *qemu still runs!* :
>>>>
>>>> [ INFO  ] TASK [ovirt.hosted_engine_setup : Get local VM IP]
>>>>
>>>> [ ERROR ] fatal: [localhost]: FAILED! => {"attempts": 50, "changed":
>>>> true, "cmd": "virsh -r net-dhcp-leases default | grep -i 00:16:3e:6c:e8:f9
>>>> | awk '{ print $5 }' | cut -f1 -d'/'", "delta": "0:00:00.092436", "end":
>>>> "2019-02-25 16:09:38.863263", "rc": 0, "start": "2019-02-25
>>>> 16:09:38.770827", "stderr": "", "stderr_lines": [], "stdout": "",
>>>> "stdout_lines": []}
>>>> [ INFO  ] TASK [ovirt.hosted_engine_setup : include_tasks]
>>>> [ INFO  ] ok: [localhost]
>>>> [ INFO  ] TASK [ovirt.hosted_engine_setup : Remove local vm dir]
>>>> [ INFO  ] changed: [localhost]
>>>> [ INFO  ] TASK [ovirt.hosted_engine_setup : Remove temporary entry in
>>>> /etc/hosts for the local VM]
>>>> [ INFO  ] ok: [localhost]
>>>> [ INFO  ] TASK [ovirt.hosted_engine_setup : Notify the user about a
>>>> failure]
>>>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "The
>>>> system may not be provisioned according to the playbook results: please
>>>> check the logs for the issue, fix accordingly or re-deploy from 
>>>> scratch.\n"}
>>>> [ ERROR ] Failed to execute stage 'Closing up': Failed executing
>>>> ansible-playbook
>>>> [ INFO  ] Stage: Clean up
>>>> [ INFO  ] Cleaning temporary resources
>>>> ...
>>>>
>>>> [ INFO  ] TASK [ovirt.hosted_engine_setup : Remove local vm dir]
>>>> [ INFO  ] ok: [localhost]
>>>> [ INFO  ] TASK [ovirt.hosted_engine_setup : Remove temporary entry in
>>>> /etc/hosts for the local VM]
>>>> [ INFO  ] ok: [localhost]
>>>> [ INFO  ] Generating answer file
>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20190225161011.conf'
>>>> [ INFO  ] Stage: Pre-termination
>>>> [ INFO  ] Stage: Termination
>>>> [ ERROR ] Hosted Engine deployment failed: please check the logs for
>>>> the issue, fix accordingly or re-deploy from scratch.
>>>>
>>>>
>>>>
>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# ps aux | grep qemu
>>>> root      4021  0.0  0.0  24844  1788 ?        Ss   févr.22   0:00
>>>> /usr/bin/qemu-ga --method=virtio-serial
>>>> --path=/dev/virtio-ports/org.qemu.guest_agent.0
>>>> --blacklist=guest-file-open,guest-file-close,guest-file-read,guest-file-write,guest-file-seek,guest-file-flush,guest-exec,guest-exec-status
>>>> -F/etc/qemu-ga/fsfreeze-hook
>>>> qemu     26463 22.9  4.8 17684512 1088844 ?    Sl   16:01   3:09
>>>> /usr/libexec/qemu-kvm -name guest=HostedEngineLocal,debug-threads=on -S
>>>> -object
>>>> secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/domain-1-HostedEngineLocal/master-key.aes
>>>> -machine pc-i440fx-rhel7.6.0,accel=kvm,usb=off,dump-guest-core=off -cpu
>>>> Haswell-noTSX,+kvmclock -m 16384 -realtime mlock=off -smp
>>>> 4,sockets=4,cores=1,threads=1 -uuid 316eca5f-81de-4a0b-af1f-58f910402a8e
>>>> -no-user-config -nodefaults -chardev
>>>> socket,id=charmonitor,fd=27,server,nowait -mon
>>>> chardev=charmonitor,id=monitor,mode=control -rtc base=utc -no-shutdown
>>>> -global PIIX4_PM.disable_s3=1 -global PIIX4_PM.disable_s4=1 -boot
>>>> menu=off,strict=on -device
>>>> virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -drive
>>>> file=/var/tmp/localvmdRIozH/images/15023c8a-e3a7-4851-a97d-3b90996b423b/07fdcff3-11ce-4f7c-af05-0a878593e78e,format=qcow2,if=none,id=drive-virtio-disk0
>>>> -device
>>>> virtio-blk-pci,scsi=off,bus=pci.0,addr=0x5,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1
>>>> -drive
>>>> file=/var/tmp/localvmdRIozH/seed.iso,format=raw,if=none,id=drive-ide0-0-0,readonly=on
>>>> -device ide-cd,bus=ide.0,unit=0,drive=drive-ide0-0-0,id=ide0-0-0 -netdev
>>>> tap,fd=29,id=hostnet0,vhost=on,vhostfd=30 -device
>>>> virtio-net-pci,netdev=hostnet0,id=net0,mac=00:16:3e:6c:e8:f9,bus=pci.0,addr=0x3
>>>> -chardev pty,id=charserial0 -device
>>>> isa-serial,chardev=charserial0,id=serial0 -chardev
>>>> socket,id=charchannel0,fd=31,server,nowait -device
>>>> virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=org.qemu.guest_agent.0
>>>> -vnc 127.0.0.1:0 -device VGA,id=video0,vgamem_mb=16,bus=pci.0,addr=0x2
>>>> -object rng-random,id=objrng0,filename=/dev/random -device
>>>> virtio-rng-pci,rng=objrng0,id=rng0,bus=pci.0,addr=0x6 -sandbox
>>>> on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny
>>>> -msg timestamp=on
>>>> root     28416  0.0  0.0 112712   980 pts/3    S+   16:14   0:00 grep
>>>> --color=auto qemu
>>>>
>>>>
>>>> Before the first Error, while the vm was running for sure and the disk
>>>> was there, I also unsuccessfuly tried to connect to it with VNC  and got
>>>> the same error I got before :
>>>>
>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# debug1: Connection to port 5900
>>>> forwarding to vs-inf-int-kvm-fr-301-210.hostics.fr port 5900 requested.
>>>> debug1: channel 3: new [direct-tcpip]
>>>> channel 3: open failed: connect failed: Connection refused
>>>> debug1: channel 3: free: direct-tcpip: listening port 5900 for
>>>> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900, connect from 127.0.0.1
>>>> port 37002 to 127.0.0.1 port 5900, nchannels 4
>>>>
>>>>
>>>> Guillaume Pavese
>>>> Ingénieur Système et Réseau
>>>> Interactiv-Group
>>>>
>>>>
>>>> On Mon, Feb 25, 2019 at 11:57 PM Guillaume Pavese <
>>>> guillaume.pav...@interactiv-group.com> wrote:
>>>>
>>>>> Something was definitely wrong ; as indicated, qemu process
>>>>> for guest=HostedEngineLocal was running but the disk file did not exist
>>>>> anymore...
>>>>> No surprise I could not connect
>>>>>
>>>>> I am retrying
>>>>>
>>>>>
>>>>> Guillaume Pavese
>>>>> Ingénieur Système et Réseau
>>>>> Interactiv-Group
>>>>>
>>>>>
>>>>> On Mon, Feb 25, 2019 at 11:15 PM Guillaume Pavese <
>>>>> guillaume.pav...@interactiv-group.com> wrote:
>>>>>
>>>>>> It fails too :
>>>>>> I made sure PermitTunnel=yes in sshd config but when I try to connect
>>>>>> to the forwarded port I get the following error on the openened host ssh
>>>>>> session :
>>>>>>
>>>>>> [gpavese@sheepora-X230 ~]$ ssh -v -L 5900:
>>>>>> vs-inf-int-kvm-fr-301-210.hostics.fr:5900
>>>>>> r...@vs-inf-int-kvm-fr-301-210.hostics.fr
>>>>>> ...
>>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]#
>>>>>> debug1: channel 3: free: direct-tcpip: listening port 5900 for
>>>>>> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900, connect from ::1
>>>>>> port 42144 to ::1 port 5900, nchannels 4
>>>>>> debug1: Connection to port 5900 forwarding to
>>>>>> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900 requested.
>>>>>> debug1: channel 3: new [direct-tcpip]
>>>>>> channel 3: open failed: connect failed: Connection refused
>>>>>> debug1: channel 3: free: direct-tcpip: listening port 5900 for
>>>>>> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900, connect from
>>>>>> 127.0.0.1 port 32778 to 127.0.0.1 port 5900, nchannels 4
>>>>>>
>>>>>>
>>>>>> and in journalctl :
>>>>>>
>>>>>> févr. 25 14:55:38 vs-inf-int-kvm-fr-301-210.hostics.fr sshd[19595]:
>>>>>> error: connect_to vs-inf-int-kvm-fr-301-210.hostics.fr port 5900:
>>>>>> failed.
>>>>>>
>>>>>>
>>>>>> Guillaume Pavese
>>>>>> Ingénieur Système et Réseau
>>>>>> Interactiv-Group
>>>>>>
>>>>>>
>>>>>> On Mon, Feb 25, 2019 at 10:44 PM Simone Tiraboschi <
>>>>>> stira...@redhat.com> wrote:
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Feb 25, 2019 at 2:35 PM Guillaume Pavese <
>>>>>>> guillaume.pav...@interactiv-group.com> wrote:
>>>>>>>
>>>>>>>> I made sure of everything and even stopped firewalld but still
>>>>>>>> can't connect :
>>>>>>>>
>>>>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# cat
>>>>>>>> /var/run/libvirt/qemu/HostedEngineLocal.xml
>>>>>>>>      <graphics type='vnc' port='*5900*' autoport='yes'
>>>>>>>> *listen='127.0.0.1*'>
>>>>>>>>         <listen type='address' address='*127.0.0.1*'
>>>>>>>> fromConfig='1' autoGenerated='no'/>
>>>>>>>>
>>>>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# netstat -pan | grep 59
>>>>>>>> tcp        0      0 127.0.0.1:5900          0.0.0.0:*
>>>>>>>>  LISTEN      13376/qemu-kvm
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> I suggest to try ssh tunneling, run
>>>>>>> ssh -L 5900:vs-inf-int-kvm-fr-301-210.hostics.fr:5900
>>>>>>> r...@vs-inf-int-kvm-fr-301-210.hostics.fr
>>>>>>>
>>>>>>> and then
>>>>>>> remote-viewer vnc://localhost:5900
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>>
>>>>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# systemctl status
>>>>>>>> firewalld.service
>>>>>>>> ● firewalld.service - firewalld - dynamic firewall daemon
>>>>>>>>    Loaded: loaded (/usr/lib/systemd/system/firewalld.service;
>>>>>>>> enabled; vendor preset: enabled)
>>>>>>>>    *Active: inactive (dead)*
>>>>>>>> *févr. 25 14:24:03 vs-inf-int-kvm-fr-301-210.hostics.fr
>>>>>>>> <http://vs-inf-int-kvm-fr-301-210.hostics.fr> systemd[1]: Stopped 
>>>>>>>> firewalld
>>>>>>>> - dynamic firewall daemon.*
>>>>>>>>
>>>>>>>> From my laptop :
>>>>>>>> [gpavese@sheepora-X230 ~]$ telnet
>>>>>>>> vs-inf-int-kvm-fr-301-210.hostics.fr *5900*
>>>>>>>> Trying 10.199.210.11...
>>>>>>>> [*nothing gets through...*]
>>>>>>>> ^C
>>>>>>>>
>>>>>>>> For making sure :
>>>>>>>> [gpavese@sheepora-X230 ~]$ telnet
>>>>>>>> vs-inf-int-kvm-fr-301-210.hostics.fr *9090*
>>>>>>>> Trying 10.199.210.11...
>>>>>>>> *Connected* to vs-inf-int-kvm-fr-301-210.hostics.fr.
>>>>>>>> Escape character is '^]'.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> Guillaume Pavese
>>>>>>>> Ingénieur Système et Réseau
>>>>>>>> Interactiv-Group
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Feb 25, 2019 at 10:24 PM Parth Dhanjal <dpa...@redhat.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Hey!
>>>>>>>>>
>>>>>>>>> You can check under /var/run/libvirt/qemu/HostedEngine.xml
>>>>>>>>> Search for 'vnc'
>>>>>>>>> From there you can look up the port on which the HE VM is
>>>>>>>>> available and connect to the same.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Mon, Feb 25, 2019 at 6:47 PM Guillaume Pavese <
>>>>>>>>> guillaume.pav...@interactiv-group.com> wrote:
>>>>>>>>>
>>>>>>>>>> 1) I am running in a Nested env, but under libvirt/kvm on remote
>>>>>>>>>> Centos 7.4 Hosts
>>>>>>>>>>
>>>>>>>>>> Please advise how to connect with VNC to the local HE vm. I see
>>>>>>>>>> it's running, but this is on a remote host, not my local machine :
>>>>>>>>>> qemu     13376  100  3.7 17679424 845216 ?     Sl   12:46  85:08
>>>>>>>>>> /usr/libexec/qemu-kvm -name guest=HostedEngineLocal,debug-threads=on 
>>>>>>>>>> -S
>>>>>>>>>> -object
>>>>>>>>>> secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/domain-1-HostedEngineLocal/master-key.aes
>>>>>>>>>> -machine pc-i440fx-rhel7.6.0,accel=kvm,usb=off,dump-guest-core=off 
>>>>>>>>>> -cpu
>>>>>>>>>> Haswell-noTSX,+kvmclock -m 16384 -realtime mlock=off -smp
>>>>>>>>>> 4,sockets=4,cores=1,threads=1 -uuid 
>>>>>>>>>> 6fe7c1c3-ea93-4343-a385-0d9e14bb563a
>>>>>>>>>> -no-user-config -nodefaults -chardev
>>>>>>>>>> socket,id=charmonitor,fd=27,server,nowait -mon
>>>>>>>>>> chardev=charmonitor,id=monitor,mode=control -rtc base=utc 
>>>>>>>>>> -no-shutdown
>>>>>>>>>> -global PIIX4_PM.disable_s3=1 -global PIIX4_PM.disable_s4=1 -boot
>>>>>>>>>> menu=off,strict=on -device
>>>>>>>>>> virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -drive
>>>>>>>>>> file=/var/tmp/localvmgmyYik/images/15023c8a-e3a7-4851-a97d-3b90996b423b/07fdcff3-11ce-4f7c-af05-0a878593e78e,format=qcow2,if=none,id=drive-virtio-disk0
>>>>>>>>>> -device
>>>>>>>>>> virtio-blk-pci,scsi=off,bus=pci.0,addr=0x5,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1
>>>>>>>>>> -drive
>>>>>>>>>> file=/var/tmp/localvmgmyYik/seed.iso,format=raw,if=none,id=drive-ide0-0-0,readonly=on
>>>>>>>>>> -device ide-cd,bus=ide.0,unit=0,drive=drive-ide0-0-0,id=ide0-0-0 
>>>>>>>>>> -netdev
>>>>>>>>>> tap,fd=29,id=hostnet0,vhost=on,vhostfd=30 -device
>>>>>>>>>> virtio-net-pci,netdev=hostnet0,id=net0,mac=00:16:3e:3e:fe:28,bus=pci.0,addr=0x3
>>>>>>>>>> -chardev pty,id=charserial0 -device
>>>>>>>>>> isa-serial,chardev=charserial0,id=serial0 -chardev
>>>>>>>>>> socket,id=charchannel0,fd=31,server,nowait -device
>>>>>>>>>> virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=org.qemu.guest_agent.0
>>>>>>>>>> *-vnc 127.0.0.1:0 <http://127.0.0.1:0> -device 
>>>>>>>>>> VGA*,id=video0,vgamem_mb=16,bus=pci.0,addr=0x2
>>>>>>>>>> -object rng-random,id=objrng0,filename=/dev/random -device
>>>>>>>>>> virtio-rng-pci,rng=objrng0,id=rng0,bus=pci.0,addr=0x6 -sandbox
>>>>>>>>>> on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny
>>>>>>>>>> -msg timestamp=on
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> 2) [root@vs-inf-int-kvm-fr-301-210 ~]# cat
>>>>>>>>>> /etc/libvirt/qemu/networks/default.xml
>>>>>>>>>> <!--
>>>>>>>>>> WARNING: THIS IS AN AUTO-GENERATED FILE. CHANGES TO IT ARE LIKELY
>>>>>>>>>> TO BE
>>>>>>>>>> OVERWRITTEN AND LOST. Changes to this xml configuration should be
>>>>>>>>>> made using:
>>>>>>>>>>   virsh net-edit default
>>>>>>>>>> or other application using the libvirt API.
>>>>>>>>>> -->
>>>>>>>>>>
>>>>>>>>>> <network>
>>>>>>>>>>   <name>default</name>
>>>>>>>>>>   <uuid>ba7bbfc8-28b8-459e-a42d-c2d6218e2cb6</uuid>
>>>>>>>>>>   <forward mode='nat'/>
>>>>>>>>>>   <bridge name='virbr0' stp='on' delay='0'/>
>>>>>>>>>>   <mac address='52:54:00:e5:fe:3b'/>
>>>>>>>>>>   <ip address='192.168.122.1' netmask='255.255.255.0'>
>>>>>>>>>>     <dhcp>
>>>>>>>>>>       <range start='192.168.122.2' end='192.168.122.254'/>
>>>>>>>>>>     </dhcp>
>>>>>>>>>>   </ip>
>>>>>>>>>> </network>
>>>>>>>>>> You have new mail in /var/spool/mail/root
>>>>>>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Guillaume Pavese
>>>>>>>>>> Ingénieur Système et Réseau
>>>>>>>>>> Interactiv-Group
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Mon, Feb 25, 2019 at 9:57 PM Simone Tiraboschi <
>>>>>>>>>> stira...@redhat.com> wrote:
>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Mon, Feb 25, 2019 at 1:14 PM Guillaume Pavese <
>>>>>>>>>>> guillaume.pav...@interactiv-group.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> He deployment with "hosted-engine --deploy" fails at TASK
>>>>>>>>>>>> [ovirt.hosted_engine_setup : Get local VM IP]
>>>>>>>>>>>>
>>>>>>>>>>>> See following Error :
>>>>>>>>>>>>
>>>>>>>>>>>> 2019-02-25 12:46:50,154+0100 INFO
>>>>>>>>>>>> otopi.ovirt_hosted_engine_setup.ansible_utils
>>>>>>>>>>>> ansible_utils._process_output:109 TASK [ovirt.hosted_engine_setup 
>>>>>>>>>>>> : Get
>>>>>>>>>>>> local VM IP]
>>>>>>>>>>>> 2019-02-25 12:55:26,823+0100 DEBUG
>>>>>>>>>>>> otopi.ovirt_hosted_engine_setup.ansible_utils
>>>>>>>>>>>> ansible_utils._process_output:103 {u'_ansible_parsed': True,
>>>>>>>>>>>> u'stderr_lines': [], u'cmd': u"virsh -r net-dhcp-leases default | 
>>>>>>>>>>>> grep -i 00
>>>>>>>>>>>> :16:3e:3e:fe:28 | awk '{ print $5 }' | cut -f1 -d'/'", u'end':
>>>>>>>>>>>> u'2019-02-25 12:55:26.666925', u'_ansible_no_log': False, 
>>>>>>>>>>>> u'stdout': u'',
>>>>>>>>>>>> u'changed': True, u'invocation': {u'module_args': {u'warn': True,
>>>>>>>>>>>> u'executable':
>>>>>>>>>>>> None, u'_uses_shell': True, u'_raw_params': u"virsh -r
>>>>>>>>>>>> net-dhcp-leases default | grep -i 00:16:3e:3e:fe:28 | awk '{ print 
>>>>>>>>>>>> $5 }' |
>>>>>>>>>>>> cut -f1 -d'/'", u'removes': None, u'argv': None, u'creates': None,
>>>>>>>>>>>> u'chdir': None, u'std
>>>>>>>>>>>> in': None}}, u'start': u'2019-02-25 12:55:26.584686',
>>>>>>>>>>>> u'attempts': 50, u'stderr': u'', u'rc': 0, u'delta': 
>>>>>>>>>>>> u'0:00:00.082239',
>>>>>>>>>>>> u'stdout_lines': []}
>>>>>>>>>>>> 2019-02-25 12:55:26,924+0100 ERROR
>>>>>>>>>>>> otopi.ovirt_hosted_engine_setup.ansible_utils
>>>>>>>>>>>> ansible_utils._process_output:107 fatal: [localhost]: FAILED! =>
>>>>>>>>>>>> {"attempts": 50, "changed": true, "cmd": "virsh -r net-dhcp-leases 
>>>>>>>>>>>> default
>>>>>>>>>>>> | grep -i 00:16:3e:3e:fe:28 | awk '{ print $5 }' | cut -f1 -d'/'", 
>>>>>>>>>>>> "delta":
>>>>>>>>>>>> "0:00:00.082239", "end": "2019-02-25 12:55:26.666925", "rc": 0, 
>>>>>>>>>>>> "start":
>>>>>>>>>>>> "2019-02-25 12:55:26.584686", "stderr": "", "stderr_lines": [], 
>>>>>>>>>>>> "stdout":
>>>>>>>>>>>> "", "stdout_lines": []}
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Here we are just waiting for the bootstrap engine VM to fetch an
>>>>>>>>>>> IP address from default libvirt network over DHCP but it your case 
>>>>>>>>>>> it never
>>>>>>>>>>> happened.
>>>>>>>>>>> Possible issues: something went wrong in the bootstrap process
>>>>>>>>>>> for the engine VM or the default libvirt network is not correctly
>>>>>>>>>>> configured.
>>>>>>>>>>>
>>>>>>>>>>> 1. can you try to reach the engine VM via VNC and check what's
>>>>>>>>>>> happening there? (another question, are you running it nested? 
>>>>>>>>>>> AFAIK it
>>>>>>>>>>> will not work if nested over ESXi)
>>>>>>>>>>> 2. can you please share the output of
>>>>>>>>>>> cat /etc/libvirt/qemu/networks/default.xml
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Guillaume Pavese
>>>>>>>>>>>> Ingénieur Système et Réseau
>>>>>>>>>>>> Interactiv-Group
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> Users mailing list -- users@ovirt.org
>>>>>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>>>>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>>>>>>>>> oVirt Code of Conduct:
>>>>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>>>>>>>>> List Archives:
>>>>>>>>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/VXRMU3SQWTMB2YYNMOMD7I5NX7RZQ2IW/
>>>>>>>>>>>>
>>>>>>>>>>> _______________________________________________
>>>>>>>>>> Users mailing list -- users@ovirt.org
>>>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>>>>>>> oVirt Code of Conduct:
>>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>>>>>>> List Archives:
>>>>>>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/45UR44ITQTV7YVE6GLJWEV2AV22TIGXK/
>>>>>>>>>>
>>>>>>>>>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3QETHEV3PFK2VS7EH35RVN437XAGPRXX/

Reply via email to