OK, try this:
temporary
edit 
/usr/share/ansible/roles/ovirt.hosted_engine_setup/tasks/bootstrap_local_vm/02_create_local_vm.yml
around line 120
and edit tasks "Get local VM IP"
changing from "retries: 50" to  "retries: 500" so that you have more time
to debug it



On Mon, Feb 25, 2019 at 4:20 PM Guillaume Pavese <
guillaume.pav...@interactiv-group.com> wrote:

> I retried after killing the remaining qemu process and
> doing ovirt-hosted-engine-cleanup
> The new attempt failed again at the same step. Then after it fails, it
> cleans the temporary files (and vm disk) but *qemu still runs!* :
>
> [ INFO  ] TASK [ovirt.hosted_engine_setup : Get local VM IP]
>
> [ ERROR ] fatal: [localhost]: FAILED! => {"attempts": 50, "changed": true,
> "cmd": "virsh -r net-dhcp-leases default | grep -i 00:16:3e:6c:e8:f9 | awk
> '{ print $5 }' | cut -f1 -d'/'", "delta": "0:00:00.092436", "end":
> "2019-02-25 16:09:38.863263", "rc": 0, "start": "2019-02-25
> 16:09:38.770827", "stderr": "", "stderr_lines": [], "stdout": "",
> "stdout_lines": []}
> [ INFO  ] TASK [ovirt.hosted_engine_setup : include_tasks]
> [ INFO  ] ok: [localhost]
> [ INFO  ] TASK [ovirt.hosted_engine_setup : Remove local vm dir]
> [ INFO  ] changed: [localhost]
> [ INFO  ] TASK [ovirt.hosted_engine_setup : Remove temporary entry in
> /etc/hosts for the local VM]
> [ INFO  ] ok: [localhost]
> [ INFO  ] TASK [ovirt.hosted_engine_setup : Notify the user about a
> failure]
> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "The
> system may not be provisioned according to the playbook results: please
> check the logs for the issue, fix accordingly or re-deploy from scratch.\n"}
> [ ERROR ] Failed to execute stage 'Closing up': Failed executing
> ansible-playbook
> [ INFO  ] Stage: Clean up
> [ INFO  ] Cleaning temporary resources
> ...
>
> [ INFO  ] TASK [ovirt.hosted_engine_setup : Remove local vm dir]
> [ INFO  ] ok: [localhost]
> [ INFO  ] TASK [ovirt.hosted_engine_setup : Remove temporary entry in
> /etc/hosts for the local VM]
> [ INFO  ] ok: [localhost]
> [ INFO  ] Generating answer file
> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20190225161011.conf'
> [ INFO  ] Stage: Pre-termination
> [ INFO  ] Stage: Termination
> [ ERROR ] Hosted Engine deployment failed: please check the logs for the
> issue, fix accordingly or re-deploy from scratch.
>
>
>
> [root@vs-inf-int-kvm-fr-301-210 ~]# ps aux | grep qemu
> root      4021  0.0  0.0  24844  1788 ?        Ss   févr.22   0:00
> /usr/bin/qemu-ga --method=virtio-serial
> --path=/dev/virtio-ports/org.qemu.guest_agent.0
> --blacklist=guest-file-open,guest-file-close,guest-file-read,guest-file-write,guest-file-seek,guest-file-flush,guest-exec,guest-exec-status
> -F/etc/qemu-ga/fsfreeze-hook
> qemu     26463 22.9  4.8 17684512 1088844 ?    Sl   16:01   3:09
> /usr/libexec/qemu-kvm -name guest=HostedEngineLocal,debug-threads=on -S
> -object
> secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/domain-1-HostedEngineLocal/master-key.aes
> -machine pc-i440fx-rhel7.6.0,accel=kvm,usb=off,dump-guest-core=off -cpu
> Haswell-noTSX,+kvmclock -m 16384 -realtime mlock=off -smp
> 4,sockets=4,cores=1,threads=1 -uuid 316eca5f-81de-4a0b-af1f-58f910402a8e
> -no-user-config -nodefaults -chardev
> socket,id=charmonitor,fd=27,server,nowait -mon
> chardev=charmonitor,id=monitor,mode=control -rtc base=utc -no-shutdown
> -global PIIX4_PM.disable_s3=1 -global PIIX4_PM.disable_s4=1 -boot
> menu=off,strict=on -device
> virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -drive
> file=/var/tmp/localvmdRIozH/images/15023c8a-e3a7-4851-a97d-3b90996b423b/07fdcff3-11ce-4f7c-af05-0a878593e78e,format=qcow2,if=none,id=drive-virtio-disk0
> -device
> virtio-blk-pci,scsi=off,bus=pci.0,addr=0x5,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1
> -drive
> file=/var/tmp/localvmdRIozH/seed.iso,format=raw,if=none,id=drive-ide0-0-0,readonly=on
> -device ide-cd,bus=ide.0,unit=0,drive=drive-ide0-0-0,id=ide0-0-0 -netdev
> tap,fd=29,id=hostnet0,vhost=on,vhostfd=30 -device
> virtio-net-pci,netdev=hostnet0,id=net0,mac=00:16:3e:6c:e8:f9,bus=pci.0,addr=0x3
> -chardev pty,id=charserial0 -device
> isa-serial,chardev=charserial0,id=serial0 -chardev
> socket,id=charchannel0,fd=31,server,nowait -device
> virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=org.qemu.guest_agent.0
> -vnc 127.0.0.1:0 -device VGA,id=video0,vgamem_mb=16,bus=pci.0,addr=0x2
> -object rng-random,id=objrng0,filename=/dev/random -device
> virtio-rng-pci,rng=objrng0,id=rng0,bus=pci.0,addr=0x6 -sandbox
> on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny
> -msg timestamp=on
> root     28416  0.0  0.0 112712   980 pts/3    S+   16:14   0:00 grep
> --color=auto qemu
>
>
> Before the first Error, while the vm was running for sure and the disk was
> there, I also unsuccessfuly tried to connect to it with VNC  and got the
> same error I got before :
>
> [root@vs-inf-int-kvm-fr-301-210 ~]# debug1: Connection to port 5900
> forwarding to vs-inf-int-kvm-fr-301-210.hostics.fr port 5900 requested.
> debug1: channel 3: new [direct-tcpip]
> channel 3: open failed: connect failed: Connection refused
> debug1: channel 3: free: direct-tcpip: listening port 5900 for
> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900, connect from 127.0.0.1
> port 37002 to 127.0.0.1 port 5900, nchannels 4
>
>
> Guillaume Pavese
> Ingénieur Système et Réseau
> Interactiv-Group
>
>
> On Mon, Feb 25, 2019 at 11:57 PM Guillaume Pavese <
> guillaume.pav...@interactiv-group.com> wrote:
>
>> Something was definitely wrong ; as indicated, qemu process
>> for guest=HostedEngineLocal was running but the disk file did not exist
>> anymore...
>> No surprise I could not connect
>>
>> I am retrying
>>
>>
>> Guillaume Pavese
>> Ingénieur Système et Réseau
>> Interactiv-Group
>>
>>
>> On Mon, Feb 25, 2019 at 11:15 PM Guillaume Pavese <
>> guillaume.pav...@interactiv-group.com> wrote:
>>
>>> It fails too :
>>> I made sure PermitTunnel=yes in sshd config but when I try to connect to
>>> the forwarded port I get the following error on the openened host ssh
>>> session :
>>>
>>> [gpavese@sheepora-X230 ~]$ ssh -v -L 5900:
>>> vs-inf-int-kvm-fr-301-210.hostics.fr:5900
>>> r...@vs-inf-int-kvm-fr-301-210.hostics.fr
>>> ...
>>> [root@vs-inf-int-kvm-fr-301-210 ~]#
>>> debug1: channel 3: free: direct-tcpip: listening port 5900 for
>>> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900, connect from ::1 port
>>> 42144 to ::1 port 5900, nchannels 4
>>> debug1: Connection to port 5900 forwarding to
>>> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900 requested.
>>> debug1: channel 3: new [direct-tcpip]
>>> channel 3: open failed: connect failed: Connection refused
>>> debug1: channel 3: free: direct-tcpip: listening port 5900 for
>>> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900, connect from 127.0.0.1
>>> port 32778 to 127.0.0.1 port 5900, nchannels 4
>>>
>>>
>>> and in journalctl :
>>>
>>> févr. 25 14:55:38 vs-inf-int-kvm-fr-301-210.hostics.fr sshd[19595]:
>>> error: connect_to vs-inf-int-kvm-fr-301-210.hostics.fr port 5900:
>>> failed.
>>>
>>>
>>> Guillaume Pavese
>>> Ingénieur Système et Réseau
>>> Interactiv-Group
>>>
>>>
>>> On Mon, Feb 25, 2019 at 10:44 PM Simone Tiraboschi <stira...@redhat.com>
>>> wrote:
>>>
>>>>
>>>>
>>>>
>>>> On Mon, Feb 25, 2019 at 2:35 PM Guillaume Pavese <
>>>> guillaume.pav...@interactiv-group.com> wrote:
>>>>
>>>>> I made sure of everything and even stopped firewalld but still can't
>>>>> connect :
>>>>>
>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# cat
>>>>> /var/run/libvirt/qemu/HostedEngineLocal.xml
>>>>>      <graphics type='vnc' port='*5900*' autoport='yes'
>>>>> *listen='127.0.0.1*'>
>>>>>         <listen type='address' address='*127.0.0.1*' fromConfig='1'
>>>>> autoGenerated='no'/>
>>>>>
>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# netstat -pan | grep 59
>>>>> tcp        0      0 127.0.0.1:5900          0.0.0.0:*
>>>>>  LISTEN      13376/qemu-kvm
>>>>>
>>>>
>>>>
>>>> I suggest to try ssh tunneling, run
>>>> ssh -L 5900:vs-inf-int-kvm-fr-301-210.hostics.fr:5900
>>>> r...@vs-inf-int-kvm-fr-301-210.hostics.fr
>>>>
>>>> and then
>>>> remote-viewer vnc://localhost:5900
>>>>
>>>>
>>>>
>>>>>
>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]# systemctl status
>>>>> firewalld.service
>>>>> ● firewalld.service - firewalld - dynamic firewall daemon
>>>>>    Loaded: loaded (/usr/lib/systemd/system/firewalld.service; enabled;
>>>>> vendor preset: enabled)
>>>>>    *Active: inactive (dead)*
>>>>> *févr. 25 14:24:03 vs-inf-int-kvm-fr-301-210.hostics.fr
>>>>> <http://vs-inf-int-kvm-fr-301-210.hostics.fr> systemd[1]: Stopped 
>>>>> firewalld
>>>>> - dynamic firewall daemon.*
>>>>>
>>>>> From my laptop :
>>>>> [gpavese@sheepora-X230 ~]$ telnet vs-inf-int-kvm-fr-301-210.hostics.fr
>>>>> *5900*
>>>>> Trying 10.199.210.11...
>>>>> [*nothing gets through...*]
>>>>> ^C
>>>>>
>>>>> For making sure :
>>>>> [gpavese@sheepora-X230 ~]$ telnet vs-inf-int-kvm-fr-301-210.hostics.fr
>>>>> *9090*
>>>>> Trying 10.199.210.11...
>>>>> *Connected* to vs-inf-int-kvm-fr-301-210.hostics.fr.
>>>>> Escape character is '^]'.
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> Guillaume Pavese
>>>>> Ingénieur Système et Réseau
>>>>> Interactiv-Group
>>>>>
>>>>>
>>>>> On Mon, Feb 25, 2019 at 10:24 PM Parth Dhanjal <dpa...@redhat.com>
>>>>> wrote:
>>>>>
>>>>>> Hey!
>>>>>>
>>>>>> You can check under /var/run/libvirt/qemu/HostedEngine.xml
>>>>>> Search for 'vnc'
>>>>>> From there you can look up the port on which the HE VM is available
>>>>>> and connect to the same.
>>>>>>
>>>>>>
>>>>>> On Mon, Feb 25, 2019 at 6:47 PM Guillaume Pavese <
>>>>>> guillaume.pav...@interactiv-group.com> wrote:
>>>>>>
>>>>>>> 1) I am running in a Nested env, but under libvirt/kvm on remote
>>>>>>> Centos 7.4 Hosts
>>>>>>>
>>>>>>> Please advise how to connect with VNC to the local HE vm. I see it's
>>>>>>> running, but this is on a remote host, not my local machine :
>>>>>>> qemu     13376  100  3.7 17679424 845216 ?     Sl   12:46  85:08
>>>>>>> /usr/libexec/qemu-kvm -name guest=HostedEngineLocal,debug-threads=on -S
>>>>>>> -object
>>>>>>> secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/domain-1-HostedEngineLocal/master-key.aes
>>>>>>> -machine pc-i440fx-rhel7.6.0,accel=kvm,usb=off,dump-guest-core=off -cpu
>>>>>>> Haswell-noTSX,+kvmclock -m 16384 -realtime mlock=off -smp
>>>>>>> 4,sockets=4,cores=1,threads=1 -uuid 6fe7c1c3-ea93-4343-a385-0d9e14bb563a
>>>>>>> -no-user-config -nodefaults -chardev
>>>>>>> socket,id=charmonitor,fd=27,server,nowait -mon
>>>>>>> chardev=charmonitor,id=monitor,mode=control -rtc base=utc -no-shutdown
>>>>>>> -global PIIX4_PM.disable_s3=1 -global PIIX4_PM.disable_s4=1 -boot
>>>>>>> menu=off,strict=on -device
>>>>>>> virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -drive
>>>>>>> file=/var/tmp/localvmgmyYik/images/15023c8a-e3a7-4851-a97d-3b90996b423b/07fdcff3-11ce-4f7c-af05-0a878593e78e,format=qcow2,if=none,id=drive-virtio-disk0
>>>>>>> -device
>>>>>>> virtio-blk-pci,scsi=off,bus=pci.0,addr=0x5,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1
>>>>>>> -drive
>>>>>>> file=/var/tmp/localvmgmyYik/seed.iso,format=raw,if=none,id=drive-ide0-0-0,readonly=on
>>>>>>> -device ide-cd,bus=ide.0,unit=0,drive=drive-ide0-0-0,id=ide0-0-0 -netdev
>>>>>>> tap,fd=29,id=hostnet0,vhost=on,vhostfd=30 -device
>>>>>>> virtio-net-pci,netdev=hostnet0,id=net0,mac=00:16:3e:3e:fe:28,bus=pci.0,addr=0x3
>>>>>>> -chardev pty,id=charserial0 -device
>>>>>>> isa-serial,chardev=charserial0,id=serial0 -chardev
>>>>>>> socket,id=charchannel0,fd=31,server,nowait -device
>>>>>>> virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=org.qemu.guest_agent.0
>>>>>>> *-vnc 127.0.0.1:0 <http://127.0.0.1:0> -device 
>>>>>>> VGA*,id=video0,vgamem_mb=16,bus=pci.0,addr=0x2
>>>>>>> -object rng-random,id=objrng0,filename=/dev/random -device
>>>>>>> virtio-rng-pci,rng=objrng0,id=rng0,bus=pci.0,addr=0x6 -sandbox
>>>>>>> on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny
>>>>>>> -msg timestamp=on
>>>>>>>
>>>>>>>
>>>>>>> 2) [root@vs-inf-int-kvm-fr-301-210 ~]# cat
>>>>>>> /etc/libvirt/qemu/networks/default.xml
>>>>>>> <!--
>>>>>>> WARNING: THIS IS AN AUTO-GENERATED FILE. CHANGES TO IT ARE LIKELY TO
>>>>>>> BE
>>>>>>> OVERWRITTEN AND LOST. Changes to this xml configuration should be
>>>>>>> made using:
>>>>>>>   virsh net-edit default
>>>>>>> or other application using the libvirt API.
>>>>>>> -->
>>>>>>>
>>>>>>> <network>
>>>>>>>   <name>default</name>
>>>>>>>   <uuid>ba7bbfc8-28b8-459e-a42d-c2d6218e2cb6</uuid>
>>>>>>>   <forward mode='nat'/>
>>>>>>>   <bridge name='virbr0' stp='on' delay='0'/>
>>>>>>>   <mac address='52:54:00:e5:fe:3b'/>
>>>>>>>   <ip address='192.168.122.1' netmask='255.255.255.0'>
>>>>>>>     <dhcp>
>>>>>>>       <range start='192.168.122.2' end='192.168.122.254'/>
>>>>>>>     </dhcp>
>>>>>>>   </ip>
>>>>>>> </network>
>>>>>>> You have new mail in /var/spool/mail/root
>>>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Guillaume Pavese
>>>>>>> Ingénieur Système et Réseau
>>>>>>> Interactiv-Group
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Feb 25, 2019 at 9:57 PM Simone Tiraboschi <
>>>>>>> stira...@redhat.com> wrote:
>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Feb 25, 2019 at 1:14 PM Guillaume Pavese <
>>>>>>>> guillaume.pav...@interactiv-group.com> wrote:
>>>>>>>>
>>>>>>>>> He deployment with "hosted-engine --deploy" fails at TASK
>>>>>>>>> [ovirt.hosted_engine_setup : Get local VM IP]
>>>>>>>>>
>>>>>>>>> See following Error :
>>>>>>>>>
>>>>>>>>> 2019-02-25 12:46:50,154+0100 INFO
>>>>>>>>> otopi.ovirt_hosted_engine_setup.ansible_utils
>>>>>>>>> ansible_utils._process_output:109 TASK [ovirt.hosted_engine_setup : 
>>>>>>>>> Get
>>>>>>>>> local VM IP]
>>>>>>>>> 2019-02-25 12:55:26,823+0100 DEBUG
>>>>>>>>> otopi.ovirt_hosted_engine_setup.ansible_utils
>>>>>>>>> ansible_utils._process_output:103 {u'_ansible_parsed': True,
>>>>>>>>> u'stderr_lines': [], u'cmd': u"virsh -r net-dhcp-leases default | 
>>>>>>>>> grep -i 00
>>>>>>>>> :16:3e:3e:fe:28 | awk '{ print $5 }' | cut -f1 -d'/'", u'end':
>>>>>>>>> u'2019-02-25 12:55:26.666925', u'_ansible_no_log': False, u'stdout': 
>>>>>>>>> u'',
>>>>>>>>> u'changed': True, u'invocation': {u'module_args': {u'warn': True,
>>>>>>>>> u'executable':
>>>>>>>>> None, u'_uses_shell': True, u'_raw_params': u"virsh -r
>>>>>>>>> net-dhcp-leases default | grep -i 00:16:3e:3e:fe:28 | awk '{ print $5 
>>>>>>>>> }' |
>>>>>>>>> cut -f1 -d'/'", u'removes': None, u'argv': None, u'creates': None,
>>>>>>>>> u'chdir': None, u'std
>>>>>>>>> in': None}}, u'start': u'2019-02-25 12:55:26.584686', u'attempts':
>>>>>>>>> 50, u'stderr': u'', u'rc': 0, u'delta': u'0:00:00.082239', 
>>>>>>>>> u'stdout_lines':
>>>>>>>>> []}
>>>>>>>>> 2019-02-25 12:55:26,924+0100 ERROR
>>>>>>>>> otopi.ovirt_hosted_engine_setup.ansible_utils
>>>>>>>>> ansible_utils._process_output:107 fatal: [localhost]: FAILED! =>
>>>>>>>>> {"attempts": 50, "changed": true, "cmd": "virsh -r net-dhcp-leases 
>>>>>>>>> default
>>>>>>>>> | grep -i 00:16:3e:3e:fe:28 | awk '{ print $5 }' | cut -f1 -d'/'", 
>>>>>>>>> "delta":
>>>>>>>>> "0:00:00.082239", "end": "2019-02-25 12:55:26.666925", "rc": 0, 
>>>>>>>>> "start":
>>>>>>>>> "2019-02-25 12:55:26.584686", "stderr": "", "stderr_lines": [], 
>>>>>>>>> "stdout":
>>>>>>>>> "", "stdout_lines": []}
>>>>>>>>>
>>>>>>>>
>>>>>>>> Here we are just waiting for the bootstrap engine VM to fetch an IP
>>>>>>>> address from default libvirt network over DHCP but it your case it 
>>>>>>>> never
>>>>>>>> happened.
>>>>>>>> Possible issues: something went wrong in the bootstrap process for
>>>>>>>> the engine VM or the default libvirt network is not correctly 
>>>>>>>> configured.
>>>>>>>>
>>>>>>>> 1. can you try to reach the engine VM via VNC and check what's
>>>>>>>> happening there? (another question, are you running it nested? AFAIK it
>>>>>>>> will not work if nested over ESXi)
>>>>>>>> 2. can you please share the output of
>>>>>>>> cat /etc/libvirt/qemu/networks/default.xml
>>>>>>>>
>>>>>>>>
>>>>>>>>>
>>>>>>>>> Guillaume Pavese
>>>>>>>>> Ingénieur Système et Réseau
>>>>>>>>> Interactiv-Group
>>>>>>>>> _______________________________________________
>>>>>>>>> Users mailing list -- users@ovirt.org
>>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>>>>>> oVirt Code of Conduct:
>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>>>>>> List Archives:
>>>>>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/VXRMU3SQWTMB2YYNMOMD7I5NX7RZQ2IW/
>>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>> Users mailing list -- users@ovirt.org
>>>>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>>>> oVirt Code of Conduct:
>>>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>>>> List Archives:
>>>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/45UR44ITQTV7YVE6GLJWEV2AV22TIGXK/
>>>>>>>
>>>>>>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/RB5STPI7G7WTEX34K4IMPR2Z57CFZGKO/

Reply via email to