On Mon, Jul 25, 2016 at 2:15 PM, <aleksey.maksi...@it-kb.ru> wrote: > # ss -plutn > > Netid State Recv-Q Send-Q > Local Address:Port Peer Address:Port > > udp UNCONN 0 0 > *:111 > *:* > users:(("rpcbind",pid=827,fd=6)) > udp UNCONN 0 0 > *:161 > *:* > users:(("snmpd",pid=1609,fd=6)) > udp UNCONN 0 0 > 127.0.0.1:323 > *:* > users:(("chronyd",pid=795,fd=1)) > udp UNCONN 0 0 > *:959 > *:* > users:(("rpcbind",pid=827,fd=7)) > udp UNCONN 0 0 > 127.0.0.1:25375 > *:* > users:(("snmpd",pid=1609,fd=8)) > udp UNCONN 0 0 > 127.0.0.1:25376 > *:* > users:(("cmapeerd",pid=2056,fd=5)) > udp UNCONN 0 0 > 127.0.0.1:25393 > *:* > users:(("cmanicd",pid=2278,fd=3)) > udp UNCONN 0 0 > :::111 > :::* > users:(("rpcbind",pid=827,fd=9)) > udp UNCONN 0 0 > :::959 > :::* > users:(("rpcbind",pid=827,fd=10)) > tcp LISTEN 0 128 > *:2381 > *:* > > users:(("hpsmhd",pid=3903,fd=4),("hpsmhd",pid=3901,fd=4),("hpsmhd",pid=3900,fd=4),("hpsmhd",pid=3899,fd=4),("hpsmhd",pid=3898,fd=4),("hpsmhd",pid=3893,fd=4)) > tcp LISTEN 0 128 > *:111 > *:* > users:(("rpcbind",pid=827,fd=8)) > tcp LISTEN 0 5 > *:54322 > *:* > users:(("ovirt-imageio-d",pid=753,fd=3)) > tcp LISTEN 0 128 > *:22 > *:* > users:(("sshd",pid=1606,fd=3)) > tcp LISTEN 0 100 > 127.0.0.1:25 > *:* > users:(("master",pid=1948,fd=13)) > tcp LISTEN 0 128 > *:2301 > *:* > > users:(("hpsmhd",pid=3903,fd=3),("hpsmhd",pid=3901,fd=3),("hpsmhd",pid=3900,fd=3),("hpsmhd",pid=3899,fd=3),("hpsmhd",pid=3898,fd=3),("hpsmhd",pid=3893,fd=3)) > tcp LISTEN 0 30 > *:16514 > *:* > users:(("libvirtd",pid=10688,fd=13)) > tcp LISTEN 0 128 > 127.0.0.1:199 > *:* > users:(("snmpd",pid=1609,fd=9)) > tcp LISTEN 0 128 > :::111 > :::* > users:(("rpcbind",pid=827,fd=11)) > tcp LISTEN 0 5 > :::54321 > :::* > users:(("vdsm",pid=11077,fd=23))
vdsm is properly bind over ipv6. Can you please check if you can connect to vdsm with: telnet kom-ad01-vm31.holding.com 54321 and with telnet ::1 54321 ? > tcp LISTEN 0 30 > :::16514 > :::* > users:(("libvirtd",pid=10688,fd=14)) > > 25.07.2016, 15:11, "Simone Tiraboschi" <stira...@redhat.com>: >> On Mon, Jul 25, 2016 at 2:03 PM, <aleksey.maksi...@it-kb.ru> wrote: >>> Yes. >>> >>> # ping $(python -c 'import socket; print(socket.gethostname())') >>> >>> PING KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data. >>> 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1 ttl=64 >>> time=0.030 ms >>> 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2 ttl=64 >>> time=0.015 ms >>> 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3 ttl=64 >>> time=0.011 ms >>> ^C >>> --- KOM-AD01-VM31.holding.com ping statistics --- >>> 3 packets transmitted, 3 received, 0% packet loss, time 2001ms >>> rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms >>> >>> but... >> >> and the output of >> ss -plutn >> >>> # vdsClient -s 0 getVdsCaps >>> >>> Traceback (most recent call last): >>> File "/usr/share/vdsm/vdsClient.py", line 2980, in <module> >>> code, message = commands[command][0](commandArgs) >>> File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap >>> return self.ExecAndExit(self.s.getVdsCapabilities()) >>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__ >>> return self.__send(self.__name, args) >>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request >>> verbose=self.__verbose >>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request >>> return self.single_request(host, handler, request_body, verbose) >>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request >>> self.send_content(h, request_body) >>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content >>> connection.endheaders(request_body) >>> File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders >>> self._send_output(message_body) >>> File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output >>> self.send(msg) >>> File "/usr/lib64/python2.7/httplib.py", line 797, in send >>> self.connect() >>> File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in >>> connect >>> sock = socket.create_connection((self.host, self.port), self.timeout) >>> File "/usr/lib64/python2.7/socket.py", line 571, in create_connection >>> raise err >>> error: [Errno 101] Network is unreachable >>> >>> 25.07.2016, 14:58, "aleksey.maksi...@it-kb.ru" <aleksey.maksi...@it-kb.ru>: >>>> Ok. >>>> >>>> 1) I stopped and disabled the service NetworkManager >>>> # systemctl stop NetworkManager >>>> # systemctl disable NetworkManager >>>> >>>> 2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add PEERDNS=no in >>>> ifcfg-* file. >>>> >>>> 3) Reboot server >>>> >>>> 4) Try deploy oVirt HE 4 and I get the same error >>>> >>>> [ INFO ] Creating Volume Group >>>> [ INFO ] Creating Storage Domain >>>> [ INFO ] Creating Storage Pool >>>> [ INFO ] Connecting Storage Pool >>>> [ INFO ] Verifying sanlock lockspace initialization >>>> [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] >>>> Network is unreachable >>>> [ INFO ] Stage: Clean up >>>> [ INFO ] Generating answer file >>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf' >>>> [ INFO ] Stage: Pre-termination >>>> [ INFO ] Stage: Termination >>>> [ ERROR ] Hosted Engine deployment failed: this system is not reliable, >>>> please check the issue, fix and redeploy >>>> Log file is located at >>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725142534-t81kwf.log >>>> >>>> What ideas further? >>>> >>>> 25.07.2016, 13:06, "Simone Tiraboschi" <stira...@redhat.com>: >>>>> On Mon, Jul 25, 2016 at 11:54 AM, <aleksey.maksi...@it-kb.ru> wrote: >>>>>> What am I supposed to do for successfully deploy ovirt 4 ? >>>>>> Any ideas ? >>>>> >>>>> Can you please try to explicitly configure your DNS with nameserver >>>>> under /etc/resolv.conf and remove DNS1 and DNS2 and set PEERDNS=no for >>>>> the interface you are going to use? >>>>> >>>>>> 25.07.2016, 12:47, "aleksey.maksi...@it-kb.ru" >>>>>> <aleksey.maksi...@it-kb.ru>: >>>>>>> "Aleksey, was your DNS configured with DNS1 and DNS2 just on the >>>>>>> interface you used to create the management bridge on?" >>>>>>> >>>>>>> Yes. Of course >>>>>>> >>>>>>> 25.07.2016, 12:27, "Simone Tiraboschi" <stira...@redhat.com>: >>>>>>>> On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski >>>>>>>> <piotr.kliczew...@gmail.com> wrote: >>>>>>>>> This could be the issue here as well as for BZ #1358530 >>>>>>>>> >>>>>>>>> On Mon, Jul 25, 2016 at 10:53 AM, <aleksey.maksi...@it-kb.ru> >>>>>>>>> wrote: >>>>>>>>>> Could this be due to the fact that the ovirt installer has >>>>>>>>>> changed network configuration files (ifcfg-*, resolv.conf) ? >>>>>>>>>> After the error in ovirt installation process I see from >>>>>>>>>> resolv.conf disappeared on my DNS servers entry and now the server >>>>>>>>>> is unable to resolve names. >>>>>>>> >>>>>>>> So it could be related to >>>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1160423 >>>>>>>> >>>>>>>> Aleksey, was your DNS configured with DNS1 and DNS2 just on the >>>>>>>> interface you used to create the management bridge on? >>>>>>>> Can you please try the workaround described here >>>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ? >>>>>>>> >>>>>>>>>> 25.07.2016, 11:26, "Simone Tiraboschi" <stira...@redhat.com>: >>>>>>>>>>> On Mon, Jul 25, 2016 at 10:22 AM, <aleksey.maksi...@it-kb.ru> >>>>>>>>>>> wrote: >>>>>>>>>>>> # vdsClient -s 0 getVdsCaps >>>>>>>>>>>> >>>>>>>>>>>> Traceback (most recent call last): >>>>>>>>>>>> File "/usr/share/vdsm/vdsClient.py", line 2980, in <module> >>>>>>>>>>>> code, message = commands[command][0](commandArgs) >>>>>>>>>>>> File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap >>>>>>>>>>>> return self.ExecAndExit(self.s.getVdsCapabilities()) >>>>>>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in >>>>>>>>>>>> __call__ >>>>>>>>>>>> return self.__send(self.__name, args) >>>>>>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in >>>>>>>>>>>> __request >>>>>>>>>>>> verbose=self.__verbose >>>>>>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in >>>>>>>>>>>> request >>>>>>>>>>>> return self.single_request(host, handler, request_body, >>>>>>>>>>>> verbose) >>>>>>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in >>>>>>>>>>>> single_request >>>>>>>>>>>> self.send_content(h, request_body) >>>>>>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in >>>>>>>>>>>> send_content >>>>>>>>>>>> connection.endheaders(request_body) >>>>>>>>>>>> File "/usr/lib64/python2.7/httplib.py", line 975, in >>>>>>>>>>>> endheaders >>>>>>>>>>>> self._send_output(message_body) >>>>>>>>>>>> File "/usr/lib64/python2.7/httplib.py", line 835, in >>>>>>>>>>>> _send_output >>>>>>>>>>>> self.send(msg) >>>>>>>>>>>> File "/usr/lib64/python2.7/httplib.py", line 797, in send >>>>>>>>>>>> self.connect() >>>>>>>>>>>> File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", >>>>>>>>>>>> line 203, in connect >>>>>>>>>>>> sock = socket.create_connection((self.host, self.port), >>>>>>>>>>>> self.timeout) >>>>>>>>>>>> File "/usr/lib64/python2.7/socket.py", line 571, in >>>>>>>>>>>> create_connection >>>>>>>>>>>> raise err >>>>>>>>>>>> error: [Errno 101] Network is unreachable >>>>>>>>>>> >>>>>>>>>>> Yaniv, can you please take also a look to this one? >>>>>>>>>>> it's exactly the opposite of >>>>>>>>>>> https://bugzilla.redhat.com/1358530 >>>>>>>>>>> Here the jsonrpcclient works but not the xmlrpc one. >>>>>>>>>>> >>>>>>>>>>>> 25.07.2016, 11:17, "Simone Tiraboschi" <stira...@redhat.com>: >>>>>>>>>>>>> On Mon, Jul 25, 2016 at 7:51 AM, >>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru> wrote: >>>>>>>>>>>>>> Simone, there is something interesting in the vdsm.log? >>>>>>>>>>>>> >>>>>>>>>>>>> For what I saw the issue is not related to the storage but >>>>>>>>>>>>> to the network. >>>>>>>>>>>>> ovirt-hosted-engine-setup uses the jsonrpc client, instead >>>>>>>>>>>>> the code >>>>>>>>>>>>> from ovirt-hosted-engine-ha still uses the xmlrpc client >>>>>>>>>>>>> somewhere and >>>>>>>>>>>>> this happens also when the setup asks to create the >>>>>>>>>>>>> lockspace volume. >>>>>>>>>>>>> It seams that in your case the xmlrpc client could not >>>>>>>>>>>>> connect vdsm on >>>>>>>>>>>>> the localhost. >>>>>>>>>>>>> It could be somehow related to: >>>>>>>>>>>>> https://bugzilla.redhat.com/1358530 >>>>>>>>>>>>> >>>>>>>>>>>>> Can you please try executing >>>>>>>>>>>>> sudo vdsClient -s 0 getVdsCaps >>>>>>>>>>>>> on that host? >>>>>>>>>>>>> >>>>>>>>>>>>>> 22.07.2016, 19:36, "aleksey.maksi...@it-kb.ru" >>>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru>: >>>>>>>>>>>>>>> Simone, thanks for link. >>>>>>>>>>>>>>> vdsm.log attached >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> 22.07.2016, 19:28, "Simone Tiraboschi" >>>>>>>>>>>>>>> <stira...@redhat.com>: >>>>>>>>>>>>>>>> On Fri, Jul 22, 2016 at 5:59 PM, >>>>>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru> wrote: >>>>>>>>>>>>>>>>> Thank you for your response, Simone. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Log attached. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> It seams it comes from VDSM, can you please attach >>>>>>>>>>>>>>>> also vdsm.log? >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> I don't use ovirt-engine-appliance because I have >>>>>>>>>>>>>>>>> not found "how-to" for ovirt-engine-appliance deployment in >>>>>>>>>>>>>>>>> hosted engine configuration. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> yum install ovirt-engine-appliance >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Then follow the instruction here: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> http://www.ovirt.org/develop/release-management/features/heapplianceflow/ >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> 22.07.2016, 17:09, "Simone Tiraboschi" >>>>>>>>>>>>>>>>> <stira...@redhat.com>: >>>>>>>>>>>>>>>>>> Hi Aleksey, >>>>>>>>>>>>>>>>>> Can you please attach hosted-engine-setup logs? >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> On Fri, Jul 22, 2016 at 3:46 PM, >>>>>>>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru> wrote: >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Hello oVirt guru`s ! >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> I have problem with initial deploy of ovirt 4.0 >>>>>>>>>>>>>>>>>>> hosted engine. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> My environment : >>>>>>>>>>>>>>>>>>> ============================ >>>>>>>>>>>>>>>>>>> * Two servers HP ProLiant DL 360 G5 with Qlogic >>>>>>>>>>>>>>>>>>> FC HBA connected (with >>>>>>>>>>>>>>>>>>> multipathd) to storage HP 3PAR 7200 >>>>>>>>>>>>>>>>>>> * On each server installed CentOS 7.2 Linux >>>>>>>>>>>>>>>>>>> (3.10.0-327.22.2.el7.x86_64) >>>>>>>>>>>>>>>>>>> * On 3PAR storage I created 2 LUNs for oVirt. >>>>>>>>>>>>>>>>>>> - First LUN for oVirt Hosted Engine VM (60GB) >>>>>>>>>>>>>>>>>>> - Second LUN for all other VMs (2TB) >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> # multipath -ll >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> 3par-vv1 (360002ac0000000000000001b0000cec9) dm-0 >>>>>>>>>>>>>>>>>>> 3PARdata,VV >>>>>>>>>>>>>>>>>>> size=60G features='1 queue_if_no_path' >>>>>>>>>>>>>>>>>>> hwhandler='1 alua' wp=rw >>>>>>>>>>>>>>>>>>> `-+- policy='round-robin 0' prio=50 status=active >>>>>>>>>>>>>>>>>>> |- 2:0:1:1 sdd 8:48 active ready running >>>>>>>>>>>>>>>>>>> |- 3:0:0:1 sdf 8:80 active ready running >>>>>>>>>>>>>>>>>>> |- 2:0:0:1 sdb 8:16 active ready running >>>>>>>>>>>>>>>>>>> `- 3:0:1:1 sdh 8:112 active ready running >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> 3par-vv2 (360002ac000000000000000160000cec9) dm-1 >>>>>>>>>>>>>>>>>>> 3PARdata,VV >>>>>>>>>>>>>>>>>>> size=2.0T features='1 queue_if_no_path' >>>>>>>>>>>>>>>>>>> hwhandler='1 alua' wp=rw >>>>>>>>>>>>>>>>>>> `-+- policy='round-robin 0' prio=50 status=active >>>>>>>>>>>>>>>>>>> |- 2:0:0:0 sda 8:0 active ready running >>>>>>>>>>>>>>>>>>> |- 3:0:0:0 sde 8:64 active ready running >>>>>>>>>>>>>>>>>>> |- 2:0:1:0 sdc 8:32 active ready running >>>>>>>>>>>>>>>>>>> `- 3:0:1:0 sdg 8:96 active ready running >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> My steps on first server (initial deploy of ovirt >>>>>>>>>>>>>>>>>>> 4.0 hosted engine): >>>>>>>>>>>>>>>>>>> ============================ >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> # systemctl stop NetworkManager >>>>>>>>>>>>>>>>>>> # systemctl disable NetworkManager >>>>>>>>>>>>>>>>>>> # yum -y install >>>>>>>>>>>>>>>>>>> http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm >>>>>>>>>>>>>>>>>>> # yum -y install epel-release >>>>>>>>>>>>>>>>>>> # wget >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1511.iso >>>>>>>>>>>>>>>>>>> -P /tmp/ >>>>>>>>>>>>>>>>>>> # yum install ovirt-hosted-engine-setup >>>>>>>>>>>>>>>>>>> # yum install screen >>>>>>>>>>>>>>>>>>> # screen -RD >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> ...in screen session : >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> # hosted-engine --deploy >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>>>>>> in configuration process I chose "fc" as storage >>>>>>>>>>>>>>>>>>> type for oVirt hosted >>>>>>>>>>>>>>>>>>> engine vm and select 60GB LUN... >>>>>>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> --== CONFIGURATION PREVIEW ==-- >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>>>>>> Firewall manager : iptables >>>>>>>>>>>>>>>>>>> Gateway address : 10.1.0.1 >>>>>>>>>>>>>>>>>>> Host name for web application : >>>>>>>>>>>>>>>>>>> KOM-AD01-OVIRT1 >>>>>>>>>>>>>>>>>>> Storage Domain type : fc >>>>>>>>>>>>>>>>>>> Host ID : 1 >>>>>>>>>>>>>>>>>>> LUN ID : >>>>>>>>>>>>>>>>>>> 360002ac0000000000000001b0000cec9 >>>>>>>>>>>>>>>>>>> Image size GB : 40 >>>>>>>>>>>>>>>>>>> Console type : vnc >>>>>>>>>>>>>>>>>>> Memory size MB : 4096 >>>>>>>>>>>>>>>>>>> MAC address : 00:16:3e:77:1d:07 >>>>>>>>>>>>>>>>>>> Boot type : cdrom >>>>>>>>>>>>>>>>>>> Number of CPUs : 2 >>>>>>>>>>>>>>>>>>> ISO image (cdrom boot/cloud-init) : >>>>>>>>>>>>>>>>>>> /tmp/CentOS-7-x86_64-NetInstall-1511.iso >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Can I ask why you prefer/need to manually create a >>>>>>>>>>>>>>>>>> VM installing from >>>>>>>>>>>>>>>>>> a CD instead of using the ready-to-use >>>>>>>>>>>>>>>>>> ovirt-engine-appliance? >>>>>>>>>>>>>>>>>> Using the appliance makes the setup process a lot >>>>>>>>>>>>>>>>>> shorted and more comfortable. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> CPU Type : model_Penryn >>>>>>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>>>>>> and get error after step "Verifying sanlock >>>>>>>>>>>>>>>>>>> lockspace initialization" >>>>>>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> [ INFO ] Verifying sanlock lockspace >>>>>>>>>>>>>>>>>>> initialization >>>>>>>>>>>>>>>>>>> [ ERROR ] Failed to execute stage 'Misc >>>>>>>>>>>>>>>>>>> configuration': [Errno 101] Network >>>>>>>>>>>>>>>>>>> is unreachable >>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Clean up >>>>>>>>>>>>>>>>>>> [ INFO ] Generating answer file >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf' >>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Pre-termination >>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Termination >>>>>>>>>>>>>>>>>>> [ ERROR ] Hosted Engine deployment failed: this >>>>>>>>>>>>>>>>>>> system is not reliable, >>>>>>>>>>>>>>>>>>> please check the issue, fix and redeploy >>>>>>>>>>>>>>>>>>> Log file is located at >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160722123404-t26vw0.log >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Interestingly >>>>>>>>>>>>>>>>>>> ============================ >>>>>>>>>>>>>>>>>>> If I try to deploy hosted-engine v3.6, everything >>>>>>>>>>>>>>>>>>> goes well in the same >>>>>>>>>>>>>>>>>>> configuration !! : >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> .... >>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Transaction setup >>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Misc configuration >>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Package installation >>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Misc configuration >>>>>>>>>>>>>>>>>>> [ INFO ] Configuring libvirt >>>>>>>>>>>>>>>>>>> [ INFO ] Configuring VDSM >>>>>>>>>>>>>>>>>>> [ INFO ] Starting vdsmd >>>>>>>>>>>>>>>>>>> [ INFO ] Waiting for VDSM hardware info >>>>>>>>>>>>>>>>>>> [ INFO ] Configuring the management bridge >>>>>>>>>>>>>>>>>>> [ INFO ] Creating Volume Group >>>>>>>>>>>>>>>>>>> [ INFO ] Creating Storage Domain >>>>>>>>>>>>>>>>>>> [ INFO ] Creating Storage Pool >>>>>>>>>>>>>>>>>>> [ INFO ] Connecting Storage Pool >>>>>>>>>>>>>>>>>>> [ INFO ] Verifying sanlock lockspace >>>>>>>>>>>>>>>>>>> initialization >>>>>>>>>>>>>>>>>>> [ INFO ] Creating Image for >>>>>>>>>>>>>>>>>>> 'hosted-engine.lockspace' ... >>>>>>>>>>>>>>>>>>> [ INFO ] Image for 'hosted-engine.lockspace' >>>>>>>>>>>>>>>>>>> created successfully >>>>>>>>>>>>>>>>>>> [ INFO ] Creating Image for >>>>>>>>>>>>>>>>>>> 'hosted-engine.metadata' ... >>>>>>>>>>>>>>>>>>> [ INFO ] Image for 'hosted-engine.metadata' >>>>>>>>>>>>>>>>>>> created successfully >>>>>>>>>>>>>>>>>>> [ INFO ] Creating VM Image >>>>>>>>>>>>>>>>>>> [ INFO ] Destroying Storage Pool >>>>>>>>>>>>>>>>>>> [ INFO ] Start monitoring domain >>>>>>>>>>>>>>>>>>> [ INFO ] Configuring VM >>>>>>>>>>>>>>>>>>> [ INFO ] Updating hosted-engine configuration >>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Transaction commit >>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Closing up >>>>>>>>>>>>>>>>>>> [ INFO ] Creating VM >>>>>>>>>>>>>>>>>>> You can now connect to the VM with the >>>>>>>>>>>>>>>>>>> following command: >>>>>>>>>>>>>>>>>>> /bin/remote-viewer >>>>>>>>>>>>>>>>>>> vnc://localhost:5900 >>>>>>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> What could be the problem? >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>>> Users mailing list >>>>>>>>>>>>>>>>>>> Users@ovirt.org >>>>>>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>> _______________________________________________ >>>>>>>>>> Users mailing list >>>>>>>>>> Users@ovirt.org >>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>> >>>>>>> _______________________________________________ >>>>>>> Users mailing list >>>>>>> Users@ovirt.org >>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>> >>>> _______________________________________________ >>>> Users mailing list >>>> Users@ovirt.org >>>> http://lists.ovirt.org/mailman/listinfo/users _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users