telnet kom-ad01-vm31.holding.com 54321 = success connection telnet ::1 54321 Trying ::1... telnet: connect to address ::1: Network is unreachable (ipv6 on my server disabled)
25.07.2016, 15:35, "Simone Tiraboschi" <stira...@redhat.com>: > On Mon, Jul 25, 2016 at 2:15 PM, <aleksey.maksi...@it-kb.ru> wrote: >> # ss -plutn >> >> Netid State Recv-Q Send-Q Local Address:Port Peer Address:Port >> >> udp UNCONN 0 0 *:111 *:* users:(("rpcbind",pid=827,fd=6)) >> udp UNCONN 0 0 *:161 *:* users:(("snmpd",pid=1609,fd=6)) >> udp UNCONN 0 0 127.0.0.1:323 *:* users:(("chronyd",pid=795,fd=1)) >> udp UNCONN 0 0 *:959 *:* users:(("rpcbind",pid=827,fd=7)) >> udp UNCONN 0 0 127.0.0.1:25375 *:* users:(("snmpd",pid=1609,fd=8)) >> udp UNCONN 0 0 127.0.0.1:25376 *:* users:(("cmapeerd",pid=2056,fd=5)) >> udp UNCONN 0 0 127.0.0.1:25393 *:* users:(("cmanicd",pid=2278,fd=3)) >> udp UNCONN 0 0 :::111 :::* users:(("rpcbind",pid=827,fd=9)) >> udp UNCONN 0 0 :::959 :::* users:(("rpcbind",pid=827,fd=10)) >> tcp LISTEN 0 128 *:2381 *:* >> users:(("hpsmhd",pid=3903,fd=4),("hpsmhd",pid=3901,fd=4),("hpsmhd",pid=3900,fd=4),("hpsmhd",pid=3899,fd=4),("hpsmhd",pid=3898,fd=4),("hpsmhd",pid=3893,fd=4)) >> tcp LISTEN 0 128 *:111 *:* users:(("rpcbind",pid=827,fd=8)) >> tcp LISTEN 0 5 *:54322 *:* users:(("ovirt-imageio-d",pid=753,fd=3)) >> tcp LISTEN 0 128 *:22 *:* users:(("sshd",pid=1606,fd=3)) >> tcp LISTEN 0 100 127.0.0.1:25 *:* users:(("master",pid=1948,fd=13)) >> tcp LISTEN 0 128 *:2301 *:* >> users:(("hpsmhd",pid=3903,fd=3),("hpsmhd",pid=3901,fd=3),("hpsmhd",pid=3900,fd=3),("hpsmhd",pid=3899,fd=3),("hpsmhd",pid=3898,fd=3),("hpsmhd",pid=3893,fd=3)) >> tcp LISTEN 0 30 *:16514 *:* users:(("libvirtd",pid=10688,fd=13)) >> tcp LISTEN 0 128 127.0.0.1:199 *:* users:(("snmpd",pid=1609,fd=9)) >> tcp LISTEN 0 128 :::111 :::* users:(("rpcbind",pid=827,fd=11)) >> tcp LISTEN 0 5 :::54321 :::* users:(("vdsm",pid=11077,fd=23)) > > vdsm is properly bind over ipv6. > > Can you please check if you can connect to vdsm with: > telnet kom-ad01-vm31.holding.com 54321 > and with > telnet ::1 54321 > ? > >> tcp LISTEN 0 30 :::16514 :::* users:(("libvirtd",pid=10688,fd=14)) >> >> 25.07.2016, 15:11, "Simone Tiraboschi" <stira...@redhat.com>: >>> On Mon, Jul 25, 2016 at 2:03 PM, <aleksey.maksi...@it-kb.ru> wrote: >>>> Yes. >>>> >>>> # ping $(python -c 'import socket; print(socket.gethostname())') >>>> >>>> PING KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data. >>>> 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1 ttl=64 >>>> time=0.030 ms >>>> 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2 ttl=64 >>>> time=0.015 ms >>>> 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3 ttl=64 >>>> time=0.011 ms >>>> ^C >>>> --- KOM-AD01-VM31.holding.com ping statistics --- >>>> 3 packets transmitted, 3 received, 0% packet loss, time 2001ms >>>> rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms >>>> >>>> but... >>> >>> and the output of >>> ss -plutn >>> >>>> # vdsClient -s 0 getVdsCaps >>>> >>>> Traceback (most recent call last): >>>> File "/usr/share/vdsm/vdsClient.py", line 2980, in <module> >>>> code, message = commands[command][0](commandArgs) >>>> File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap >>>> return self.ExecAndExit(self.s.getVdsCapabilities()) >>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__ >>>> return self.__send(self.__name, args) >>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request >>>> verbose=self.__verbose >>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request >>>> return self.single_request(host, handler, request_body, verbose) >>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request >>>> self.send_content(h, request_body) >>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content >>>> connection.endheaders(request_body) >>>> File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders >>>> self._send_output(message_body) >>>> File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output >>>> self.send(msg) >>>> File "/usr/lib64/python2.7/httplib.py", line 797, in send >>>> self.connect() >>>> File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in >>>> connect >>>> sock = socket.create_connection((self.host, self.port), self.timeout) >>>> File "/usr/lib64/python2.7/socket.py", line 571, in create_connection >>>> raise err >>>> error: [Errno 101] Network is unreachable >>>> >>>> 25.07.2016, 14:58, "aleksey.maksi...@it-kb.ru" >>>> <aleksey.maksi...@it-kb.ru>: >>>>> Ok. >>>>> >>>>> 1) I stopped and disabled the service NetworkManager >>>>> # systemctl stop NetworkManager >>>>> # systemctl disable NetworkManager >>>>> >>>>> 2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add PEERDNS=no in >>>>> ifcfg-* file. >>>>> >>>>> 3) Reboot server >>>>> >>>>> 4) Try deploy oVirt HE 4 and I get the same error >>>>> >>>>> [ INFO ] Creating Volume Group >>>>> [ INFO ] Creating Storage Domain >>>>> [ INFO ] Creating Storage Pool >>>>> [ INFO ] Connecting Storage Pool >>>>> [ INFO ] Verifying sanlock lockspace initialization >>>>> [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] >>>>> Network is unreachable >>>>> [ INFO ] Stage: Clean up >>>>> [ INFO ] Generating answer file >>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf' >>>>> [ INFO ] Stage: Pre-termination >>>>> [ INFO ] Stage: Termination >>>>> [ ERROR ] Hosted Engine deployment failed: this system is not reliable, >>>>> please check the issue, fix and redeploy >>>>> Log file is located at >>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725142534-t81kwf.log >>>>> >>>>> What ideas further? >>>>> >>>>> 25.07.2016, 13:06, "Simone Tiraboschi" <stira...@redhat.com>: >>>>>> On Mon, Jul 25, 2016 at 11:54 AM, <aleksey.maksi...@it-kb.ru> wrote: >>>>>>> What am I supposed to do for successfully deploy ovirt 4 ? >>>>>>> Any ideas ? >>>>>> >>>>>> Can you please try to explicitly configure your DNS with nameserver >>>>>> under /etc/resolv.conf and remove DNS1 and DNS2 and set PEERDNS=no for >>>>>> the interface you are going to use? >>>>>> >>>>>>> 25.07.2016, 12:47, "aleksey.maksi...@it-kb.ru" >>>>>>> <aleksey.maksi...@it-kb.ru>: >>>>>>>> "Aleksey, was your DNS configured with DNS1 and DNS2 just on the >>>>>>>> interface you used to create the management bridge on?" >>>>>>>> >>>>>>>> Yes. Of course >>>>>>>> >>>>>>>> 25.07.2016, 12:27, "Simone Tiraboschi" <stira...@redhat.com>: >>>>>>>>> On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski >>>>>>>>> <piotr.kliczew...@gmail.com> wrote: >>>>>>>>>> This could be the issue here as well as for BZ #1358530 >>>>>>>>>> >>>>>>>>>> On Mon, Jul 25, 2016 at 10:53 AM, <aleksey.maksi...@it-kb.ru> >>>>>>>>>> wrote: >>>>>>>>>>> Could this be due to the fact that the ovirt installer has >>>>>>>>>>> changed network configuration files (ifcfg-*, resolv.conf) ? >>>>>>>>>>> After the error in ovirt installation process I see from >>>>>>>>>>> resolv.conf disappeared on my DNS servers entry and now the server >>>>>>>>>>> is unable to resolve names. >>>>>>>>> >>>>>>>>> So it could be related to >>>>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1160423 >>>>>>>>> >>>>>>>>> Aleksey, was your DNS configured with DNS1 and DNS2 just on the >>>>>>>>> interface you used to create the management bridge on? >>>>>>>>> Can you please try the workaround described here >>>>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ? >>>>>>>>> >>>>>>>>>>> 25.07.2016, 11:26, "Simone Tiraboschi" <stira...@redhat.com>: >>>>>>>>>>>> On Mon, Jul 25, 2016 at 10:22 AM, >>>>>>>>>>>> <aleksey.maksi...@it-kb.ru> wrote: >>>>>>>>>>>>> # vdsClient -s 0 getVdsCaps >>>>>>>>>>>>> >>>>>>>>>>>>> Traceback (most recent call last): >>>>>>>>>>>>> File "/usr/share/vdsm/vdsClient.py", line 2980, in >>>>>>>>>>>>> <module> >>>>>>>>>>>>> code, message = commands[command][0](commandArgs) >>>>>>>>>>>>> File "/usr/share/vdsm/vdsClient.py", line 543, in >>>>>>>>>>>>> do_getCap >>>>>>>>>>>>> return self.ExecAndExit(self.s.getVdsCapabilities()) >>>>>>>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in >>>>>>>>>>>>> __call__ >>>>>>>>>>>>> return self.__send(self.__name, args) >>>>>>>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in >>>>>>>>>>>>> __request >>>>>>>>>>>>> verbose=self.__verbose >>>>>>>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in >>>>>>>>>>>>> request >>>>>>>>>>>>> return self.single_request(host, handler, >>>>>>>>>>>>> request_body, verbose) >>>>>>>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in >>>>>>>>>>>>> single_request >>>>>>>>>>>>> self.send_content(h, request_body) >>>>>>>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in >>>>>>>>>>>>> send_content >>>>>>>>>>>>> connection.endheaders(request_body) >>>>>>>>>>>>> File "/usr/lib64/python2.7/httplib.py", line 975, in >>>>>>>>>>>>> endheaders >>>>>>>>>>>>> self._send_output(message_body) >>>>>>>>>>>>> File "/usr/lib64/python2.7/httplib.py", line 835, in >>>>>>>>>>>>> _send_output >>>>>>>>>>>>> self.send(msg) >>>>>>>>>>>>> File "/usr/lib64/python2.7/httplib.py", line 797, in send >>>>>>>>>>>>> self.connect() >>>>>>>>>>>>> File >>>>>>>>>>>>> "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in >>>>>>>>>>>>> connect >>>>>>>>>>>>> sock = socket.create_connection((self.host, >>>>>>>>>>>>> self.port), self.timeout) >>>>>>>>>>>>> File "/usr/lib64/python2.7/socket.py", line 571, in >>>>>>>>>>>>> create_connection >>>>>>>>>>>>> raise err >>>>>>>>>>>>> error: [Errno 101] Network is unreachable >>>>>>>>>>>> >>>>>>>>>>>> Yaniv, can you please take also a look to this one? >>>>>>>>>>>> it's exactly the opposite of >>>>>>>>>>>> https://bugzilla.redhat.com/1358530 >>>>>>>>>>>> Here the jsonrpcclient works but not the xmlrpc one. >>>>>>>>>>>> >>>>>>>>>>>>> 25.07.2016, 11:17, "Simone Tiraboschi" >>>>>>>>>>>>> <stira...@redhat.com>: >>>>>>>>>>>>>> On Mon, Jul 25, 2016 at 7:51 AM, >>>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru> wrote: >>>>>>>>>>>>>>> Simone, there is something interesting in the vdsm.log? >>>>>>>>>>>>>> >>>>>>>>>>>>>> For what I saw the issue is not related to the storage >>>>>>>>>>>>>> but to the network. >>>>>>>>>>>>>> ovirt-hosted-engine-setup uses the jsonrpc client, >>>>>>>>>>>>>> instead the code >>>>>>>>>>>>>> from ovirt-hosted-engine-ha still uses the xmlrpc client >>>>>>>>>>>>>> somewhere and >>>>>>>>>>>>>> this happens also when the setup asks to create the >>>>>>>>>>>>>> lockspace volume. >>>>>>>>>>>>>> It seams that in your case the xmlrpc client could not >>>>>>>>>>>>>> connect vdsm on >>>>>>>>>>>>>> the localhost. >>>>>>>>>>>>>> It could be somehow related to: >>>>>>>>>>>>>> https://bugzilla.redhat.com/1358530 >>>>>>>>>>>>>> >>>>>>>>>>>>>> Can you please try executing >>>>>>>>>>>>>> sudo vdsClient -s 0 getVdsCaps >>>>>>>>>>>>>> on that host? >>>>>>>>>>>>>> >>>>>>>>>>>>>>> 22.07.2016, 19:36, "aleksey.maksi...@it-kb.ru" >>>>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru>: >>>>>>>>>>>>>>>> Simone, thanks for link. >>>>>>>>>>>>>>>> vdsm.log attached >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> 22.07.2016, 19:28, "Simone Tiraboschi" >>>>>>>>>>>>>>>> <stira...@redhat.com>: >>>>>>>>>>>>>>>>> On Fri, Jul 22, 2016 at 5:59 PM, >>>>>>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru> wrote: >>>>>>>>>>>>>>>>>> Thank you for your response, Simone. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Log attached. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> It seams it comes from VDSM, can you please attach >>>>>>>>>>>>>>>>> also vdsm.log? >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> I don't use ovirt-engine-appliance because I have >>>>>>>>>>>>>>>>>> not found "how-to" for ovirt-engine-appliance deployment in >>>>>>>>>>>>>>>>>> hosted engine configuration. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> yum install ovirt-engine-appliance >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Then follow the instruction here: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> http://www.ovirt.org/develop/release-management/features/heapplianceflow/ >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> 22.07.2016, 17:09, "Simone Tiraboschi" >>>>>>>>>>>>>>>>>> <stira...@redhat.com>: >>>>>>>>>>>>>>>>>>> Hi Aleksey, >>>>>>>>>>>>>>>>>>> Can you please attach hosted-engine-setup logs? >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> On Fri, Jul 22, 2016 at 3:46 PM, >>>>>>>>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru> wrote: >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Hello oVirt guru`s ! >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> I have problem with initial deploy of ovirt 4.0 >>>>>>>>>>>>>>>>>>>> hosted engine. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> My environment : >>>>>>>>>>>>>>>>>>>> ============================ >>>>>>>>>>>>>>>>>>>> * Two servers HP ProLiant DL 360 G5 with Qlogic >>>>>>>>>>>>>>>>>>>> FC HBA connected (with >>>>>>>>>>>>>>>>>>>> multipathd) to storage HP 3PAR 7200 >>>>>>>>>>>>>>>>>>>> * On each server installed CentOS 7.2 Linux >>>>>>>>>>>>>>>>>>>> (3.10.0-327.22.2.el7.x86_64) >>>>>>>>>>>>>>>>>>>> * On 3PAR storage I created 2 LUNs for oVirt. >>>>>>>>>>>>>>>>>>>> - First LUN for oVirt Hosted Engine VM (60GB) >>>>>>>>>>>>>>>>>>>> - Second LUN for all other VMs (2TB) >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> # multipath -ll >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> 3par-vv1 (360002ac0000000000000001b0000cec9) >>>>>>>>>>>>>>>>>>>> dm-0 3PARdata,VV >>>>>>>>>>>>>>>>>>>> size=60G features='1 queue_if_no_path' >>>>>>>>>>>>>>>>>>>> hwhandler='1 alua' wp=rw >>>>>>>>>>>>>>>>>>>> `-+- policy='round-robin 0' prio=50 >>>>>>>>>>>>>>>>>>>> status=active >>>>>>>>>>>>>>>>>>>> |- 2:0:1:1 sdd 8:48 active ready running >>>>>>>>>>>>>>>>>>>> |- 3:0:0:1 sdf 8:80 active ready running >>>>>>>>>>>>>>>>>>>> |- 2:0:0:1 sdb 8:16 active ready running >>>>>>>>>>>>>>>>>>>> `- 3:0:1:1 sdh 8:112 active ready running >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> 3par-vv2 (360002ac000000000000000160000cec9) >>>>>>>>>>>>>>>>>>>> dm-1 3PARdata,VV >>>>>>>>>>>>>>>>>>>> size=2.0T features='1 queue_if_no_path' >>>>>>>>>>>>>>>>>>>> hwhandler='1 alua' wp=rw >>>>>>>>>>>>>>>>>>>> `-+- policy='round-robin 0' prio=50 >>>>>>>>>>>>>>>>>>>> status=active >>>>>>>>>>>>>>>>>>>> |- 2:0:0:0 sda 8:0 active ready running >>>>>>>>>>>>>>>>>>>> |- 3:0:0:0 sde 8:64 active ready running >>>>>>>>>>>>>>>>>>>> |- 2:0:1:0 sdc 8:32 active ready running >>>>>>>>>>>>>>>>>>>> `- 3:0:1:0 sdg 8:96 active ready running >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> My steps on first server (initial deploy of >>>>>>>>>>>>>>>>>>>> ovirt 4.0 hosted engine): >>>>>>>>>>>>>>>>>>>> ============================ >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> # systemctl stop NetworkManager >>>>>>>>>>>>>>>>>>>> # systemctl disable NetworkManager >>>>>>>>>>>>>>>>>>>> # yum -y install >>>>>>>>>>>>>>>>>>>> http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm >>>>>>>>>>>>>>>>>>>> # yum -y install epel-release >>>>>>>>>>>>>>>>>>>> # wget >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1511.iso >>>>>>>>>>>>>>>>>>>> -P /tmp/ >>>>>>>>>>>>>>>>>>>> # yum install ovirt-hosted-engine-setup >>>>>>>>>>>>>>>>>>>> # yum install screen >>>>>>>>>>>>>>>>>>>> # screen -RD >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> ...in screen session : >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> # hosted-engine --deploy >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>>>>>>> in configuration process I chose "fc" as >>>>>>>>>>>>>>>>>>>> storage type for oVirt hosted >>>>>>>>>>>>>>>>>>>> engine vm and select 60GB LUN... >>>>>>>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> --== CONFIGURATION PREVIEW ==-- >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>>>>>>> Firewall manager : iptables >>>>>>>>>>>>>>>>>>>> Gateway address : 10.1.0.1 >>>>>>>>>>>>>>>>>>>> Host name for web application : >>>>>>>>>>>>>>>>>>>> KOM-AD01-OVIRT1 >>>>>>>>>>>>>>>>>>>> Storage Domain type : fc >>>>>>>>>>>>>>>>>>>> Host ID : 1 >>>>>>>>>>>>>>>>>>>> LUN ID : >>>>>>>>>>>>>>>>>>>> 360002ac0000000000000001b0000cec9 >>>>>>>>>>>>>>>>>>>> Image size GB : 40 >>>>>>>>>>>>>>>>>>>> Console type : vnc >>>>>>>>>>>>>>>>>>>> Memory size MB : 4096 >>>>>>>>>>>>>>>>>>>> MAC address : 00:16:3e:77:1d:07 >>>>>>>>>>>>>>>>>>>> Boot type : cdrom >>>>>>>>>>>>>>>>>>>> Number of CPUs : 2 >>>>>>>>>>>>>>>>>>>> ISO image (cdrom boot/cloud-init) : >>>>>>>>>>>>>>>>>>>> /tmp/CentOS-7-x86_64-NetInstall-1511.iso >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Can I ask why you prefer/need to manually create >>>>>>>>>>>>>>>>>>> a VM installing from >>>>>>>>>>>>>>>>>>> a CD instead of using the ready-to-use >>>>>>>>>>>>>>>>>>> ovirt-engine-appliance? >>>>>>>>>>>>>>>>>>> Using the appliance makes the setup process a lot >>>>>>>>>>>>>>>>>>> shorted and more comfortable. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> CPU Type : model_Penryn >>>>>>>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>>>>>>> and get error after step "Verifying sanlock >>>>>>>>>>>>>>>>>>>> lockspace initialization" >>>>>>>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> [ INFO ] Verifying sanlock lockspace >>>>>>>>>>>>>>>>>>>> initialization >>>>>>>>>>>>>>>>>>>> [ ERROR ] Failed to execute stage 'Misc >>>>>>>>>>>>>>>>>>>> configuration': [Errno 101] Network >>>>>>>>>>>>>>>>>>>> is unreachable >>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Clean up >>>>>>>>>>>>>>>>>>>> [ INFO ] Generating answer file >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf' >>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Pre-termination >>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Termination >>>>>>>>>>>>>>>>>>>> [ ERROR ] Hosted Engine deployment failed: this >>>>>>>>>>>>>>>>>>>> system is not reliable, >>>>>>>>>>>>>>>>>>>> please check the issue, fix and redeploy >>>>>>>>>>>>>>>>>>>> Log file is located at >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160722123404-t26vw0.log >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Interestingly >>>>>>>>>>>>>>>>>>>> ============================ >>>>>>>>>>>>>>>>>>>> If I try to deploy hosted-engine v3.6, >>>>>>>>>>>>>>>>>>>> everything goes well in the same >>>>>>>>>>>>>>>>>>>> configuration !! : >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> .... >>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Transaction setup >>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Misc configuration >>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Package installation >>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Misc configuration >>>>>>>>>>>>>>>>>>>> [ INFO ] Configuring libvirt >>>>>>>>>>>>>>>>>>>> [ INFO ] Configuring VDSM >>>>>>>>>>>>>>>>>>>> [ INFO ] Starting vdsmd >>>>>>>>>>>>>>>>>>>> [ INFO ] Waiting for VDSM hardware info >>>>>>>>>>>>>>>>>>>> [ INFO ] Configuring the management bridge >>>>>>>>>>>>>>>>>>>> [ INFO ] Creating Volume Group >>>>>>>>>>>>>>>>>>>> [ INFO ] Creating Storage Domain >>>>>>>>>>>>>>>>>>>> [ INFO ] Creating Storage Pool >>>>>>>>>>>>>>>>>>>> [ INFO ] Connecting Storage Pool >>>>>>>>>>>>>>>>>>>> [ INFO ] Verifying sanlock lockspace >>>>>>>>>>>>>>>>>>>> initialization >>>>>>>>>>>>>>>>>>>> [ INFO ] Creating Image for >>>>>>>>>>>>>>>>>>>> 'hosted-engine.lockspace' ... >>>>>>>>>>>>>>>>>>>> [ INFO ] Image for 'hosted-engine.lockspace' >>>>>>>>>>>>>>>>>>>> created successfully >>>>>>>>>>>>>>>>>>>> [ INFO ] Creating Image for >>>>>>>>>>>>>>>>>>>> 'hosted-engine.metadata' ... >>>>>>>>>>>>>>>>>>>> [ INFO ] Image for 'hosted-engine.metadata' >>>>>>>>>>>>>>>>>>>> created successfully >>>>>>>>>>>>>>>>>>>> [ INFO ] Creating VM Image >>>>>>>>>>>>>>>>>>>> [ INFO ] Destroying Storage Pool >>>>>>>>>>>>>>>>>>>> [ INFO ] Start monitoring domain >>>>>>>>>>>>>>>>>>>> [ INFO ] Configuring VM >>>>>>>>>>>>>>>>>>>> [ INFO ] Updating hosted-engine configuration >>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Transaction commit >>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Closing up >>>>>>>>>>>>>>>>>>>> [ INFO ] Creating VM >>>>>>>>>>>>>>>>>>>> You can now connect to the VM with >>>>>>>>>>>>>>>>>>>> the following command: >>>>>>>>>>>>>>>>>>>> /bin/remote-viewer >>>>>>>>>>>>>>>>>>>> vnc://localhost:5900 >>>>>>>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> What could be the problem? >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>>>> Users mailing list >>>>>>>>>>>>>>>>>>>> Users@ovirt.org >>>>>>>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Users mailing list >>>>>>>>>>> Users@ovirt.org >>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> Users mailing list >>>>>>>> Users@ovirt.org >>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>> >>>>> _______________________________________________ >>>>> Users mailing list >>>>> Users@ovirt.org >>>>> http://lists.ovirt.org/mailman/listinfo/users _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users