# ss -plutn

Netid State      Recv-Q Send-Q                                                  
                     Local Address:Port     Peer Address:Port

udp   UNCONN     0      0                                                       
                                 *:111                                          
                                                        *:*                   
users:(("rpcbind",pid=827,fd=6))
udp   UNCONN     0      0                                                       
                                 *:161                                          
                                                        *:*                   
users:(("snmpd",pid=1609,fd=6))
udp   UNCONN     0      0                                                       
                         127.0.0.1:323                                          
                                                        *:*                   
users:(("chronyd",pid=795,fd=1))
udp   UNCONN     0      0                                                       
                                 *:959                                          
                                                        *:*                   
users:(("rpcbind",pid=827,fd=7))
udp   UNCONN     0      0                                                       
                         127.0.0.1:25375                                        
                                                        *:*                   
users:(("snmpd",pid=1609,fd=8))
udp   UNCONN     0      0                                                       
                         127.0.0.1:25376                                        
                                                        *:*                   
users:(("cmapeerd",pid=2056,fd=5))
udp   UNCONN     0      0                                                       
                         127.0.0.1:25393                                        
                                                        *:*                   
users:(("cmanicd",pid=2278,fd=3))
udp   UNCONN     0      0                                                       
                                :::111                                          
                                                       :::*                   
users:(("rpcbind",pid=827,fd=9))
udp   UNCONN     0      0                                                       
                                :::959                                          
                                                       :::*                   
users:(("rpcbind",pid=827,fd=10))
tcp   LISTEN     0      128                                                     
                                 *:2381                                         
                                                        *:*                   
users:(("hpsmhd",pid=3903,fd=4),("hpsmhd",pid=3901,fd=4),("hpsmhd",pid=3900,fd=4),("hpsmhd",pid=3899,fd=4),("hpsmhd",pid=3898,fd=4),("hpsmhd",pid=3893,fd=4))
tcp   LISTEN     0      128                                                     
                                 *:111                                          
                                                        *:*                   
users:(("rpcbind",pid=827,fd=8))
tcp   LISTEN     0      5                                                       
                                 *:54322                                        
                                                        *:*                   
users:(("ovirt-imageio-d",pid=753,fd=3))
tcp   LISTEN     0      128                                                     
                                 *:22                                           
                                                        *:*                   
users:(("sshd",pid=1606,fd=3))
tcp   LISTEN     0      100                                                     
                         127.0.0.1:25                                           
                                                        *:*                   
users:(("master",pid=1948,fd=13))
tcp   LISTEN     0      128                                                     
                                 *:2301                                         
                                                        *:*                   
users:(("hpsmhd",pid=3903,fd=3),("hpsmhd",pid=3901,fd=3),("hpsmhd",pid=3900,fd=3),("hpsmhd",pid=3899,fd=3),("hpsmhd",pid=3898,fd=3),("hpsmhd",pid=3893,fd=3))
tcp   LISTEN     0      30                                                      
                                 *:16514                                        
                                                        *:*                   
users:(("libvirtd",pid=10688,fd=13))
tcp   LISTEN     0      128                                                     
                         127.0.0.1:199                                          
                                                        *:*                   
users:(("snmpd",pid=1609,fd=9))
tcp   LISTEN     0      128                                                     
                                :::111                                          
                                                       :::*                   
users:(("rpcbind",pid=827,fd=11))
tcp   LISTEN     0      5                                                       
                                :::54321                                        
                                                       :::*                   
users:(("vdsm",pid=11077,fd=23))
tcp   LISTEN     0      30                                                      
                                :::16514                                        
                                                       :::*                   
users:(("libvirtd",pid=10688,fd=14))

25.07.2016, 15:11, "Simone Tiraboschi" <stira...@redhat.com>:
> On Mon, Jul 25, 2016 at 2:03 PM, <aleksey.maksi...@it-kb.ru> wrote:
>>  Yes.
>>
>>  # ping $(python -c 'import socket; print(socket.gethostname())')
>>
>>  PING KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data.
>>  64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1 ttl=64 
>> time=0.030 ms
>>  64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2 ttl=64 
>> time=0.015 ms
>>  64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3 ttl=64 
>> time=0.011 ms
>>  ^C
>>  --- KOM-AD01-VM31.holding.com ping statistics ---
>>  3 packets transmitted, 3 received, 0% packet loss, time 2001ms
>>  rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms
>>
>>  but...
>
> and the output of
>  ss -plutn
>
>>  # vdsClient -s 0 getVdsCaps
>>
>>  Traceback (most recent call last):
>>    File "/usr/share/vdsm/vdsClient.py", line 2980, in <module>
>>      code, message = commands[command][0](commandArgs)
>>    File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
>>      return self.ExecAndExit(self.s.getVdsCapabilities())
>>    File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
>>      return self.__send(self.__name, args)
>>    File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
>>      verbose=self.__verbose
>>    File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
>>      return self.single_request(host, handler, request_body, verbose)
>>    File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request
>>      self.send_content(h, request_body)
>>    File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content
>>      connection.endheaders(request_body)
>>    File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders
>>      self._send_output(message_body)
>>    File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output
>>      self.send(msg)
>>    File "/usr/lib64/python2.7/httplib.py", line 797, in send
>>      self.connect()
>>    File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in 
>> connect
>>      sock = socket.create_connection((self.host, self.port), self.timeout)
>>    File "/usr/lib64/python2.7/socket.py", line 571, in create_connection
>>      raise err
>>  error: [Errno 101] Network is unreachable
>>
>>  25.07.2016, 14:58, "aleksey.maksi...@it-kb.ru" <aleksey.maksi...@it-kb.ru>:
>>>  Ok.
>>>
>>>  1) I stopped and disabled the service NetworkManager
>>>  # systemctl stop NetworkManager
>>>  # systemctl disable NetworkManager
>>>
>>>  2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add PEERDNS=no in 
>>> ifcfg-* file.
>>>
>>>  3) Reboot server
>>>
>>>  4) Try deploy oVirt HE 4 and I get the same error
>>>
>>>  [ INFO ] Creating Volume Group
>>>  [ INFO ] Creating Storage Domain
>>>  [ INFO ] Creating Storage Pool
>>>  [ INFO ] Connecting Storage Pool
>>>  [ INFO ] Verifying sanlock lockspace initialization
>>>  [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] 
>>> Network is unreachable
>>>  [ INFO ] Stage: Clean up
>>>  [ INFO ] Generating answer file 
>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf'
>>>  [ INFO ] Stage: Pre-termination
>>>  [ INFO ] Stage: Termination
>>>  [ ERROR ] Hosted Engine deployment failed: this system is not reliable, 
>>> please check the issue, fix and redeploy
>>>            Log file is located at 
>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725142534-t81kwf.log
>>>
>>>  What ideas further?
>>>
>>>  25.07.2016, 13:06, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>   On Mon, Jul 25, 2016 at 11:54 AM, <aleksey.maksi...@it-kb.ru> wrote:
>>>>>    What am I supposed to do for successfully deploy ovirt 4 ?
>>>>>    Any ideas ?
>>>>
>>>>   Can you please try to explicitly configure your DNS with nameserver
>>>>   under /etc/resolv.conf and remove DNS1 and DNS2 and set PEERDNS=no for
>>>>   the interface you are going to use?
>>>>
>>>>>    25.07.2016, 12:47, "aleksey.maksi...@it-kb.ru" 
>>>>> <aleksey.maksi...@it-kb.ru>:
>>>>>>    "Aleksey, was your DNS configured with DNS1 and DNS2 just on the 
>>>>>> interface you used to create the management bridge on?"
>>>>>>
>>>>>>    Yes. Of course
>>>>>>
>>>>>>    25.07.2016, 12:27, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>>>     On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski
>>>>>>>     <piotr.kliczew...@gmail.com> wrote:
>>>>>>>>      This could be the issue here as well as for BZ #1358530
>>>>>>>>
>>>>>>>>      On Mon, Jul 25, 2016 at 10:53 AM, <aleksey.maksi...@it-kb.ru> 
>>>>>>>> wrote:
>>>>>>>>>      Could this be due to the fact that the ovirt installer has 
>>>>>>>>> changed network configuration files (ifcfg-*, resolv.conf) ?
>>>>>>>>>      After the error in ovirt installation process I see from 
>>>>>>>>> resolv.conf disappeared on my DNS servers entry and now the server is 
>>>>>>>>> unable to resolve names.
>>>>>>>
>>>>>>>     So it could be related to 
>>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1160423
>>>>>>>
>>>>>>>     Aleksey, was your DNS configured with DNS1 and DNS2 just on the
>>>>>>>     interface you used to create the management bridge on?
>>>>>>>     Can you please try the workaround described here
>>>>>>>     https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ?
>>>>>>>
>>>>>>>>>      25.07.2016, 11:26, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>>>>>>      On Mon, Jul 25, 2016 at 10:22 AM, <aleksey.maksi...@it-kb.ru> 
>>>>>>>>>> wrote:
>>>>>>>>>>>       # vdsClient -s 0 getVdsCaps
>>>>>>>>>>>
>>>>>>>>>>>       Traceback (most recent call last):
>>>>>>>>>>>         File "/usr/share/vdsm/vdsClient.py", line 2980, in <module>
>>>>>>>>>>>           code, message = commands[command][0](commandArgs)
>>>>>>>>>>>         File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
>>>>>>>>>>>           return self.ExecAndExit(self.s.getVdsCapabilities())
>>>>>>>>>>>         File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in 
>>>>>>>>>>> __call__
>>>>>>>>>>>           return self.__send(self.__name, args)
>>>>>>>>>>>         File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in 
>>>>>>>>>>> __request
>>>>>>>>>>>           verbose=self.__verbose
>>>>>>>>>>>         File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in 
>>>>>>>>>>> request
>>>>>>>>>>>           return self.single_request(host, handler, request_body, 
>>>>>>>>>>> verbose)
>>>>>>>>>>>         File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in 
>>>>>>>>>>> single_request
>>>>>>>>>>>           self.send_content(h, request_body)
>>>>>>>>>>>         File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in 
>>>>>>>>>>> send_content
>>>>>>>>>>>           connection.endheaders(request_body)
>>>>>>>>>>>         File "/usr/lib64/python2.7/httplib.py", line 975, in 
>>>>>>>>>>> endheaders
>>>>>>>>>>>           self._send_output(message_body)
>>>>>>>>>>>         File "/usr/lib64/python2.7/httplib.py", line 835, in 
>>>>>>>>>>> _send_output
>>>>>>>>>>>           self.send(msg)
>>>>>>>>>>>         File "/usr/lib64/python2.7/httplib.py", line 797, in send
>>>>>>>>>>>           self.connect()
>>>>>>>>>>>         File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", 
>>>>>>>>>>> line 203, in connect
>>>>>>>>>>>           sock = socket.create_connection((self.host, self.port), 
>>>>>>>>>>> self.timeout)
>>>>>>>>>>>         File "/usr/lib64/python2.7/socket.py", line 571, in 
>>>>>>>>>>> create_connection
>>>>>>>>>>>           raise err
>>>>>>>>>>>       error: [Errno 101] Network is unreachable
>>>>>>>>>>
>>>>>>>>>>      Yaniv, can you please take also a look to this one?
>>>>>>>>>>      it's exactly the opposite of https://bugzilla.redhat.com/1358530
>>>>>>>>>>      Here the jsonrpcclient works but not the xmlrpc one.
>>>>>>>>>>
>>>>>>>>>>>       25.07.2016, 11:17, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>>>>>>>>       On Mon, Jul 25, 2016 at 7:51 AM, <aleksey.maksi...@it-kb.ru> 
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>        Simone, there is something interesting in the vdsm.log?
>>>>>>>>>>>>
>>>>>>>>>>>>       For what I saw the issue is not related to the storage but 
>>>>>>>>>>>> to the network.
>>>>>>>>>>>>       ovirt-hosted-engine-setup uses the jsonrpc client, instead 
>>>>>>>>>>>> the code
>>>>>>>>>>>>       from ovirt-hosted-engine-ha still uses the xmlrpc client 
>>>>>>>>>>>> somewhere and
>>>>>>>>>>>>       this happens also when the setup asks to create the 
>>>>>>>>>>>> lockspace volume.
>>>>>>>>>>>>       It seams that in your case the xmlrpc client could not 
>>>>>>>>>>>> connect vdsm on
>>>>>>>>>>>>       the localhost.
>>>>>>>>>>>>       It could be somehow related to:
>>>>>>>>>>>>       https://bugzilla.redhat.com/1358530
>>>>>>>>>>>>
>>>>>>>>>>>>       Can you please try executing
>>>>>>>>>>>>        sudo vdsClient -s 0 getVdsCaps
>>>>>>>>>>>>       on that host?
>>>>>>>>>>>>
>>>>>>>>>>>>>        22.07.2016, 19:36, "aleksey.maksi...@it-kb.ru" 
>>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru>:
>>>>>>>>>>>>>>        Simone, thanks for link.
>>>>>>>>>>>>>>        vdsm.log attached
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        22.07.2016, 19:28, "Simone Tiraboschi" 
>>>>>>>>>>>>>> <stira...@redhat.com>:
>>>>>>>>>>>>>>>         On Fri, Jul 22, 2016 at 5:59 PM, 
>>>>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>>>>>>>>>>>          Thank you for your response, Simone.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>          Log attached.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>         It seams it comes from VDSM, can you please attach also 
>>>>>>>>>>>>>>> vdsm.log?
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>          I don't use ovirt-engine-appliance because I have not 
>>>>>>>>>>>>>>>> found "how-to" for ovirt-engine-appliance deployment in hosted 
>>>>>>>>>>>>>>>> engine configuration.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>         yum install ovirt-engine-appliance
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>         Then follow the instruction here:
>>>>>>>>>>>>>>>         
>>>>>>>>>>>>>>> http://www.ovirt.org/develop/release-management/features/heapplianceflow/
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>          22.07.2016, 17:09, "Simone Tiraboschi" 
>>>>>>>>>>>>>>>> <stira...@redhat.com>:
>>>>>>>>>>>>>>>>>          Hi Aleksey,
>>>>>>>>>>>>>>>>>          Can you please attach hosted-engine-setup logs?
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          On Fri, Jul 22, 2016 at 3:46 PM, 
>>>>>>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           Hello oVirt guru`s !
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           I have problem with initial deploy of ovirt 4.0 
>>>>>>>>>>>>>>>>>> hosted engine.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           My environment :
>>>>>>>>>>>>>>>>>>           ============================
>>>>>>>>>>>>>>>>>>           * Two servers HP ProLiant DL 360 G5 with Qlogic FC 
>>>>>>>>>>>>>>>>>> HBA connected (with
>>>>>>>>>>>>>>>>>>           multipathd) to storage HP 3PAR 7200
>>>>>>>>>>>>>>>>>>           * On each server installed CentOS 7.2 Linux 
>>>>>>>>>>>>>>>>>> (3.10.0-327.22.2.el7.x86_64)
>>>>>>>>>>>>>>>>>>           * On 3PAR storage I created 2 LUNs for oVirt.
>>>>>>>>>>>>>>>>>>           - First LUN for oVirt Hosted Engine VM (60GB)
>>>>>>>>>>>>>>>>>>           - Second LUN for all other VMs (2TB)
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           # multipath -ll
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           3par-vv1 (360002ac0000000000000001b0000cec9) dm-0 
>>>>>>>>>>>>>>>>>> 3PARdata,VV
>>>>>>>>>>>>>>>>>>           size=60G features='1 queue_if_no_path' 
>>>>>>>>>>>>>>>>>> hwhandler='1 alua' wp=rw
>>>>>>>>>>>>>>>>>>           `-+- policy='round-robin 0' prio=50 status=active
>>>>>>>>>>>>>>>>>>             |- 2:0:1:1 sdd 8:48 active ready running
>>>>>>>>>>>>>>>>>>             |- 3:0:0:1 sdf 8:80 active ready running
>>>>>>>>>>>>>>>>>>             |- 2:0:0:1 sdb 8:16 active ready running
>>>>>>>>>>>>>>>>>>             `- 3:0:1:1 sdh 8:112 active ready running
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           3par-vv2 (360002ac000000000000000160000cec9) dm-1 
>>>>>>>>>>>>>>>>>> 3PARdata,VV
>>>>>>>>>>>>>>>>>>           size=2.0T features='1 queue_if_no_path' 
>>>>>>>>>>>>>>>>>> hwhandler='1 alua' wp=rw
>>>>>>>>>>>>>>>>>>           `-+- policy='round-robin 0' prio=50 status=active
>>>>>>>>>>>>>>>>>>             |- 2:0:0:0 sda 8:0 active ready running
>>>>>>>>>>>>>>>>>>             |- 3:0:0:0 sde 8:64 active ready running
>>>>>>>>>>>>>>>>>>             |- 2:0:1:0 sdc 8:32 active ready running
>>>>>>>>>>>>>>>>>>             `- 3:0:1:0 sdg 8:96 active ready running
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           My steps on first server (initial deploy of ovirt 
>>>>>>>>>>>>>>>>>> 4.0 hosted engine):
>>>>>>>>>>>>>>>>>>           ============================
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           # systemctl stop NetworkManager
>>>>>>>>>>>>>>>>>>           # systemctl disable NetworkManager
>>>>>>>>>>>>>>>>>>           # yum -y install 
>>>>>>>>>>>>>>>>>> http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm
>>>>>>>>>>>>>>>>>>           # yum -y install epel-release
>>>>>>>>>>>>>>>>>>           # wget
>>>>>>>>>>>>>>>>>>           
>>>>>>>>>>>>>>>>>> http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1511.iso
>>>>>>>>>>>>>>>>>>           -P /tmp/
>>>>>>>>>>>>>>>>>>           # yum install ovirt-hosted-engine-setup
>>>>>>>>>>>>>>>>>>           # yum install screen
>>>>>>>>>>>>>>>>>>           # screen -RD
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           ...in screen session :
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           # hosted-engine --deploy
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           ...
>>>>>>>>>>>>>>>>>>           in configuration process I chose "fc" as storage 
>>>>>>>>>>>>>>>>>> type for oVirt hosted
>>>>>>>>>>>>>>>>>>           engine vm and select 60GB LUN...
>>>>>>>>>>>>>>>>>>           ...
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           --== CONFIGURATION PREVIEW ==--
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           ...
>>>>>>>>>>>>>>>>>>                     Firewall manager : iptables
>>>>>>>>>>>>>>>>>>                     Gateway address : 10.1.0.1
>>>>>>>>>>>>>>>>>>                     Host name for web application : 
>>>>>>>>>>>>>>>>>> KOM-AD01-OVIRT1
>>>>>>>>>>>>>>>>>>                     Storage Domain type : fc
>>>>>>>>>>>>>>>>>>                     Host ID : 1
>>>>>>>>>>>>>>>>>>                     LUN ID :
>>>>>>>>>>>>>>>>>>           360002ac0000000000000001b0000cec9
>>>>>>>>>>>>>>>>>>                     Image size GB : 40
>>>>>>>>>>>>>>>>>>                     Console type : vnc
>>>>>>>>>>>>>>>>>>                     Memory size MB : 4096
>>>>>>>>>>>>>>>>>>                     MAC address : 00:16:3e:77:1d:07
>>>>>>>>>>>>>>>>>>                     Boot type : cdrom
>>>>>>>>>>>>>>>>>>                     Number of CPUs : 2
>>>>>>>>>>>>>>>>>>                     ISO image (cdrom boot/cloud-init) :
>>>>>>>>>>>>>>>>>>           /tmp/CentOS-7-x86_64-NetInstall-1511.iso
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          Can I ask why you prefer/need to manually create a 
>>>>>>>>>>>>>>>>> VM installing from
>>>>>>>>>>>>>>>>>          a CD instead of using the ready-to-use 
>>>>>>>>>>>>>>>>> ovirt-engine-appliance?
>>>>>>>>>>>>>>>>>          Using the appliance makes the setup process a lot 
>>>>>>>>>>>>>>>>> shorted and more comfortable.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>                     CPU Type : model_Penryn
>>>>>>>>>>>>>>>>>>           ...
>>>>>>>>>>>>>>>>>>           and get error after step "Verifying sanlock 
>>>>>>>>>>>>>>>>>> lockspace initialization"
>>>>>>>>>>>>>>>>>>           ...
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           [ INFO ] Verifying sanlock lockspace initialization
>>>>>>>>>>>>>>>>>>           [ ERROR ] Failed to execute stage 'Misc 
>>>>>>>>>>>>>>>>>> configuration': [Errno 101] Network
>>>>>>>>>>>>>>>>>>           is unreachable
>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Clean up
>>>>>>>>>>>>>>>>>>           [ INFO ] Generating answer file
>>>>>>>>>>>>>>>>>>           
>>>>>>>>>>>>>>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf'
>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Pre-termination
>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Termination
>>>>>>>>>>>>>>>>>>           [ ERROR ] Hosted Engine deployment failed: this 
>>>>>>>>>>>>>>>>>> system is not reliable,
>>>>>>>>>>>>>>>>>>           please check the issue, fix and redeploy
>>>>>>>>>>>>>>>>>>                     Log file is located at
>>>>>>>>>>>>>>>>>>           
>>>>>>>>>>>>>>>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160722123404-t26vw0.log
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           Interestingly
>>>>>>>>>>>>>>>>>>           ============================
>>>>>>>>>>>>>>>>>>           If I try to deploy hosted-engine v3.6, everything 
>>>>>>>>>>>>>>>>>> goes well in the same
>>>>>>>>>>>>>>>>>>           configuration !! :
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           ....
>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Transaction setup
>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Misc configuration
>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Package installation
>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Misc configuration
>>>>>>>>>>>>>>>>>>           [ INFO ] Configuring libvirt
>>>>>>>>>>>>>>>>>>           [ INFO ] Configuring VDSM
>>>>>>>>>>>>>>>>>>           [ INFO ] Starting vdsmd
>>>>>>>>>>>>>>>>>>           [ INFO ] Waiting for VDSM hardware info
>>>>>>>>>>>>>>>>>>           [ INFO ] Configuring the management bridge
>>>>>>>>>>>>>>>>>>           [ INFO ] Creating Volume Group
>>>>>>>>>>>>>>>>>>           [ INFO ] Creating Storage Domain
>>>>>>>>>>>>>>>>>>           [ INFO ] Creating Storage Pool
>>>>>>>>>>>>>>>>>>           [ INFO ] Connecting Storage Pool
>>>>>>>>>>>>>>>>>>           [ INFO ] Verifying sanlock lockspace initialization
>>>>>>>>>>>>>>>>>>           [ INFO ] Creating Image for 
>>>>>>>>>>>>>>>>>> 'hosted-engine.lockspace' ...
>>>>>>>>>>>>>>>>>>           [ INFO ] Image for 'hosted-engine.lockspace' 
>>>>>>>>>>>>>>>>>> created successfully
>>>>>>>>>>>>>>>>>>           [ INFO ] Creating Image for 
>>>>>>>>>>>>>>>>>> 'hosted-engine.metadata' ...
>>>>>>>>>>>>>>>>>>           [ INFO ] Image for 'hosted-engine.metadata' 
>>>>>>>>>>>>>>>>>> created successfully
>>>>>>>>>>>>>>>>>>           [ INFO ] Creating VM Image
>>>>>>>>>>>>>>>>>>           [ INFO ] Destroying Storage Pool
>>>>>>>>>>>>>>>>>>           [ INFO ] Start monitoring domain
>>>>>>>>>>>>>>>>>>           [ INFO ] Configuring VM
>>>>>>>>>>>>>>>>>>           [ INFO ] Updating hosted-engine configuration
>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Transaction commit
>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Closing up
>>>>>>>>>>>>>>>>>>           [ INFO ] Creating VM
>>>>>>>>>>>>>>>>>>                     You can now connect to the VM with the 
>>>>>>>>>>>>>>>>>> following command:
>>>>>>>>>>>>>>>>>>                           /bin/remote-viewer 
>>>>>>>>>>>>>>>>>> vnc://localhost:5900
>>>>>>>>>>>>>>>>>>           ...
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           What could be the problem?
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>           _______________________________________________
>>>>>>>>>>>>>>>>>>           Users mailing list
>>>>>>>>>>>>>>>>>>           Users@ovirt.org
>>>>>>>>>>>>>>>>>>           http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>      _______________________________________________
>>>>>>>>>      Users mailing list
>>>>>>>>>      Users@ovirt.org
>>>>>>>>>      http://lists.ovirt.org/mailman/listinfo/users
>>>>>>
>>>>>>    _______________________________________________
>>>>>>    Users mailing list
>>>>>>    Users@ovirt.org
>>>>>>    http://lists.ovirt.org/mailman/listinfo/users
>>>
>>>  _______________________________________________
>>>  Users mailing list
>>>  Users@ovirt.org
>>>  http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to