On Mon, Jul 25, 2016 at 2:15 PM,  <aleksey.maksi...@it-kb.ru> wrote:
> # ss -plutn
>
> Netid State      Recv-Q Send-Q                                                
>                        Local Address:Port     Peer Address:Port
>
> udp   UNCONN     0      0                                                     
>                                    *:111                                      
>                                                             *:*               
>     users:(("rpcbind",pid=827,fd=6))
> udp   UNCONN     0      0                                                     
>                                    *:161                                      
>                                                             *:*               
>     users:(("snmpd",pid=1609,fd=6))
> udp   UNCONN     0      0                                                     
>                            127.0.0.1:323                                      
>                                                             *:*               
>     users:(("chronyd",pid=795,fd=1))
> udp   UNCONN     0      0                                                     
>                                    *:959                                      
>                                                             *:*               
>     users:(("rpcbind",pid=827,fd=7))
> udp   UNCONN     0      0                                                     
>                            127.0.0.1:25375                                    
>                                                             *:*               
>     users:(("snmpd",pid=1609,fd=8))
> udp   UNCONN     0      0                                                     
>                            127.0.0.1:25376                                    
>                                                             *:*               
>     users:(("cmapeerd",pid=2056,fd=5))
> udp   UNCONN     0      0                                                     
>                            127.0.0.1:25393                                    
>                                                             *:*               
>     users:(("cmanicd",pid=2278,fd=3))
> udp   UNCONN     0      0                                                     
>                                   :::111                                      
>                                                            :::*               
>     users:(("rpcbind",pid=827,fd=9))
> udp   UNCONN     0      0                                                     
>                                   :::959                                      
>                                                            :::*               
>     users:(("rpcbind",pid=827,fd=10))
> tcp   LISTEN     0      128                                                   
>                                    *:2381                                     
>                                                             *:*               
>     
> users:(("hpsmhd",pid=3903,fd=4),("hpsmhd",pid=3901,fd=4),("hpsmhd",pid=3900,fd=4),("hpsmhd",pid=3899,fd=4),("hpsmhd",pid=3898,fd=4),("hpsmhd",pid=3893,fd=4))
> tcp   LISTEN     0      128                                                   
>                                    *:111                                      
>                                                             *:*               
>     users:(("rpcbind",pid=827,fd=8))
> tcp   LISTEN     0      5                                                     
>                                    *:54322                                    
>                                                             *:*               
>     users:(("ovirt-imageio-d",pid=753,fd=3))
> tcp   LISTEN     0      128                                                   
>                                    *:22                                       
>                                                             *:*               
>     users:(("sshd",pid=1606,fd=3))
> tcp   LISTEN     0      100                                                   
>                            127.0.0.1:25                                       
>                                                             *:*               
>     users:(("master",pid=1948,fd=13))
> tcp   LISTEN     0      128                                                   
>                                    *:2301                                     
>                                                             *:*               
>     
> users:(("hpsmhd",pid=3903,fd=3),("hpsmhd",pid=3901,fd=3),("hpsmhd",pid=3900,fd=3),("hpsmhd",pid=3899,fd=3),("hpsmhd",pid=3898,fd=3),("hpsmhd",pid=3893,fd=3))
> tcp   LISTEN     0      30                                                    
>                                    *:16514                                    
>                                                             *:*               
>     users:(("libvirtd",pid=10688,fd=13))
> tcp   LISTEN     0      128                                                   
>                            127.0.0.1:199                                      
>                                                             *:*               
>     users:(("snmpd",pid=1609,fd=9))
> tcp   LISTEN     0      128                                                   
>                                   :::111                                      
>                                                            :::*               
>     users:(("rpcbind",pid=827,fd=11))
> tcp   LISTEN     0      5                                                     
>                                   :::54321                                    
>                                                            :::*               
>     users:(("vdsm",pid=11077,fd=23))

vdsm is properly bind over ipv6.

Can you please check if you can connect to vdsm with:
 telnet  kom-ad01-vm31.holding.com 54321
and with
 telnet ::1 54321
?

> tcp   LISTEN     0      30                                                    
>                                   :::16514                                    
>                                                            :::*               
>     users:(("libvirtd",pid=10688,fd=14))
>
> 25.07.2016, 15:11, "Simone Tiraboschi" <stira...@redhat.com>:
>> On Mon, Jul 25, 2016 at 2:03 PM, <aleksey.maksi...@it-kb.ru> wrote:
>>>  Yes.
>>>
>>>  # ping $(python -c 'import socket; print(socket.gethostname())')
>>>
>>>  PING KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data.
>>>  64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1 ttl=64 
>>> time=0.030 ms
>>>  64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2 ttl=64 
>>> time=0.015 ms
>>>  64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3 ttl=64 
>>> time=0.011 ms
>>>  ^C
>>>  --- KOM-AD01-VM31.holding.com ping statistics ---
>>>  3 packets transmitted, 3 received, 0% packet loss, time 2001ms
>>>  rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms
>>>
>>>  but...
>>
>> and the output of
>>  ss -plutn
>>
>>>  # vdsClient -s 0 getVdsCaps
>>>
>>>  Traceback (most recent call last):
>>>    File "/usr/share/vdsm/vdsClient.py", line 2980, in <module>
>>>      code, message = commands[command][0](commandArgs)
>>>    File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
>>>      return self.ExecAndExit(self.s.getVdsCapabilities())
>>>    File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
>>>      return self.__send(self.__name, args)
>>>    File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
>>>      verbose=self.__verbose
>>>    File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
>>>      return self.single_request(host, handler, request_body, verbose)
>>>    File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request
>>>      self.send_content(h, request_body)
>>>    File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content
>>>      connection.endheaders(request_body)
>>>    File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders
>>>      self._send_output(message_body)
>>>    File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output
>>>      self.send(msg)
>>>    File "/usr/lib64/python2.7/httplib.py", line 797, in send
>>>      self.connect()
>>>    File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in 
>>> connect
>>>      sock = socket.create_connection((self.host, self.port), self.timeout)
>>>    File "/usr/lib64/python2.7/socket.py", line 571, in create_connection
>>>      raise err
>>>  error: [Errno 101] Network is unreachable
>>>
>>>  25.07.2016, 14:58, "aleksey.maksi...@it-kb.ru" <aleksey.maksi...@it-kb.ru>:
>>>>  Ok.
>>>>
>>>>  1) I stopped and disabled the service NetworkManager
>>>>  # systemctl stop NetworkManager
>>>>  # systemctl disable NetworkManager
>>>>
>>>>  2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add PEERDNS=no in 
>>>> ifcfg-* file.
>>>>
>>>>  3) Reboot server
>>>>
>>>>  4) Try deploy oVirt HE 4 and I get the same error
>>>>
>>>>  [ INFO ] Creating Volume Group
>>>>  [ INFO ] Creating Storage Domain
>>>>  [ INFO ] Creating Storage Pool
>>>>  [ INFO ] Connecting Storage Pool
>>>>  [ INFO ] Verifying sanlock lockspace initialization
>>>>  [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] 
>>>> Network is unreachable
>>>>  [ INFO ] Stage: Clean up
>>>>  [ INFO ] Generating answer file 
>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf'
>>>>  [ INFO ] Stage: Pre-termination
>>>>  [ INFO ] Stage: Termination
>>>>  [ ERROR ] Hosted Engine deployment failed: this system is not reliable, 
>>>> please check the issue, fix and redeploy
>>>>            Log file is located at 
>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725142534-t81kwf.log
>>>>
>>>>  What ideas further?
>>>>
>>>>  25.07.2016, 13:06, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>   On Mon, Jul 25, 2016 at 11:54 AM, <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>    What am I supposed to do for successfully deploy ovirt 4 ?
>>>>>>    Any ideas ?
>>>>>
>>>>>   Can you please try to explicitly configure your DNS with nameserver
>>>>>   under /etc/resolv.conf and remove DNS1 and DNS2 and set PEERDNS=no for
>>>>>   the interface you are going to use?
>>>>>
>>>>>>    25.07.2016, 12:47, "aleksey.maksi...@it-kb.ru" 
>>>>>> <aleksey.maksi...@it-kb.ru>:
>>>>>>>    "Aleksey, was your DNS configured with DNS1 and DNS2 just on the 
>>>>>>> interface you used to create the management bridge on?"
>>>>>>>
>>>>>>>    Yes. Of course
>>>>>>>
>>>>>>>    25.07.2016, 12:27, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>>>>     On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski
>>>>>>>>     <piotr.kliczew...@gmail.com> wrote:
>>>>>>>>>      This could be the issue here as well as for BZ #1358530
>>>>>>>>>
>>>>>>>>>      On Mon, Jul 25, 2016 at 10:53 AM, <aleksey.maksi...@it-kb.ru> 
>>>>>>>>> wrote:
>>>>>>>>>>      Could this be due to the fact that the ovirt installer has 
>>>>>>>>>> changed network configuration files (ifcfg-*, resolv.conf) ?
>>>>>>>>>>      After the error in ovirt installation process I see from 
>>>>>>>>>> resolv.conf disappeared on my DNS servers entry and now the server 
>>>>>>>>>> is unable to resolve names.
>>>>>>>>
>>>>>>>>     So it could be related to 
>>>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1160423
>>>>>>>>
>>>>>>>>     Aleksey, was your DNS configured with DNS1 and DNS2 just on the
>>>>>>>>     interface you used to create the management bridge on?
>>>>>>>>     Can you please try the workaround described here
>>>>>>>>     https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ?
>>>>>>>>
>>>>>>>>>>      25.07.2016, 11:26, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>>>>>>>      On Mon, Jul 25, 2016 at 10:22 AM, <aleksey.maksi...@it-kb.ru> 
>>>>>>>>>>> wrote:
>>>>>>>>>>>>       # vdsClient -s 0 getVdsCaps
>>>>>>>>>>>>
>>>>>>>>>>>>       Traceback (most recent call last):
>>>>>>>>>>>>         File "/usr/share/vdsm/vdsClient.py", line 2980, in <module>
>>>>>>>>>>>>           code, message = commands[command][0](commandArgs)
>>>>>>>>>>>>         File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
>>>>>>>>>>>>           return self.ExecAndExit(self.s.getVdsCapabilities())
>>>>>>>>>>>>         File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in 
>>>>>>>>>>>> __call__
>>>>>>>>>>>>           return self.__send(self.__name, args)
>>>>>>>>>>>>         File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in 
>>>>>>>>>>>> __request
>>>>>>>>>>>>           verbose=self.__verbose
>>>>>>>>>>>>         File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in 
>>>>>>>>>>>> request
>>>>>>>>>>>>           return self.single_request(host, handler, request_body, 
>>>>>>>>>>>> verbose)
>>>>>>>>>>>>         File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in 
>>>>>>>>>>>> single_request
>>>>>>>>>>>>           self.send_content(h, request_body)
>>>>>>>>>>>>         File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in 
>>>>>>>>>>>> send_content
>>>>>>>>>>>>           connection.endheaders(request_body)
>>>>>>>>>>>>         File "/usr/lib64/python2.7/httplib.py", line 975, in 
>>>>>>>>>>>> endheaders
>>>>>>>>>>>>           self._send_output(message_body)
>>>>>>>>>>>>         File "/usr/lib64/python2.7/httplib.py", line 835, in 
>>>>>>>>>>>> _send_output
>>>>>>>>>>>>           self.send(msg)
>>>>>>>>>>>>         File "/usr/lib64/python2.7/httplib.py", line 797, in send
>>>>>>>>>>>>           self.connect()
>>>>>>>>>>>>         File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", 
>>>>>>>>>>>> line 203, in connect
>>>>>>>>>>>>           sock = socket.create_connection((self.host, self.port), 
>>>>>>>>>>>> self.timeout)
>>>>>>>>>>>>         File "/usr/lib64/python2.7/socket.py", line 571, in 
>>>>>>>>>>>> create_connection
>>>>>>>>>>>>           raise err
>>>>>>>>>>>>       error: [Errno 101] Network is unreachable
>>>>>>>>>>>
>>>>>>>>>>>      Yaniv, can you please take also a look to this one?
>>>>>>>>>>>      it's exactly the opposite of 
>>>>>>>>>>> https://bugzilla.redhat.com/1358530
>>>>>>>>>>>      Here the jsonrpcclient works but not the xmlrpc one.
>>>>>>>>>>>
>>>>>>>>>>>>       25.07.2016, 11:17, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>>>>>>>>>       On Mon, Jul 25, 2016 at 7:51 AM, 
>>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>>>>>>>>>        Simone, there is something interesting in the vdsm.log?
>>>>>>>>>>>>>
>>>>>>>>>>>>>       For what I saw the issue is not related to the storage but 
>>>>>>>>>>>>> to the network.
>>>>>>>>>>>>>       ovirt-hosted-engine-setup uses the jsonrpc client, instead 
>>>>>>>>>>>>> the code
>>>>>>>>>>>>>       from ovirt-hosted-engine-ha still uses the xmlrpc client 
>>>>>>>>>>>>> somewhere and
>>>>>>>>>>>>>       this happens also when the setup asks to create the 
>>>>>>>>>>>>> lockspace volume.
>>>>>>>>>>>>>       It seams that in your case the xmlrpc client could not 
>>>>>>>>>>>>> connect vdsm on
>>>>>>>>>>>>>       the localhost.
>>>>>>>>>>>>>       It could be somehow related to:
>>>>>>>>>>>>>       https://bugzilla.redhat.com/1358530
>>>>>>>>>>>>>
>>>>>>>>>>>>>       Can you please try executing
>>>>>>>>>>>>>        sudo vdsClient -s 0 getVdsCaps
>>>>>>>>>>>>>       on that host?
>>>>>>>>>>>>>
>>>>>>>>>>>>>>        22.07.2016, 19:36, "aleksey.maksi...@it-kb.ru" 
>>>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru>:
>>>>>>>>>>>>>>>        Simone, thanks for link.
>>>>>>>>>>>>>>>        vdsm.log attached
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>        22.07.2016, 19:28, "Simone Tiraboschi" 
>>>>>>>>>>>>>>> <stira...@redhat.com>:
>>>>>>>>>>>>>>>>         On Fri, Jul 22, 2016 at 5:59 PM, 
>>>>>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>>>>>>>>>>>>          Thank you for your response, Simone.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          Log attached.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>         It seams it comes from VDSM, can you please attach 
>>>>>>>>>>>>>>>> also vdsm.log?
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          I don't use ovirt-engine-appliance because I have 
>>>>>>>>>>>>>>>>> not found "how-to" for ovirt-engine-appliance deployment in 
>>>>>>>>>>>>>>>>> hosted engine configuration.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>         yum install ovirt-engine-appliance
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>         Then follow the instruction here:
>>>>>>>>>>>>>>>>         
>>>>>>>>>>>>>>>> http://www.ovirt.org/develop/release-management/features/heapplianceflow/
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          22.07.2016, 17:09, "Simone Tiraboschi" 
>>>>>>>>>>>>>>>>> <stira...@redhat.com>:
>>>>>>>>>>>>>>>>>>          Hi Aleksey,
>>>>>>>>>>>>>>>>>>          Can you please attach hosted-engine-setup logs?
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>          On Fri, Jul 22, 2016 at 3:46 PM, 
>>>>>>>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           Hello oVirt guru`s !
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           I have problem with initial deploy of ovirt 4.0 
>>>>>>>>>>>>>>>>>>> hosted engine.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           My environment :
>>>>>>>>>>>>>>>>>>>           ============================
>>>>>>>>>>>>>>>>>>>           * Two servers HP ProLiant DL 360 G5 with Qlogic 
>>>>>>>>>>>>>>>>>>> FC HBA connected (with
>>>>>>>>>>>>>>>>>>>           multipathd) to storage HP 3PAR 7200
>>>>>>>>>>>>>>>>>>>           * On each server installed CentOS 7.2 Linux 
>>>>>>>>>>>>>>>>>>> (3.10.0-327.22.2.el7.x86_64)
>>>>>>>>>>>>>>>>>>>           * On 3PAR storage I created 2 LUNs for oVirt.
>>>>>>>>>>>>>>>>>>>           - First LUN for oVirt Hosted Engine VM (60GB)
>>>>>>>>>>>>>>>>>>>           - Second LUN for all other VMs (2TB)
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           # multipath -ll
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           3par-vv1 (360002ac0000000000000001b0000cec9) dm-0 
>>>>>>>>>>>>>>>>>>> 3PARdata,VV
>>>>>>>>>>>>>>>>>>>           size=60G features='1 queue_if_no_path' 
>>>>>>>>>>>>>>>>>>> hwhandler='1 alua' wp=rw
>>>>>>>>>>>>>>>>>>>           `-+- policy='round-robin 0' prio=50 status=active
>>>>>>>>>>>>>>>>>>>             |- 2:0:1:1 sdd 8:48 active ready running
>>>>>>>>>>>>>>>>>>>             |- 3:0:0:1 sdf 8:80 active ready running
>>>>>>>>>>>>>>>>>>>             |- 2:0:0:1 sdb 8:16 active ready running
>>>>>>>>>>>>>>>>>>>             `- 3:0:1:1 sdh 8:112 active ready running
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           3par-vv2 (360002ac000000000000000160000cec9) dm-1 
>>>>>>>>>>>>>>>>>>> 3PARdata,VV
>>>>>>>>>>>>>>>>>>>           size=2.0T features='1 queue_if_no_path' 
>>>>>>>>>>>>>>>>>>> hwhandler='1 alua' wp=rw
>>>>>>>>>>>>>>>>>>>           `-+- policy='round-robin 0' prio=50 status=active
>>>>>>>>>>>>>>>>>>>             |- 2:0:0:0 sda 8:0 active ready running
>>>>>>>>>>>>>>>>>>>             |- 3:0:0:0 sde 8:64 active ready running
>>>>>>>>>>>>>>>>>>>             |- 2:0:1:0 sdc 8:32 active ready running
>>>>>>>>>>>>>>>>>>>             `- 3:0:1:0 sdg 8:96 active ready running
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           My steps on first server (initial deploy of ovirt 
>>>>>>>>>>>>>>>>>>> 4.0 hosted engine):
>>>>>>>>>>>>>>>>>>>           ============================
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           # systemctl stop NetworkManager
>>>>>>>>>>>>>>>>>>>           # systemctl disable NetworkManager
>>>>>>>>>>>>>>>>>>>           # yum -y install 
>>>>>>>>>>>>>>>>>>> http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm
>>>>>>>>>>>>>>>>>>>           # yum -y install epel-release
>>>>>>>>>>>>>>>>>>>           # wget
>>>>>>>>>>>>>>>>>>>           
>>>>>>>>>>>>>>>>>>> http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1511.iso
>>>>>>>>>>>>>>>>>>>           -P /tmp/
>>>>>>>>>>>>>>>>>>>           # yum install ovirt-hosted-engine-setup
>>>>>>>>>>>>>>>>>>>           # yum install screen
>>>>>>>>>>>>>>>>>>>           # screen -RD
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           ...in screen session :
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           # hosted-engine --deploy
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           ...
>>>>>>>>>>>>>>>>>>>           in configuration process I chose "fc" as storage 
>>>>>>>>>>>>>>>>>>> type for oVirt hosted
>>>>>>>>>>>>>>>>>>>           engine vm and select 60GB LUN...
>>>>>>>>>>>>>>>>>>>           ...
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           --== CONFIGURATION PREVIEW ==--
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           ...
>>>>>>>>>>>>>>>>>>>                     Firewall manager : iptables
>>>>>>>>>>>>>>>>>>>                     Gateway address : 10.1.0.1
>>>>>>>>>>>>>>>>>>>                     Host name for web application : 
>>>>>>>>>>>>>>>>>>> KOM-AD01-OVIRT1
>>>>>>>>>>>>>>>>>>>                     Storage Domain type : fc
>>>>>>>>>>>>>>>>>>>                     Host ID : 1
>>>>>>>>>>>>>>>>>>>                     LUN ID :
>>>>>>>>>>>>>>>>>>>           360002ac0000000000000001b0000cec9
>>>>>>>>>>>>>>>>>>>                     Image size GB : 40
>>>>>>>>>>>>>>>>>>>                     Console type : vnc
>>>>>>>>>>>>>>>>>>>                     Memory size MB : 4096
>>>>>>>>>>>>>>>>>>>                     MAC address : 00:16:3e:77:1d:07
>>>>>>>>>>>>>>>>>>>                     Boot type : cdrom
>>>>>>>>>>>>>>>>>>>                     Number of CPUs : 2
>>>>>>>>>>>>>>>>>>>                     ISO image (cdrom boot/cloud-init) :
>>>>>>>>>>>>>>>>>>>           /tmp/CentOS-7-x86_64-NetInstall-1511.iso
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>          Can I ask why you prefer/need to manually create a 
>>>>>>>>>>>>>>>>>> VM installing from
>>>>>>>>>>>>>>>>>>          a CD instead of using the ready-to-use 
>>>>>>>>>>>>>>>>>> ovirt-engine-appliance?
>>>>>>>>>>>>>>>>>>          Using the appliance makes the setup process a lot 
>>>>>>>>>>>>>>>>>> shorted and more comfortable.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>                     CPU Type : model_Penryn
>>>>>>>>>>>>>>>>>>>           ...
>>>>>>>>>>>>>>>>>>>           and get error after step "Verifying sanlock 
>>>>>>>>>>>>>>>>>>> lockspace initialization"
>>>>>>>>>>>>>>>>>>>           ...
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           [ INFO ] Verifying sanlock lockspace 
>>>>>>>>>>>>>>>>>>> initialization
>>>>>>>>>>>>>>>>>>>           [ ERROR ] Failed to execute stage 'Misc 
>>>>>>>>>>>>>>>>>>> configuration': [Errno 101] Network
>>>>>>>>>>>>>>>>>>>           is unreachable
>>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Clean up
>>>>>>>>>>>>>>>>>>>           [ INFO ] Generating answer file
>>>>>>>>>>>>>>>>>>>           
>>>>>>>>>>>>>>>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf'
>>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Pre-termination
>>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Termination
>>>>>>>>>>>>>>>>>>>           [ ERROR ] Hosted Engine deployment failed: this 
>>>>>>>>>>>>>>>>>>> system is not reliable,
>>>>>>>>>>>>>>>>>>>           please check the issue, fix and redeploy
>>>>>>>>>>>>>>>>>>>                     Log file is located at
>>>>>>>>>>>>>>>>>>>           
>>>>>>>>>>>>>>>>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160722123404-t26vw0.log
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           Interestingly
>>>>>>>>>>>>>>>>>>>           ============================
>>>>>>>>>>>>>>>>>>>           If I try to deploy hosted-engine v3.6, everything 
>>>>>>>>>>>>>>>>>>> goes well in the same
>>>>>>>>>>>>>>>>>>>           configuration !! :
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           ....
>>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Transaction setup
>>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Misc configuration
>>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Package installation
>>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Misc configuration
>>>>>>>>>>>>>>>>>>>           [ INFO ] Configuring libvirt
>>>>>>>>>>>>>>>>>>>           [ INFO ] Configuring VDSM
>>>>>>>>>>>>>>>>>>>           [ INFO ] Starting vdsmd
>>>>>>>>>>>>>>>>>>>           [ INFO ] Waiting for VDSM hardware info
>>>>>>>>>>>>>>>>>>>           [ INFO ] Configuring the management bridge
>>>>>>>>>>>>>>>>>>>           [ INFO ] Creating Volume Group
>>>>>>>>>>>>>>>>>>>           [ INFO ] Creating Storage Domain
>>>>>>>>>>>>>>>>>>>           [ INFO ] Creating Storage Pool
>>>>>>>>>>>>>>>>>>>           [ INFO ] Connecting Storage Pool
>>>>>>>>>>>>>>>>>>>           [ INFO ] Verifying sanlock lockspace 
>>>>>>>>>>>>>>>>>>> initialization
>>>>>>>>>>>>>>>>>>>           [ INFO ] Creating Image for 
>>>>>>>>>>>>>>>>>>> 'hosted-engine.lockspace' ...
>>>>>>>>>>>>>>>>>>>           [ INFO ] Image for 'hosted-engine.lockspace' 
>>>>>>>>>>>>>>>>>>> created successfully
>>>>>>>>>>>>>>>>>>>           [ INFO ] Creating Image for 
>>>>>>>>>>>>>>>>>>> 'hosted-engine.metadata' ...
>>>>>>>>>>>>>>>>>>>           [ INFO ] Image for 'hosted-engine.metadata' 
>>>>>>>>>>>>>>>>>>> created successfully
>>>>>>>>>>>>>>>>>>>           [ INFO ] Creating VM Image
>>>>>>>>>>>>>>>>>>>           [ INFO ] Destroying Storage Pool
>>>>>>>>>>>>>>>>>>>           [ INFO ] Start monitoring domain
>>>>>>>>>>>>>>>>>>>           [ INFO ] Configuring VM
>>>>>>>>>>>>>>>>>>>           [ INFO ] Updating hosted-engine configuration
>>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Transaction commit
>>>>>>>>>>>>>>>>>>>           [ INFO ] Stage: Closing up
>>>>>>>>>>>>>>>>>>>           [ INFO ] Creating VM
>>>>>>>>>>>>>>>>>>>                     You can now connect to the VM with the 
>>>>>>>>>>>>>>>>>>> following command:
>>>>>>>>>>>>>>>>>>>                           /bin/remote-viewer 
>>>>>>>>>>>>>>>>>>> vnc://localhost:5900
>>>>>>>>>>>>>>>>>>>           ...
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           What could be the problem?
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>           _______________________________________________
>>>>>>>>>>>>>>>>>>>           Users mailing list
>>>>>>>>>>>>>>>>>>>           Users@ovirt.org
>>>>>>>>>>>>>>>>>>>           http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>      _______________________________________________
>>>>>>>>>>      Users mailing list
>>>>>>>>>>      Users@ovirt.org
>>>>>>>>>>      http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>
>>>>>>>    _______________________________________________
>>>>>>>    Users mailing list
>>>>>>>    Users@ovirt.org
>>>>>>>    http://lists.ovirt.org/mailman/listinfo/users
>>>>
>>>>  _______________________________________________
>>>>  Users mailing list
>>>>  Users@ovirt.org
>>>>  http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to