On Mon, Jul 25, 2016 at 11:54 AM,  <aleksey.maksi...@it-kb.ru> wrote:
> What am I supposed to do for successfully deploy ovirt 4 ?
> Any ideas ?

Can you please try to explicitly configure your DNS with nameserver
under /etc/resolv.conf and remove DNS1 and DNS2 and set PEERDNS=no for
the interface you are going to use?

> 25.07.2016, 12:47, "aleksey.maksi...@it-kb.ru" <aleksey.maksi...@it-kb.ru>:
>> "Aleksey, was your DNS configured with DNS1 and DNS2 just on the interface 
>> you used to create the management bridge on?"
>>
>> Yes. Of course
>>
>> 25.07.2016, 12:27, "Simone Tiraboschi" <stira...@redhat.com>:
>>>  On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski
>>>  <piotr.kliczew...@gmail.com> wrote:
>>>>   This could be the issue here as well as for BZ #1358530
>>>>
>>>>   On Mon, Jul 25, 2016 at 10:53 AM, <aleksey.maksi...@it-kb.ru> wrote:
>>>>>   Could this be due to the fact that the ovirt installer has changed 
>>>>> network configuration files (ifcfg-*, resolv.conf) ?
>>>>>   After the error in ovirt installation process I see from resolv.conf 
>>>>> disappeared on my DNS servers entry and now the server is unable to 
>>>>> resolve names.
>>>
>>>  So it could be related to 
>>> https://bugzilla.redhat.com/show_bug.cgi?id=1160423
>>>
>>>  Aleksey, was your DNS configured with DNS1 and DNS2 just on the
>>>  interface you used to create the management bridge on?
>>>  Can you please try the workaround described here
>>>  https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ?
>>>
>>>>>   25.07.2016, 11:26, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>>   On Mon, Jul 25, 2016 at 10:22 AM, <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>>    # vdsClient -s 0 getVdsCaps
>>>>>>>
>>>>>>>    Traceback (most recent call last):
>>>>>>>      File "/usr/share/vdsm/vdsClient.py", line 2980, in <module>
>>>>>>>        code, message = commands[command][0](commandArgs)
>>>>>>>      File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
>>>>>>>        return self.ExecAndExit(self.s.getVdsCapabilities())
>>>>>>>      File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
>>>>>>>        return self.__send(self.__name, args)
>>>>>>>      File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
>>>>>>>        verbose=self.__verbose
>>>>>>>      File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
>>>>>>>        return self.single_request(host, handler, request_body, verbose)
>>>>>>>      File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in 
>>>>>>> single_request
>>>>>>>        self.send_content(h, request_body)
>>>>>>>      File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in 
>>>>>>> send_content
>>>>>>>        connection.endheaders(request_body)
>>>>>>>      File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders
>>>>>>>        self._send_output(message_body)
>>>>>>>      File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output
>>>>>>>        self.send(msg)
>>>>>>>      File "/usr/lib64/python2.7/httplib.py", line 797, in send
>>>>>>>        self.connect()
>>>>>>>      File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 
>>>>>>> 203, in connect
>>>>>>>        sock = socket.create_connection((self.host, self.port), 
>>>>>>> self.timeout)
>>>>>>>      File "/usr/lib64/python2.7/socket.py", line 571, in 
>>>>>>> create_connection
>>>>>>>        raise err
>>>>>>>    error: [Errno 101] Network is unreachable
>>>>>>
>>>>>>   Yaniv, can you please take also a look to this one?
>>>>>>   it's exactly the opposite of https://bugzilla.redhat.com/1358530
>>>>>>   Here the jsonrpcclient works but not the xmlrpc one.
>>>>>>
>>>>>>>    25.07.2016, 11:17, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>>>>    On Mon, Jul 25, 2016 at 7:51 AM, <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>>>>     Simone, there is something interesting in the vdsm.log?
>>>>>>>>
>>>>>>>>    For what I saw the issue is not related to the storage but to the 
>>>>>>>> network.
>>>>>>>>    ovirt-hosted-engine-setup uses the jsonrpc client, instead the code
>>>>>>>>    from ovirt-hosted-engine-ha still uses the xmlrpc client somewhere 
>>>>>>>> and
>>>>>>>>    this happens also when the setup asks to create the lockspace 
>>>>>>>> volume.
>>>>>>>>    It seams that in your case the xmlrpc client could not connect vdsm 
>>>>>>>> on
>>>>>>>>    the localhost.
>>>>>>>>    It could be somehow related to:
>>>>>>>>    https://bugzilla.redhat.com/1358530
>>>>>>>>
>>>>>>>>    Can you please try executing
>>>>>>>>     sudo vdsClient -s 0 getVdsCaps
>>>>>>>>    on that host?
>>>>>>>>
>>>>>>>>>     22.07.2016, 19:36, "aleksey.maksi...@it-kb.ru" 
>>>>>>>>> <aleksey.maksi...@it-kb.ru>:
>>>>>>>>>>     Simone, thanks for link.
>>>>>>>>>>     vdsm.log attached
>>>>>>>>>>
>>>>>>>>>>     22.07.2016, 19:28, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>>>>>>>      On Fri, Jul 22, 2016 at 5:59 PM, <aleksey.maksi...@it-kb.ru> 
>>>>>>>>>>> wrote:
>>>>>>>>>>>>       Thank you for your response, Simone.
>>>>>>>>>>>>
>>>>>>>>>>>>       Log attached.
>>>>>>>>>>>
>>>>>>>>>>>      It seams it comes from VDSM, can you please attach also 
>>>>>>>>>>> vdsm.log?
>>>>>>>>>>>
>>>>>>>>>>>>       I don't use ovirt-engine-appliance because I have not found 
>>>>>>>>>>>> "how-to" for ovirt-engine-appliance deployment in hosted engine 
>>>>>>>>>>>> configuration.
>>>>>>>>>>>
>>>>>>>>>>>      yum install ovirt-engine-appliance
>>>>>>>>>>>
>>>>>>>>>>>      Then follow the instruction here:
>>>>>>>>>>>      
>>>>>>>>>>> http://www.ovirt.org/develop/release-management/features/heapplianceflow/
>>>>>>>>>>>
>>>>>>>>>>>>       22.07.2016, 17:09, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>>>>>>>>>       Hi Aleksey,
>>>>>>>>>>>>>       Can you please attach hosted-engine-setup logs?
>>>>>>>>>>>>>
>>>>>>>>>>>>>       On Fri, Jul 22, 2016 at 3:46 PM, 
>>>>>>>>>>>>> <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>>        Hello oVirt guru`s !
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        I have problem with initial deploy of ovirt 4.0 hosted 
>>>>>>>>>>>>>> engine.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        My environment :
>>>>>>>>>>>>>>        ============================
>>>>>>>>>>>>>>        * Two servers HP ProLiant DL 360 G5 with Qlogic FC HBA 
>>>>>>>>>>>>>> connected (with
>>>>>>>>>>>>>>        multipathd) to storage HP 3PAR 7200
>>>>>>>>>>>>>>        * On each server installed CentOS 7.2 Linux 
>>>>>>>>>>>>>> (3.10.0-327.22.2.el7.x86_64)
>>>>>>>>>>>>>>        * On 3PAR storage I created 2 LUNs for oVirt.
>>>>>>>>>>>>>>        - First LUN for oVirt Hosted Engine VM (60GB)
>>>>>>>>>>>>>>        - Second LUN for all other VMs (2TB)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        # multipath -ll
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        3par-vv1 (360002ac0000000000000001b0000cec9) dm-0 
>>>>>>>>>>>>>> 3PARdata,VV
>>>>>>>>>>>>>>        size=60G features='1 queue_if_no_path' hwhandler='1 alua' 
>>>>>>>>>>>>>> wp=rw
>>>>>>>>>>>>>>        `-+- policy='round-robin 0' prio=50 status=active
>>>>>>>>>>>>>>          |- 2:0:1:1 sdd 8:48 active ready running
>>>>>>>>>>>>>>          |- 3:0:0:1 sdf 8:80 active ready running
>>>>>>>>>>>>>>          |- 2:0:0:1 sdb 8:16 active ready running
>>>>>>>>>>>>>>          `- 3:0:1:1 sdh 8:112 active ready running
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        3par-vv2 (360002ac000000000000000160000cec9) dm-1 
>>>>>>>>>>>>>> 3PARdata,VV
>>>>>>>>>>>>>>        size=2.0T features='1 queue_if_no_path' hwhandler='1 
>>>>>>>>>>>>>> alua' wp=rw
>>>>>>>>>>>>>>        `-+- policy='round-robin 0' prio=50 status=active
>>>>>>>>>>>>>>          |- 2:0:0:0 sda 8:0 active ready running
>>>>>>>>>>>>>>          |- 3:0:0:0 sde 8:64 active ready running
>>>>>>>>>>>>>>          |- 2:0:1:0 sdc 8:32 active ready running
>>>>>>>>>>>>>>          `- 3:0:1:0 sdg 8:96 active ready running
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        My steps on first server (initial deploy of ovirt 4.0 
>>>>>>>>>>>>>> hosted engine):
>>>>>>>>>>>>>>        ============================
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        # systemctl stop NetworkManager
>>>>>>>>>>>>>>        # systemctl disable NetworkManager
>>>>>>>>>>>>>>        # yum -y install 
>>>>>>>>>>>>>> http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm
>>>>>>>>>>>>>>        # yum -y install epel-release
>>>>>>>>>>>>>>        # wget
>>>>>>>>>>>>>>        
>>>>>>>>>>>>>> http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1511.iso
>>>>>>>>>>>>>>        -P /tmp/
>>>>>>>>>>>>>>        # yum install ovirt-hosted-engine-setup
>>>>>>>>>>>>>>        # yum install screen
>>>>>>>>>>>>>>        # screen -RD
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        ...in screen session :
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        # hosted-engine --deploy
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        ...
>>>>>>>>>>>>>>        in configuration process I chose "fc" as storage type for 
>>>>>>>>>>>>>> oVirt hosted
>>>>>>>>>>>>>>        engine vm and select 60GB LUN...
>>>>>>>>>>>>>>        ...
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        --== CONFIGURATION PREVIEW ==--
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        ...
>>>>>>>>>>>>>>                  Firewall manager : iptables
>>>>>>>>>>>>>>                  Gateway address : 10.1.0.1
>>>>>>>>>>>>>>                  Host name for web application : KOM-AD01-OVIRT1
>>>>>>>>>>>>>>                  Storage Domain type : fc
>>>>>>>>>>>>>>                  Host ID : 1
>>>>>>>>>>>>>>                  LUN ID :
>>>>>>>>>>>>>>        360002ac0000000000000001b0000cec9
>>>>>>>>>>>>>>                  Image size GB : 40
>>>>>>>>>>>>>>                  Console type : vnc
>>>>>>>>>>>>>>                  Memory size MB : 4096
>>>>>>>>>>>>>>                  MAC address : 00:16:3e:77:1d:07
>>>>>>>>>>>>>>                  Boot type : cdrom
>>>>>>>>>>>>>>                  Number of CPUs : 2
>>>>>>>>>>>>>>                  ISO image (cdrom boot/cloud-init) :
>>>>>>>>>>>>>>        /tmp/CentOS-7-x86_64-NetInstall-1511.iso
>>>>>>>>>>>>>
>>>>>>>>>>>>>       Can I ask why you prefer/need to manually create a VM 
>>>>>>>>>>>>> installing from
>>>>>>>>>>>>>       a CD instead of using the ready-to-use 
>>>>>>>>>>>>> ovirt-engine-appliance?
>>>>>>>>>>>>>       Using the appliance makes the setup process a lot shorted 
>>>>>>>>>>>>> and more comfortable.
>>>>>>>>>>>>>
>>>>>>>>>>>>>>                  CPU Type : model_Penryn
>>>>>>>>>>>>>>        ...
>>>>>>>>>>>>>>        and get error after step "Verifying sanlock lockspace 
>>>>>>>>>>>>>> initialization"
>>>>>>>>>>>>>>        ...
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        [ INFO ] Verifying sanlock lockspace initialization
>>>>>>>>>>>>>>        [ ERROR ] Failed to execute stage 'Misc configuration': 
>>>>>>>>>>>>>> [Errno 101] Network
>>>>>>>>>>>>>>        is unreachable
>>>>>>>>>>>>>>        [ INFO ] Stage: Clean up
>>>>>>>>>>>>>>        [ INFO ] Generating answer file
>>>>>>>>>>>>>>        
>>>>>>>>>>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf'
>>>>>>>>>>>>>>        [ INFO ] Stage: Pre-termination
>>>>>>>>>>>>>>        [ INFO ] Stage: Termination
>>>>>>>>>>>>>>        [ ERROR ] Hosted Engine deployment failed: this system is 
>>>>>>>>>>>>>> not reliable,
>>>>>>>>>>>>>>        please check the issue, fix and redeploy
>>>>>>>>>>>>>>                  Log file is located at
>>>>>>>>>>>>>>        
>>>>>>>>>>>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160722123404-t26vw0.log
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        Interestingly
>>>>>>>>>>>>>>        ============================
>>>>>>>>>>>>>>        If I try to deploy hosted-engine v3.6, everything goes 
>>>>>>>>>>>>>> well in the same
>>>>>>>>>>>>>>        configuration !! :
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        ....
>>>>>>>>>>>>>>        [ INFO ] Stage: Transaction setup
>>>>>>>>>>>>>>        [ INFO ] Stage: Misc configuration
>>>>>>>>>>>>>>        [ INFO ] Stage: Package installation
>>>>>>>>>>>>>>        [ INFO ] Stage: Misc configuration
>>>>>>>>>>>>>>        [ INFO ] Configuring libvirt
>>>>>>>>>>>>>>        [ INFO ] Configuring VDSM
>>>>>>>>>>>>>>        [ INFO ] Starting vdsmd
>>>>>>>>>>>>>>        [ INFO ] Waiting for VDSM hardware info
>>>>>>>>>>>>>>        [ INFO ] Configuring the management bridge
>>>>>>>>>>>>>>        [ INFO ] Creating Volume Group
>>>>>>>>>>>>>>        [ INFO ] Creating Storage Domain
>>>>>>>>>>>>>>        [ INFO ] Creating Storage Pool
>>>>>>>>>>>>>>        [ INFO ] Connecting Storage Pool
>>>>>>>>>>>>>>        [ INFO ] Verifying sanlock lockspace initialization
>>>>>>>>>>>>>>        [ INFO ] Creating Image for 'hosted-engine.lockspace' ...
>>>>>>>>>>>>>>        [ INFO ] Image for 'hosted-engine.lockspace' created 
>>>>>>>>>>>>>> successfully
>>>>>>>>>>>>>>        [ INFO ] Creating Image for 'hosted-engine.metadata' ...
>>>>>>>>>>>>>>        [ INFO ] Image for 'hosted-engine.metadata' created 
>>>>>>>>>>>>>> successfully
>>>>>>>>>>>>>>        [ INFO ] Creating VM Image
>>>>>>>>>>>>>>        [ INFO ] Destroying Storage Pool
>>>>>>>>>>>>>>        [ INFO ] Start monitoring domain
>>>>>>>>>>>>>>        [ INFO ] Configuring VM
>>>>>>>>>>>>>>        [ INFO ] Updating hosted-engine configuration
>>>>>>>>>>>>>>        [ INFO ] Stage: Transaction commit
>>>>>>>>>>>>>>        [ INFO ] Stage: Closing up
>>>>>>>>>>>>>>        [ INFO ] Creating VM
>>>>>>>>>>>>>>                  You can now connect to the VM with the 
>>>>>>>>>>>>>> following command:
>>>>>>>>>>>>>>                        /bin/remote-viewer vnc://localhost:5900
>>>>>>>>>>>>>>        ...
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        What could be the problem?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        _______________________________________________
>>>>>>>>>>>>>>        Users mailing list
>>>>>>>>>>>>>>        Users@ovirt.org
>>>>>>>>>>>>>>        http://lists.ovirt.org/mailman/listinfo/users
>>>>>   _______________________________________________
>>>>>   Users mailing list
>>>>>   Users@ovirt.org
>>>>>   http://lists.ovirt.org/mailman/listinfo/users
>>
>> _______________________________________________
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to