Re: [ovirt-users] Fwd: Re: HA agent fails to start

2016-04-15 Thread Simone Tiraboschi
On Fri, Apr 15, 2016 at 8:45 AM, Richard Neuboeck <h...@tbi.univie.ac.at> wrote:
> On 04/14/2016 11:03 PM, Simone Tiraboschi wrote:
>> On Thu, Apr 14, 2016 at 10:38 PM, Simone Tiraboschi <stira...@redhat.com> 
>> wrote:
>>> On Thu, Apr 14, 2016 at 6:53 PM, Richard Neuboeck <h...@tbi.univie.ac.at> 
>>> wrote:
>>>> On 14.04.16 18:46, Simone Tiraboschi wrote:
>>>>> On Thu, Apr 14, 2016 at 4:04 PM, Richard Neuboeck <h...@tbi.univie.ac.at> 
>>>>> wrote:
>>>>>> On 04/14/2016 02:14 PM, Simone Tiraboschi wrote:
>>>>>>> On Thu, Apr 14, 2016 at 12:51 PM, Richard Neuboeck
>>>>>>> <h...@tbi.univie.ac.at> wrote:
>>>>>>>> On 04/13/2016 10:00 AM, Simone Tiraboschi wrote:
>>>>>>>>> On Wed, Apr 13, 2016 at 9:38 AM, Richard Neuboeck 
>>>>>>>>> <h...@tbi.univie.ac.at> wrote:
>>>>>>>>>> The answers file shows the setup time of both machines.
>>>>>>>>>>
>>>>>>>>>> On both machines hosted-engine.conf got rotated right before I wrote
>>>>>>>>>> this mail. Is it possible that I managed to interrupt the rotation 
>>>>>>>>>> with
>>>>>>>>>> the reboot so the backup was accurate but the update not yet written 
>>>>>>>>>> to
>>>>>>>>>> hosted-engine.conf?
>>>>>>>>>
>>>>>>>>> AFAIK we don't have any rotation mechanism for that file; something
>>>>>>>>> else you have in place on that host?
>>>>>>>>
>>>>>>>> Those machines are all CentOS 7.2 minimal installs. The only
>>>>>>>> adaptation I do is installing vim, removing postfix and installing
>>>>>>>> exim, removing firewalld and installing iptables-service. Then I add
>>>>>>>> the oVirt repos (3.6 and 3.6-snapshot) and deploy the host.
>>>>>>>>
>>>>>>>> But checking lsof shows that 'ovirt-ha-agent --no-daemon' has access
>>>>>>>> to the config file (and the one ending with ~):
>>>>>>>>
>>>>>>>> # lsof | grep 'hosted-engine.conf~'
>>>>>>>> ovirt-ha- 193446   vdsm  351u  REG
>>>>>>>> 253,01021135070683
>>>>>>>> /etc/ovirt-hosted-engine/hosted-engine.conf~
>>>>>>>
>>>>>>> This is not that much relevant if the file was renamed after
>>>>>>> ovirt-ha-agent opened it.
>>>>>>> Try this:
>>>>>>>
>>>>>>> [root@c72he20160405h1 ovirt-hosted-engine-setup]# tail -n1 -f
>>>>>>> /etc/ovirt-hosted-engine/hosted-engine.conf &
>>>>>>> [1] 28866
>>>>>>> [root@c72he20160405h1 ovirt-hosted-engine-setup]# port=
>>>>>>>
>>>>>>> [root@c72he20160405h1 ovirt-hosted-engine-setup]# lsof | grep 
>>>>>>> hosted-engine.conf
>>>>>>> tail  28866  root3r  REG
>>>>>>> 253,0  10141595898 /etc/ovirt-hosted-engine/hosted-engine.conf
>>>>>>> [root@c72he20160405h1 ovirt-hosted-engine-setup]# mv
>>>>>>> /etc/ovirt-hosted-engine/hosted-engine.conf
>>>>>>> /etc/ovirt-hosted-engine/hosted-engine.conf_123
>>>>>>> [root@c72he20160405h1 ovirt-hosted-engine-setup]# lsof | grep 
>>>>>>> hosted-engine.conf
>>>>>>> tail  28866  root3r  REG
>>>>>>> 253,0  10141595898
>>>>>>> /etc/ovirt-hosted-engine/hosted-engine.conf_123
>>>>>>> [root@c72he20160405h1 ovirt-hosted-engine-setup]#
>>>>>>>
>>>>>>
>>>>>> I've issued the commands you suggested but I don't know how that
>>>>>> helps to find the process accessing the config files.
>>>>>>
>>>>>> After moving the hosted-engine.conf file the HA agent crashed
>>>>>> logging the information that the config file is not available.
>>>>>>
>>>>>> Here is the output from every command:
>>>>>>
>>>>>> # tail -n1 -f /etc/ovirt-hosted-engine/hosted-engine.

Re: [ovirt-users] Hosted Engine

2016-05-09 Thread Simone Tiraboschi
On Sun, May 8, 2016 at 1:13 PM, Budur Nagaraju  wrote:
> Hi
>
> Configured hosted engine it was up for few days and all of sudden oVirt
> Engine went down and now unable to bring it up ,
>
> below are the logs details,
>
> http://pastebin.com/xXABPNhF

Hi,
I don't see any error there.

Could you please paste the output of
 hosted-engine --vm-status
?

Can you please share your /var/log/ovirt-hosted-engine-ha/agent.log ?

> Thanks,
> Nagaraju
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] virt-in-virt problem: DHCP failing for a container in a oVirt VM

2016-05-09 Thread Simone Tiraboschi
On Sat, May 7, 2016 at 11:06 AM, Yaniv Kaul  wrote:
>
>
> On Fri, May 6, 2016 at 11:07 PM, Will Dennis  wrote:
>>
>> That’s in iptables, right? I have iptables disabled on my oVirt nodes...
>
>
> No, it's a L2 filter libvirt sets up, I believe using ebtables.
> Y.


There is also a specific VDSM hook to address this configuration:
https://github.com/oVirt/vdsm/tree/master/vdsm_hooks/macspoof

>>
>>
>>
>> From: Yaniv Kaul [mailto:yk...@redhat.com]
>> Sent: Friday, May 06, 2016 3:50 PM
>> To: Will Dennis
>> Subject: Re: [ovirt-users] virt-in-virt problem: DHCP failing for a
>> container in a oVirt VM
>>
>>
>>
>> Long shot - you need to disable the EnableMACAntiSpoofingFilterRules .
>>
>> Y.
>>
>>
>>
>> On Fri, May 6, 2016 at 8:27 PM, Will Dennis  wrote:
>>
>> Hi all,
>>
>>
>>
>> Have an interesting problem – I am running a VM in oVirt that is running
>> Proxmox VE 4.1 OS, which I have spun up a container on.  The container is
>> set for DHCP, and I have verified that it is sending Discover packets as
>> normal, and that these packets are making it out of the Proxmox VM to the
>> oVirt bridge (which is attached to a VLAN sub-interface of a bond
>> interface.) However, these packets do NOT make it past the oVirt bridge. The
>> interesting thing is that the Proxmox VM (as well as any other VM I spin up
>> on oVirt) works fine with DHCP. (I also have other oVirt VMs instantiated
>> which are using LXD to spin up containers, and I have the same problem with
>> those as well.) I checked a bunch of stuff, and the only clue I could find
>> is that it seems that the oVirt bridge is not learning the MAC for the
>> container on the VM, even though it does learn the VM’s MAC, but I can
>> capture DHCP traffic coming from the container off the ‘vnet0’ interface
>> which is joined to that bridge...
>>
>>
>>
>> Info:
>>
>>
>>
>> = off Proxmox VM =
>>
>>
>>
>> Container's MAC address: 32:62:65:61:65:33
>>
>>
>>
>> root@proxmox-02:~# ip link sh
>>
>> 1: lo:  mtu 65536 qdisc noqueue state UNKNOWN mode
>> DEFAULT group default
>>
>> link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>>
>> 2: eth0:  mtu 1500 qdisc pfifo_fast
>> master vmbr0 state UP mode DEFAULT group default qlen 1000
>>
>> link/ether 00:1a:4a:16:01:57 brd ff:ff:ff:ff:ff:ff
>>
>> 3: vmbr0:  mtu 1500 qdisc noqueue state
>> UP mode DEFAULT group default
>>
>> link/ether 00:1a:4a:16:01:57 brd ff:ff:ff:ff:ff:ff
>>
>> 7: veth100i0@if6:  mtu 1500 qdisc
>> pfifo_fast master vmbr0 state UP mode DEFAULT group default qlen 1000
>>
>> link/ether fe:50:4f:3c:bd:b8 brd ff:ff:ff:ff:ff:ff link-netnsid 0
>> <<< veth connection to container
>>
>>
>>
>> root@proxmox-02:~# brctl showmacs vmbr0
>>
>> port no mac addris local?   ageing timer
>>
>>   1 00:12:3f:24:a4:54   no   112.88
>>
>>   1 00:1a:4a:16:01:56   no 0.02
>>
>>   1 00:1a:4a:16:01:57   yes0.00
>>
>>   1 00:1a:4a:16:01:57   yes0.00
>>
>>   1 00:24:50:dd:a2:05   no 1.37
>>
>>   1 18:03:73:e3:be:5a   no21.04
>>
>>   1 18:03:73:e3:ca:24   no 4.23
>>
>>   1 18:03:73:e3:cb:5b   no48.41
>>
>>   1 18:03:73:e3:cc:e5   no91.93
>>
>>   1 18:03:73:e3:cd:b8   no   151.04
>>
>>   1 18:03:73:e3:ce:43   no 0.80
>>
>>   1 18:03:73:e3:d0:a4   no   290.74
>>
>>   1 18:03:73:e3:d4:26   no34.06
>>
>>   1 18:03:73:e3:d5:3d   no 6.36
>>
>>   1 18:03:73:e4:23:08   no88.76
>>
>>   1 18:03:73:e4:25:92   no   111.86
>>
>>   1 18:03:73:e4:26:2f   no 9.54
>>
>>   1 18:03:73:e4:2b:4c   no   114.86
>>
>>   1 18:03:73:e4:31:15   no   263.91
>>
>>   1 18:03:73:e4:6c:19   no 6.36
>>
>>   1 18:03:73:e4:7e:0a   no   103.06
>>
>>   1 18:03:73:e8:16:e0   no23.21
>>
>>   2 32:62:65:61:65:33   no 5.08   <<< container’s
>> MAC learned on Proxmox bridge
>>
>>   1 34:17:eb:9b:e0:29   no   265.22
>>
>>   1 34:17:eb:9b:f8:ea   no   114.86
>>
>>   1 44:d3:ca:7e:3c:ff   no 0.00
>>
>>   1 78:2b:cb:3b:ca:b9   no   284.70
>>
>>   1 78:2b:cb:92:cb:cb   no   279.70
>>
>>   1 78:2b:cb:93:08:a8   no   287.05
>>
>>   1 b8:ca:3a:7a:70:63   no 4.83
>>
>>   1 f8:bc:12:69:bb:a3   no   121.82
>>
>>   2 fe:50:4f:3c:bd:b8   yes0.00
>>
>>   2 

Re: [ovirt-users] Hosted Engine

2016-05-09 Thread Simone Tiraboschi
On Mon, May 9, 2016 at 10:19 AM, Budur Nagaraju <nbud...@gmail.com> wrote:
> Below are the details ,
>
> [root@asglmesx1 ~]#  hosted-engine --vm-status
>
>
> --== Host 1 status ==--
>
> Status up-to-date  : False
> Hostname   : asglmesx1
> Host ID: 1
> Engine status  : unknown stale-data
> Score  : 2400
> Local maintenance  : False
> Host timestamp : 407135
> Extra metadata (valid at timestamp):
> metadata_parse_version=1
> metadata_feature_version=1
> timestamp=407135 (Sun May  8 03:18:26 2016)
> host-id=1
> score=2400
> maintenance=False
> state=EngineUp
>
>
> --== Host 2 status ==--
>
> Status up-to-date  : False
> Hostname   : asglmesx4
> Host ID: 2
> Engine status  : unknown stale-data
> Score  : 2400
> Local maintenance  : False
> Host timestamp : 870090
> Extra metadata (valid at timestamp):
> metadata_parse_version=1
> metadata_feature_version=1
> timestamp=870090 (Mon May  9 13:46:57 2016)
> host-id=2
> score=2400
> maintenance=False
> state=EngineUp
>
>
>
>
>
>
>
> [root@asglmesx1 /]# tail -f /var/log/ovirt-hosted-engine-ha/agent.log
>
> MainThread::ERROR::2016-05-08
> 10:34:46,600::agent::182::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
> Error: 'path to storage domain 85ff2ef7-e08a-4fc4-a9e2-d684117beeca not
> found in /rhev/data-center/mnt' - trying to restart agent
> MainThread::WARNING::2016-05-08
> 10:34:51,614::agent::185::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
> Restarting agent, attempt '9'
> MainThread::ERROR::2016-05-08
> 10:34:51,614::agent::187::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
> Too many errors occurred, giving up. Please review the log and consider
> filing a bug.
> MainThread::INFO::2016-05-08
> 10:34:51,615::agent::121::ovirt_hosted_engine_ha.agent.agent.Agent::(run)
> Agent shutting down

Can you please attach a longer snippet?
ovirt-ha-agent give up since it wasn't able to access the shared
storage for 10 attempt.
where is the hosted-engine storage domain? is it available now?


> On Mon, May 9, 2016 at 1:21 PM, Simone Tiraboschi <stira...@redhat.com>
> wrote:
>>
>> On Sun, May 8, 2016 at 1:13 PM, Budur Nagaraju <nbud...@gmail.com> wrote:
>> > Hi
>> >
>> > Configured hosted engine it was up for few days and all of sudden oVirt
>> > Engine went down and now unable to bring it up ,
>> >
>> > below are the logs details,
>> >
>> > http://pastebin.com/xXABPNhF
>>
>> Hi,
>> I don't see any error there.
>>
>> Could you please paste the output of
>>  hosted-engine --vm-status
>> ?
>>
>> Can you please share your /var/log/ovirt-hosted-engine-ha/agent.log ?
>>
>> > Thanks,
>> > Nagaraju
>> >
>> >
>> > ___
>> > Users mailing list
>> > Users@ovirt.org
>> > http://lists.ovirt.org/mailman/listinfo/users
>> >
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Failed to connect Host to the Storage Domains hosted_storage.

2016-07-22 Thread Simone Tiraboschi
On Fri, Jul 22, 2016 at 3:47 PM, Robert Story  wrote:
> Hello,
>
> I'm in the process of upgrading from 3.5.x to 3.6.x. My hosted engine and
> hosts in the primary cluster are all upgraded and appear to be running fine.
>
> I have a second cluster of 2 machines which are just regular hosts, without
> the hosted-engine. Both have been marked non-operational, with the
> following messages logged about every 5 minutes:
>
>
> Failed to connect Host perses to Storage Pool Default
>
> Host perses cannot access the Storage Domain(s) hosted_storage attached to 
> the Data Center Default. Setting Host state to Non-Operational.
>
> Host perses reports about one of the Active Storage Domains as Problematic.
>
> Failed to connect Host perses to Storage Servers
>
> Failed to connect Host perses to the Storage Domains hosted_storage.
>
>
> I could see the normal storage/iso/export domains mounted on the host, and
> the VMs running on the host are fine.

In 3.5 only the hosts involved in hosted-engine have to access the
hosted-engine storage domain.
With 3.6 we introduced the capabilities to manage the engine VM from
the engine itself so the engine has to import in the hosted-engine
storage domain.
This means that all the hosts in the datacenter that contains the
cluster with the hosted-engine hosts have now to be able to connect
the hosted-engine storage domain.

Can you please check the ACL on the storage server (NFS or iSCSI) that
you use to expose the hosted-engine storage domain?

> I shut down the VMs on one host, put it in maintenance mode, installed 3.6
> repo and ran yum update. All went well, but when I activated the host, same
> deal.
>
> I've attached the engine log snippet for the activation attempt.
>
> Robert
>
> --
> Senior Software Engineer @ Parsons
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4 Hosted Engine deploy on fc storage - [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network is unreachable

2016-07-22 Thread Simone Tiraboschi
Hi Aleksey,
Can you please attach hosted-engine-setup logs?

On Fri, Jul 22, 2016 at 3:46 PM,   wrote:
>
> Hello oVirt guru`s !
>
> I have problem with initial deploy of ovirt 4.0 hosted engine.
>
> My environment :
> 
> * Two servers HP ProLiant DL 360 G5 with Qlogic FC HBA connected (with
> multipathd) to storage HP 3PAR 7200
> * On each server installed CentOS 7.2 Linux (3.10.0-327.22.2.el7.x86_64)
> * On 3PAR storage I created 2 LUNs for oVirt.
> - First LUN for oVirt Hosted Engine VM (60GB)
> - Second LUN for all other VMs (2TB)
>
> # multipath -ll
>
> 3par-vv1 (360002ac0001bcec9) dm-0 3PARdata,VV
> size=60G features='1 queue_if_no_path' hwhandler='1 alua' wp=rw
> `-+- policy='round-robin 0' prio=50 status=active
>   |- 2:0:1:1 sdd 8:48  active ready running
>   |- 3:0:0:1 sdf 8:80  active ready running
>   |- 2:0:0:1 sdb 8:16  active ready running
>   `- 3:0:1:1 sdh 8:112 active ready running
>
> 3par-vv2 (360002ac00016cec9) dm-1 3PARdata,VV
> size=2.0T features='1 queue_if_no_path' hwhandler='1 alua' wp=rw
> `-+- policy='round-robin 0' prio=50 status=active
>   |- 2:0:0:0 sda 8:0   active ready running
>   |- 3:0:0:0 sde 8:64  active ready running
>   |- 2:0:1:0 sdc 8:32  active ready running
>   `- 3:0:1:0 sdg 8:96  active ready running
>
>
>
> My steps on first server (initial deploy of ovirt 4.0 hosted engine):
> 
>
> # systemctl stop NetworkManager
> # systemctl disable NetworkManager
> # yum -y install http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm
> # yum -y install epel-release
> # wget
> http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1511.iso
> -P /tmp/
> # yum install ovirt-hosted-engine-setup
> # yum install screen
> # screen -RD
>
> ...in screen session :
>
> # hosted-engine --deploy
>
> ...
> in configuration process I chose "fc" as storage type for oVirt hosted
> engine vm and select 60GB LUN...
> ...
>
> --== CONFIGURATION PREVIEW ==--
>
> ...
>   Firewall manager   : iptables
>   Gateway address: 10.1.0.1
>   Host name for web application  : KOM-AD01-OVIRT1
>   Storage Domain type: fc
>   Host ID: 1
>   LUN ID :
> 360002ac0001bcec9
>   Image size GB  : 40
>   Console type   : vnc
>   Memory size MB : 4096
>   MAC address: 00:16:3e:77:1d:07
>   Boot type  : cdrom
>   Number of CPUs : 2
>   ISO image (cdrom boot/cloud-init)  :
> /tmp/CentOS-7-x86_64-NetInstall-1511.iso

Can I ask why you prefer/need to manually create a VM installing from
a CD instead of using the ready-to-use ovirt-engine-appliance?
Using the appliance makes the setup process a lot shorted and more comfortable.

>   CPU Type   : model_Penryn
> ...
> and get error after step "Verifying sanlock lockspace initialization"
> ...
>
> [ INFO  ] Verifying sanlock lockspace initialization
> [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network
> is unreachable
> [ INFO  ] Stage: Clean up
> [ INFO  ] Generating answer file
> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf'
> [ INFO  ] Stage: Pre-termination
> [ INFO  ] Stage: Termination
> [ ERROR ] Hosted Engine deployment failed: this system is not reliable,
> please check the issue, fix and redeploy
>   Log file is located at
> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160722123404-t26vw0.log
>
>
> Interestingly
> 
> If I try to deploy hosted-engine v3.6, everything goes well in the same
> configuration !! :
>
> 
> [ INFO  ] Stage: Transaction setup
> [ INFO  ] Stage: Misc configuration
> [ INFO  ] Stage: Package installation
> [ INFO  ] Stage: Misc configuration
> [ INFO  ] Configuring libvirt
> [ INFO  ] Configuring VDSM
> [ INFO  ] Starting vdsmd
> [ INFO  ] Waiting for VDSM hardware info
> [ INFO  ] Configuring the management bridge
> [ INFO  ] Creating Volume Group
> [ INFO  ] Creating Storage Domain
> [ INFO  ] Creating Storage Pool
> [ INFO  ] Connecting Storage Pool
> [ INFO  ] Verifying sanlock lockspace initialization
> [ INFO  ] Creating Image for 'hosted-engine.lockspace' ...
> [ INFO  ] Image for 'hosted-engine.lockspace' created successfully
> [ INFO  ] Creating Image for 'hosted-engine.metadata' ...
> [ INFO  ] Image for 'hosted-engine.metadata' created successfully
> [ INFO  ] Creating VM Image
> [ INFO  ] Destroying Storage Pool
> [ INFO  ] Start monitoring domain
> [ INFO  ] Configuring VM
> [ INFO  ] Updating hosted-engine configuration
> [ INFO  ] Stage: Transaction commit
> [ INFO  ] Stage: Closing up
> [ INFO  ] 

Re: [ovirt-users] Hosted Engine & FC Data Domain

2016-07-26 Thread Simone Tiraboschi
On Tue, Jul 26, 2016 at 12:47 AM, Anantha Raghava
 wrote:
> Hello,
>
> I had recently installed oVirt 4.0.1 Hosted Engine with FC (SAN) Storage.
>
> I could install the oVirt, deploy hosted-engine. But when I accessed the
> Engine Web Admin, I noticed the following:
>
> a. There was no Data Domain. I had to import the FC LUN, on which the Hosted
> Engine's Disk resided into Storage manually. But when I attempted to attach
> the data domain to Data Center, it failed.
>
> b. I ended up removing all configuration, and recreated the FC LUNs. This
> time, I created two LUNs, on one LUN, the Hosted Engine resided. This time,
> first, I had to import the second LUN as master Data Domain and add the LUN
> on which Engine VM resided as just another Data Domain.
>
> Now the questions:
>
> i. Will Hosted Engine not automatically import the hosted_storage as master
> data domain? Why does it throw error when attempted to import and attach to
> DC?

You have just to create the first storage domain for regular VMs which
will become the master storage domain.
At that point the datacenter can go up and the engine will import the
hosted-engine storage domain by itself.
I agree with you that the warning message is a bit confusing and
indeed we have an open bug to change it:
https://bugzilla.redhat.com/show_bug.cgi?id=1358313

> ii. Is it necessary to have another data domain as master data domain,
> before attaching hosted_storage?

Yes, you have.

> iii. How do we change the master data domain from one to another?

Please avoid it.

> --
>
> Thanks & Regards,
>
>
> Anantha Raghava
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4 Hosted Engine deploy on fc storage - [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network is unreachable

2016-07-25 Thread Simone Tiraboschi
pid=3903,fd=3),("hpsmhd",pid=3901,fd=3),("hpsmhd",pid=3900,fd=3),("hpsmhd",pid=3899,fd=3),("hpsmhd",pid=3898,fd=3),("hpsmhd",pid=3893,fd=3))
> tcp   LISTEN 0  30
>*:16514
> *:*   
> users:(("libvirtd",pid=10688,fd=13))
> tcp   LISTEN 0  128   
>127.0.0.1:199  
> *:*   
> users:(("snmpd",pid=1609,fd=9))
> tcp   LISTEN 0  128   
>   :::111  
>:::*   
> users:(("rpcbind",pid=827,fd=11))
> tcp   LISTEN 0  5 
>   :::54321
>    :::*   
> users:(("vdsm",pid=11077,fd=23))

vdsm is properly bind over ipv6.

Can you please check if you can connect to vdsm with:
 telnet  kom-ad01-vm31.holding.com 54321
and with
 telnet ::1 54321
?

> tcp   LISTEN 0  30
>   :::16514
>:::*   
> users:(("libvirtd",pid=10688,fd=14))
>
> 25.07.2016, 15:11, "Simone Tiraboschi" <stira...@redhat.com>:
>> On Mon, Jul 25, 2016 at 2:03 PM, <aleksey.maksi...@it-kb.ru> wrote:
>>>  Yes.
>>>
>>>  # ping $(python -c 'import socket; print(socket.gethostname())')
>>>
>>>  PING KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data.
>>>  64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1 ttl=64 
>>> time=0.030 ms
>>>  64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2 ttl=64 
>>> time=0.015 ms
>>>  64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3 ttl=64 
>>> time=0.011 ms
>>>  ^C
>>>  --- KOM-AD01-VM31.holding.com ping statistics ---
>>>  3 packets transmitted, 3 received, 0% packet loss, time 2001ms
>>>  rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms
>>>
>>>  but...
>>
>> and the output of
>>  ss -plutn
>>
>>>  # vdsClient -s 0 getVdsCaps
>>>
>>>  Traceback (most recent call last):
>>>File "/usr/share/vdsm/vdsClient.py", line 2980, in 
>>>  code, message = commands[command][0](commandArgs)
>>>File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
>>>  return self.ExecAndExit(self.s.getVdsCapabilities())
>>>File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
>>>  return self.__send(self.__name, args)
>>>File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
>>>  verbose=self.__verbose
>>>File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
>>>  return self.single_request(host, handler, request_body, verbose)
>>>File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request
>>>  self.send_content(h, request_body)
>>>File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content
>>>  connection.endheaders(request_body)
>>>File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders
>>>  self._send_output(message_body)
>>>File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output
>>>  self.send(msg)
>>>File "/usr/lib64/python2.7/httplib.py", line 797, in send
>>>  self.connect()
>>>File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in 
>>> connect
>>>  sock = socket.create_connection((self.host, self.port), self.timeout)
>>>File "/usr/lib64/python2.7/socket.py", line 571, in create_connection
>>>  raise err
>>>  error: [Errno 101] Network is unreachable
>>>
>>>  25.07.2016, 14:58, "aleksey.maksi...@it-kb.ru" <aleksey.maksi...@it-kb.ru>:
>>>>  Ok.
>>>

Re: [ovirt-users] oVirt 4 Hosted Engine deploy on fc storage - [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network is unreachable

2016-07-25 Thread Simone Tiraboschi
On Mon, Jul 25, 2016 at 11:54 AM,  <aleksey.maksi...@it-kb.ru> wrote:
> What am I supposed to do for successfully deploy ovirt 4 ?
> Any ideas ?

Can you please try to explicitly configure your DNS with nameserver
under /etc/resolv.conf and remove DNS1 and DNS2 and set PEERDNS=no for
the interface you are going to use?

> 25.07.2016, 12:47, "aleksey.maksi...@it-kb.ru" <aleksey.maksi...@it-kb.ru>:
>> "Aleksey, was your DNS configured with DNS1 and DNS2 just on the interface 
>> you used to create the management bridge on?"
>>
>> Yes. Of course
>>
>> 25.07.2016, 12:27, "Simone Tiraboschi" <stira...@redhat.com>:
>>>  On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski
>>>  <piotr.kliczew...@gmail.com> wrote:
>>>>   This could be the issue here as well as for BZ #1358530
>>>>
>>>>   On Mon, Jul 25, 2016 at 10:53 AM, <aleksey.maksi...@it-kb.ru> wrote:
>>>>>   Could this be due to the fact that the ovirt installer has changed 
>>>>> network configuration files (ifcfg-*, resolv.conf) ?
>>>>>   After the error in ovirt installation process I see from resolv.conf 
>>>>> disappeared on my DNS servers entry and now the server is unable to 
>>>>> resolve names.
>>>
>>>  So it could be related to 
>>> https://bugzilla.redhat.com/show_bug.cgi?id=1160423
>>>
>>>  Aleksey, was your DNS configured with DNS1 and DNS2 just on the
>>>  interface you used to create the management bridge on?
>>>  Can you please try the workaround described here
>>>  https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ?
>>>
>>>>>   25.07.2016, 11:26, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>>   On Mon, Jul 25, 2016 at 10:22 AM, <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>># vdsClient -s 0 getVdsCaps
>>>>>>>
>>>>>>>Traceback (most recent call last):
>>>>>>>  File "/usr/share/vdsm/vdsClient.py", line 2980, in 
>>>>>>>code, message = commands[command][0](commandArgs)
>>>>>>>  File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
>>>>>>>return self.ExecAndExit(self.s.getVdsCapabilities())
>>>>>>>  File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
>>>>>>>return self.__send(self.__name, args)
>>>>>>>  File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
>>>>>>>verbose=self.__verbose
>>>>>>>  File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
>>>>>>>return self.single_request(host, handler, request_body, verbose)
>>>>>>>  File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in 
>>>>>>> single_request
>>>>>>>self.send_content(h, request_body)
>>>>>>>  File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in 
>>>>>>> send_content
>>>>>>>connection.endheaders(request_body)
>>>>>>>  File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders
>>>>>>>self._send_output(message_body)
>>>>>>>  File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output
>>>>>>>self.send(msg)
>>>>>>>  File "/usr/lib64/python2.7/httplib.py", line 797, in send
>>>>>>>self.connect()
>>>>>>>  File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 
>>>>>>> 203, in connect
>>>>>>>sock = socket.create_connection((self.host, self.port), 
>>>>>>> self.timeout)
>>>>>>>  File "/usr/lib64/python2.7/socket.py", line 571, in 
>>>>>>> create_connection
>>>>>>>raise err
>>>>>>>error: [Errno 101] Network is unreachable
>>>>>>
>>>>>>   Yaniv, can you please take also a look to this one?
>>>>>>   it's exactly the opposite of https://bugzilla.redhat.com/1358530
>>>>>>   Here the jsonrpcclient works but not the xmlrpc one.
>>>>>>
>>>>>>>25.07.2016, 11:17, "Simone Tiraboschi&q

Re: [ovirt-users] oVirt 4 Hosted Engine deploy on fc storage - [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network is unreachable

2016-07-25 Thread Simone Tiraboschi
On Mon, Jul 25, 2016 at 1:46 PM,  <aleksey.maksi...@it-kb.ru> wrote:
> Ok.
>
> 1) I stopped and disabled the service NetworkManager
> # systemctl stop NetworkManager
> # systemctl disable NetworkManager
>
> 2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add PEERDNS=no in 
> ifcfg-* file.
>
> 3) Reboot server
>
> 4) Try deploy oVirt HE 4 and I get the same error
>
> [ INFO  ] Creating Volume Group
> [ INFO  ] Creating Storage Domain
> [ INFO  ] Creating Storage Pool
> [ INFO  ] Connecting Storage Pool
> [ INFO  ] Verifying sanlock lockspace initialization
> [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network 
> is unreachable
> [ INFO  ] Stage: Clean up
> [ INFO  ] Generating answer file 
> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf'
> [ INFO  ] Stage: Pre-termination
> [ INFO  ] Stage: Termination
> [ ERROR ] Hosted Engine deployment failed: this system is not reliable, 
> please check the issue, fix and redeploy
>   Log file is located at 
> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725142534-t81kwf.log
>
>
> What ideas further?

Is your host hostname resolvable now?
Can you please check it with:
 ping $(python -c 'import socket; print(socket.gethostname())')

> 25.07.2016, 13:06, "Simone Tiraboschi" <stira...@redhat.com>:
>> On Mon, Jul 25, 2016 at 11:54 AM, <aleksey.maksi...@it-kb.ru> wrote:
>>>  What am I supposed to do for successfully deploy ovirt 4 ?
>>>  Any ideas ?
>>
>> Can you please try to explicitly configure your DNS with nameserver
>> under /etc/resolv.conf and remove DNS1 and DNS2 and set PEERDNS=no for
>> the interface you are going to use?
>>
>>>  25.07.2016, 12:47, "aleksey.maksi...@it-kb.ru" <aleksey.maksi...@it-kb.ru>:
>>>>  "Aleksey, was your DNS configured with DNS1 and DNS2 just on the 
>>>> interface you used to create the management bridge on?"
>>>>
>>>>  Yes. Of course
>>>>
>>>>  25.07.2016, 12:27, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>   On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski
>>>>>   <piotr.kliczew...@gmail.com> wrote:
>>>>>>This could be the issue here as well as for BZ #1358530
>>>>>>
>>>>>>On Mon, Jul 25, 2016 at 10:53 AM, <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>>Could this be due to the fact that the ovirt installer has changed 
>>>>>>> network configuration files (ifcfg-*, resolv.conf) ?
>>>>>>>After the error in ovirt installation process I see from resolv.conf 
>>>>>>> disappeared on my DNS servers entry and now the server is unable to 
>>>>>>> resolve names.
>>>>>
>>>>>   So it could be related to 
>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1160423
>>>>>
>>>>>   Aleksey, was your DNS configured with DNS1 and DNS2 just on the
>>>>>   interface you used to create the management bridge on?
>>>>>   Can you please try the workaround described here
>>>>>   https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ?
>>>>>
>>>>>>>25.07.2016, 11:26, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>>>>On Mon, Jul 25, 2016 at 10:22 AM, <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>>>> # vdsClient -s 0 getVdsCaps
>>>>>>>>>
>>>>>>>>> Traceback (most recent call last):
>>>>>>>>>   File "/usr/share/vdsm/vdsClient.py", line 2980, in 
>>>>>>>>> code, message = commands[command][0](commandArgs)
>>>>>>>>>   File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
>>>>>>>>> return self.ExecAndExit(self.s.getVdsCapabilities())
>>>>>>>>>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
>>>>>>>>> return self.__send(self.__name, args)
>>>>>>>>>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in 
>>>>>>>>> __request
>>>>>>>>> verbose=self.__verbose
>>>>>>>>>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
>>>>>>>>> return self.singl

Re: [ovirt-users] oVirt 4 Hosted Engine deploy on fc storage - [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network is unreachable

2016-07-25 Thread Simone Tiraboschi
On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski
<piotr.kliczew...@gmail.com> wrote:
> This could be the issue here as well as for BZ #1358530
>
> On Mon, Jul 25, 2016 at 10:53 AM,  <aleksey.maksi...@it-kb.ru> wrote:
>> Could this be due to the fact that the ovirt installer has changed network 
>> configuration files (ifcfg-*, resolv.conf) ?
>> After the error in ovirt installation process I see from resolv.conf 
>> disappeared on my DNS servers entry and now the server is unable to resolve 
>> names.

So it could be related to https://bugzilla.redhat.com/show_bug.cgi?id=1160423

Aleksey, was your DNS configured with DNS1 and DNS2 just on the
interface you used to create the management bridge on?
Can you please try the workaround described here
https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ?



>> 25.07.2016, 11:26, "Simone Tiraboschi" <stira...@redhat.com>:
>>> On Mon, Jul 25, 2016 at 10:22 AM, <aleksey.maksi...@it-kb.ru> wrote:
>>>>  # vdsClient -s 0 getVdsCaps
>>>>
>>>>  Traceback (most recent call last):
>>>>File "/usr/share/vdsm/vdsClient.py", line 2980, in 
>>>>  code, message = commands[command][0](commandArgs)
>>>>File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
>>>>  return self.ExecAndExit(self.s.getVdsCapabilities())
>>>>File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
>>>>  return self.__send(self.__name, args)
>>>>File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
>>>>  verbose=self.__verbose
>>>>File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
>>>>  return self.single_request(host, handler, request_body, verbose)
>>>>File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request
>>>>  self.send_content(h, request_body)
>>>>File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content
>>>>  connection.endheaders(request_body)
>>>>File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders
>>>>  self._send_output(message_body)
>>>>File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output
>>>>  self.send(msg)
>>>>File "/usr/lib64/python2.7/httplib.py", line 797, in send
>>>>  self.connect()
>>>>File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in 
>>>> connect
>>>>  sock = socket.create_connection((self.host, self.port), self.timeout)
>>>>File "/usr/lib64/python2.7/socket.py", line 571, in create_connection
>>>>  raise err
>>>>  error: [Errno 101] Network is unreachable
>>>
>>> Yaniv, can you please take also a look to this one?
>>> it's exactly the opposite of https://bugzilla.redhat.com/1358530
>>> Here the jsonrpcclient works but not the xmlrpc one.
>>>
>>>>  25.07.2016, 11:17, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>  On Mon, Jul 25, 2016 at 7:51 AM, <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>   Simone, there is something interesting in the vdsm.log?
>>>>>
>>>>>  For what I saw the issue is not related to the storage but to the 
>>>>> network.
>>>>>  ovirt-hosted-engine-setup uses the jsonrpc client, instead the code
>>>>>  from ovirt-hosted-engine-ha still uses the xmlrpc client somewhere and
>>>>>  this happens also when the setup asks to create the lockspace volume.
>>>>>  It seams that in your case the xmlrpc client could not connect vdsm on
>>>>>  the localhost.
>>>>>  It could be somehow related to:
>>>>>  https://bugzilla.redhat.com/1358530
>>>>>
>>>>>  Can you please try executing
>>>>>   sudo vdsClient -s 0 getVdsCaps
>>>>>  on that host?
>>>>>
>>>>>>   22.07.2016, 19:36, "aleksey.maksi...@it-kb.ru" 
>>>>>> <aleksey.maksi...@it-kb.ru>:
>>>>>>>   Simone, thanks for link.
>>>>>>>   vdsm.log attached
>>>>>>>
>>>>>>>   22.07.2016, 19:28, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>>>>On Fri, Jul 22, 2016 at 5:59 PM, <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>>>> Thank you for you

Re: [ovirt-users] oVirt 4 Hosted Engine deploy on fc storage - [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network is unreachable

2016-07-25 Thread Simone Tiraboschi
On Mon, Jul 25, 2016 at 2:03 PM,  <aleksey.maksi...@it-kb.ru> wrote:
> Yes.
>
> # ping $(python -c 'import socket; print(socket.gethostname())')
>
> PING KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data.
> 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1 ttl=64 
> time=0.030 ms
> 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2 ttl=64 
> time=0.015 ms
> 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3 ttl=64 
> time=0.011 ms
> ^C
> --- KOM-AD01-VM31.holding.com ping statistics ---
> 3 packets transmitted, 3 received, 0% packet loss, time 2001ms
> rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms
>
>
> but...

and the output of
 ss -plutn

> # vdsClient -s 0 getVdsCaps
>
> Traceback (most recent call last):
>   File "/usr/share/vdsm/vdsClient.py", line 2980, in 
> code, message = commands[command][0](commandArgs)
>   File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
> return self.ExecAndExit(self.s.getVdsCapabilities())
>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
> return self.__send(self.__name, args)
>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
> verbose=self.__verbose
>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
> return self.single_request(host, handler, request_body, verbose)
>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request
> self.send_content(h, request_body)
>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content
> connection.endheaders(request_body)
>   File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders
> self._send_output(message_body)
>   File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output
> self.send(msg)
>   File "/usr/lib64/python2.7/httplib.py", line 797, in send
> self.connect()
>   File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in 
> connect
> sock = socket.create_connection((self.host, self.port), self.timeout)
>   File "/usr/lib64/python2.7/socket.py", line 571, in create_connection
> raise err
> error: [Errno 101] Network is unreachable
>
>
>
>
> 25.07.2016, 14:58, "aleksey.maksi...@it-kb.ru" <aleksey.maksi...@it-kb.ru>:
>> Ok.
>>
>> 1) I stopped and disabled the service NetworkManager
>> # systemctl stop NetworkManager
>> # systemctl disable NetworkManager
>>
>> 2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add PEERDNS=no in 
>> ifcfg-* file.
>>
>> 3) Reboot server
>>
>> 4) Try deploy oVirt HE 4 and I get the same error
>>
>> [ INFO ] Creating Volume Group
>> [ INFO ] Creating Storage Domain
>> [ INFO ] Creating Storage Pool
>> [ INFO ] Connecting Storage Pool
>> [ INFO ] Verifying sanlock lockspace initialization
>> [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network 
>> is unreachable
>> [ INFO ] Stage: Clean up
>> [ INFO ] Generating answer file 
>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf'
>> [ INFO ] Stage: Pre-termination
>> [ INFO ] Stage: Termination
>> [ ERROR ] Hosted Engine deployment failed: this system is not reliable, 
>> please check the issue, fix and redeploy
>>   Log file is located at 
>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725142534-t81kwf.log
>>
>> What ideas further?
>>
>> 25.07.2016, 13:06, "Simone Tiraboschi" <stira...@redhat.com>:
>>>  On Mon, Jul 25, 2016 at 11:54 AM, <aleksey.maksi...@it-kb.ru> wrote:
>>>>   What am I supposed to do for successfully deploy ovirt 4 ?
>>>>   Any ideas ?
>>>
>>>  Can you please try to explicitly configure your DNS with nameserver
>>>  under /etc/resolv.conf and remove DNS1 and DNS2 and set PEERDNS=no for
>>>  the interface you are going to use?
>>>
>>>>   25.07.2016, 12:47, "aleksey.maksi...@it-kb.ru" 
>>>> <aleksey.maksi...@it-kb.ru>:
>>>>>   "Aleksey, was your DNS configured with DNS1 and DNS2 just on the 
>>>>> interface you used to create the management bridge on?"
>>>>>
>>>>>   Yes. Of course
>>>>>
>>>>>   25.07.2016, 12:27, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>>On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski
>>>>>><piotr.kliczew...@gmail.com> wrote

Re: [ovirt-users] Failed to connect Host to the Storage Domains hosted_storage.

2016-07-22 Thread Simone Tiraboschi
On Fri, Jul 22, 2016 at 4:48 PM, Ralf Schenk <r...@databay.de> wrote:

> Hello,
>
> I also see from the logs that all your Storage-Domains that work are
> mounted as nfsVersion='V4' but ovirt-nfs.netsec:/ovirt/hosted-engine is
> mounted as nfsVersion='null'.
>

Hi Robert,
unfortunately Ralf is right: I reproduced the issue.

The auto-import procedure for the hosted-engine storage domain ignores the
nfsVersion parameter and so we don't have a value for that in the engine DB.
On hosted-engine hosts, the agent mounts the hosted-engine storage domain
before the engine and so everything is fine since the agent knows that it's
nfsv4.

The issue comes with the hosts of that datacenter not involved in
hosted-engine: in this case the engine simply tries to mount without
the nfsVersion parameter and so, if the NFS server cannot be access over
nfsv3, the mount could fail and the host will be declared as not operation.

I opened a ticket to track it:
https://bugzilla.redhat.com/show_bug.cgi?id=1359265

If you need a quick fix you can:
- fix the configuration of your storage server to allow it to be accessed
also over nfsv3
- edit the configuration of the storage connection in the engine DB on the
engine VM to add the missing parameter. Something like:
 # sudo -u postgresl psql
 \c engine;
 select * from storage_server_connections;
 UPDATE storage_server_connections SET nfs_version = '4' WHERE connection =
'ovirt-nfs.netsec:/ovirt/hosted-engine';
 commit;
 select * from storage_server_connections;


> Bye
>
> Am 22.07.2016 um 16:17 schrieb Simone Tiraboschi:
>
> On Fri, Jul 22, 2016 at 3:47 PM, Robert Story <rst...@tislabs.com> 
> <rst...@tislabs.com> wrote:
>
> Hello,
>
> I'm in the process of upgrading from 3.5.x to 3.6.x. My hosted engine and
> hosts in the primary cluster are all upgraded and appear to be running fine.
>
> I have a second cluster of 2 machines which are just regular hosts, without
> the hosted-engine. Both have been marked non-operational, with the
> following messages logged about every 5 minutes:
>
>
> Failed to connect Host perses to Storage Pool Default
>
> Host perses cannot access the Storage Domain(s) hosted_storage attached to 
> the Data Center Default. Setting Host state to Non-Operational.
>
> Host perses reports about one of the Active Storage Domains as Problematic.
>
> Failed to connect Host perses to Storage Servers
>
> Failed to connect Host perses to the Storage Domains hosted_storage.
>
>
> I could see the normal storage/iso/export domains mounted on the host, and
> the VMs running on the host are fine.
>
> In 3.5 only the hosts involved in hosted-engine have to access the
> hosted-engine storage domain.
> With 3.6 we introduced the capabilities to manage the engine VM from
> the engine itself so the engine has to import in the hosted-engine
> storage domain.
> This means that all the hosts in the datacenter that contains the
> cluster with the hosted-engine hosts have now to be able to connect
> the hosted-engine storage domain.
>
> Can you please check the ACL on the storage server (NFS or iSCSI) that
> you use to expose the hosted-engine storage domain?
>
>
> I shut down the VMs on one host, put it in maintenance mode, installed 3.6
> repo and ran yum update. All went well, but when I activated the host, same
> deal.
>
> I've attached the engine log snippet for the activation attempt.
>
> Robert
>
> --
> Senior Software Engineer @ Parsons
>
> ___
> Users mailing listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>
> ___
> Users mailing listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>
>
> --
>
>
> *Ralf Schenk*
> fon +49 (0) 24 05 / 40 83 70
> fax +49 (0) 24 05 / 40 83 759
> mail *r...@databay.de* <r...@databay.de>
>
> *Databay AG*
> Jens-Otto-Krag-Straße 11
> D-52146 Würselen
> *www.databay.de* <http://www.databay.de>
>
> Sitz/Amtsgericht Aachen • HRB:8437 • USt-IdNr.: DE 210844202
> Vorstand: Ralf Schenk, Dipl.-Ing. Jens Conze, Aresch Yavari, Dipl.-Kfm.
> Philipp Hermanns
> Aufsichtsratsvorsitzender: Wilhelm Dohmen
> --
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4 Hosted Engine deploy on fc storage - [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network is unreachable

2016-07-25 Thread Simone Tiraboschi
On Mon, Jul 25, 2016 at 10:22 AM,  <aleksey.maksi...@it-kb.ru> wrote:
> # vdsClient -s 0 getVdsCaps
>
> Traceback (most recent call last):
>   File "/usr/share/vdsm/vdsClient.py", line 2980, in 
> code, message = commands[command][0](commandArgs)
>   File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
> return self.ExecAndExit(self.s.getVdsCapabilities())
>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
> return self.__send(self.__name, args)
>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
> verbose=self.__verbose
>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
> return self.single_request(host, handler, request_body, verbose)
>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request
> self.send_content(h, request_body)
>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content
> connection.endheaders(request_body)
>   File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders
> self._send_output(message_body)
>   File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output
> self.send(msg)
>   File "/usr/lib64/python2.7/httplib.py", line 797, in send
> self.connect()
>   File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in 
> connect
> sock = socket.create_connection((self.host, self.port), self.timeout)
>   File "/usr/lib64/python2.7/socket.py", line 571, in create_connection
> raise err
> error: [Errno 101] Network is unreachable

Yaniv, can you please take also a look to this one?
it's exactly the opposite of https://bugzilla.redhat.com/1358530
Here the jsonrpcclient works but not the xmlrpc one.

> 25.07.2016, 11:17, "Simone Tiraboschi" <stira...@redhat.com>:
>> On Mon, Jul 25, 2016 at 7:51 AM, <aleksey.maksi...@it-kb.ru> wrote:
>>>  Simone, there is something interesting in the vdsm.log?
>>
>> For what I saw the issue is not related to the storage but to the network.
>> ovirt-hosted-engine-setup uses the jsonrpc client, instead the code
>> from ovirt-hosted-engine-ha still uses the xmlrpc client somewhere and
>> this happens also when the setup asks to create the lockspace volume.
>> It seams that in your case the xmlrpc client could not connect vdsm on
>> the localhost.
>> It could be somehow related to:
>> https://bugzilla.redhat.com/1358530
>>
>> Can you please try executing
>>  sudo vdsClient -s 0 getVdsCaps
>> on that host?
>>
>>>  22.07.2016, 19:36, "aleksey.maksi...@it-kb.ru" <aleksey.maksi...@it-kb.ru>:
>>>>  Simone, thanks for link.
>>>>  vdsm.log attached
>>>>
>>>>  22.07.2016, 19:28, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>   On Fri, Jul 22, 2016 at 5:59 PM, <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>Thank you for your response, Simone.
>>>>>>
>>>>>>Log attached.
>>>>>
>>>>>   It seams it comes from VDSM, can you please attach also vdsm.log?
>>>>>
>>>>>>I don't use ovirt-engine-appliance because I have not found "how-to" 
>>>>>> for ovirt-engine-appliance deployment in hosted engine configuration.
>>>>>
>>>>>   yum install ovirt-engine-appliance
>>>>>
>>>>>   Then follow the instruction here:
>>>>>   
>>>>> http://www.ovirt.org/develop/release-management/features/heapplianceflow/
>>>>>
>>>>>>22.07.2016, 17:09, "Simone Tiraboschi" <stira...@redhat.com>:
>>>>>>>Hi Aleksey,
>>>>>>>Can you please attach hosted-engine-setup logs?
>>>>>>>
>>>>>>>On Fri, Jul 22, 2016 at 3:46 PM, <aleksey.maksi...@it-kb.ru> wrote:
>>>>>>>
>>>>>>>> Hello oVirt guru`s !
>>>>>>>>
>>>>>>>> I have problem with initial deploy of ovirt 4.0 hosted engine.
>>>>>>>>
>>>>>>>> My environment :
>>>>>>>> 
>>>>>>>> * Two servers HP ProLiant DL 360 G5 with Qlogic FC HBA connected 
>>>>>>>> (with
>>>>>>>> multipathd) to storage HP 3PAR 7200
>>>>>>>> * On each server installed CentOS 7.2 Linux 
&g

Re: [ovirt-users] hosted-engine-deploy error

2016-07-25 Thread Simone Tiraboschi
On Mon, Jul 25, 2016 at 5:23 AM, qinglong.d...@horebdata.cn
 wrote:
> Hi, all:
> I tried to deploy an hosted engine, but I got an error:
> [root@node ~]# hosted-engine --deploy
> ..
>   --== CONFIGURATION PREVIEW ==--
>
>   Bridge interface   : enp6s0f0
>   Engine FQDN: engine.horebdata.cn
>   Bridge name: ovirtmgmt
>   Host address   : node.horebdata.cn
>   SSH daemon port: 22
>   Firewall manager   : iptables
>   Gateway address: 192.168.1.1
>   Host name for web application  : node
>   Storage Domain type: iscsi
>   Host ID: 1
>   LUN ID : 23866376332323932
>   Image size GB  : 25
>   iSCSI Portal IP Address: 192.168.128.115
>   iSCSI Target Name  : engine
>   iSCSI Portal port  : 3260
>   iSCSI Portal user  :
>   Console type   : vnc
>   Memory size MB : 4096
>   MAC address: 00:16:3e:48:db:9e
>   Boot type  : cdrom
>   Number of CPUs : 2
>   ISO image (cdrom boot/cloud-init)  : /tmp/1.iso
>   CPU Type   : model_SandyBridge
>
>   Please confirm installation settings (Yes, No)[Yes]:
> [ INFO  ] Stage: Transaction setup
> [ INFO  ] Stage: Misc configuration
> [ INFO  ] Stage: Package installation
> [ INFO  ] Stage: Misc configuration
> [ INFO  ] Configuring libvirt
> [ INFO  ] Configuring VDSM
> [ INFO  ] Starting vdsmd
> [ INFO  ] Configuring the management bridge
> [ INFO  ] Creating Volume Group
> [ ERROR ] Error creating Volume Group: Failed to initialize physical device:
> ("[u'/dev/mapper/23866376332323932']",)
>   The selected device is already used.
>   To create a vg on this device, you must use Force.
>   WARNING: This will destroy existing data on the device.
>   (Force, Abort)[Abort]? Force
> [ INFO  ] Creating Storage Domain
> [ INFO  ] Creating Storage Pool
> [ INFO  ] Connecting Storage Pool
> [ INFO  ] Verifying sanlock lockspace initialization
> [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network
> is unreachable
> [ INFO  ] Stage: Clean up
> [ INFO  ] Generating answer file
> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725101205.conf'
> [ INFO  ] Stage: Pre-termination
> [ INFO  ] Stage: Termination
> [ ERROR ] Hosted Engine deployment failed: this system is not reliable,
> please check the issue, fix and redeploy
>   Log file is located at
> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725100535-h2y78c.log
> [root@node ~]# cat
> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725100535-h2y78c.log
> | grep ERROR
> 2016-07-25 10:07:01 ERROR otopi.plugins.gr_he_setup.storage.blockd
> blockd._misc:664 Error creating Volume Group: Failed to initialize physical
> device: ("[u'/dev/mapper/23866376332323932']",)
> 2016-07-25 10:11:59 ERROR otopi.context context._executeMethod:151 Failed to
> execute stage 'Misc configuration': [Errno 101] Network is unreachable
> 2016-07-25 10:12:05 ERROR otopi.plugins.gr_he_setup.core.misc
> misc._terminate:99 Hosted Engine deployment failed: this system is not
> reliable, please check the issue, fix and redeploy
> [root@node ~]# cat /var/log/vdsm/vdsm.log | grep ERROR
> jsonrpc.Executor/4::ERROR::2016-07-25
> 10:07:01,491::lvm::737::Storage.LVM::(_initpvs) pvcreate failed with rc=5
> jsonrpc.Executor/4::ERROR::2016-07-25
> 10:07:01,492::lvm::738::Storage.LVM::(_initpvs) [], ['  WARNING: lvmetad is
> running but disabled. Restart lvmetad before enabling it!', '
> /dev/mapper/23866376332323932: read failed after 0 of 4096 at 0:
> Input/output error', '  /dev/mapper/23866376332323932: read failed after 0
> of 4096 at 42949607424: Input/output error', '
> /dev/mapper/23866376332323932: read failed after 0 of 4096 at 42949664768:
> Input/output error', '  WARNING: Error counts reached a limit of 3. Device
> /dev/mapper/23866376332323932 was disabled', '  Failed to wipe new metadata
> area at the start of the /dev/mapper/23866376332323932', '  Failed to add
> metadata area for new physical volume /dev/mapper/23866376332323932', '
> Failed to setup physical volume "/dev/mapper/23866376332323932"']
> jsonrpc.Executor/4::ERROR::2016-07-25
> 10:07:01,492::task::868::Storage.TaskManager.Task::(_setError)
> Task=`a1b55dc1-2c23-4e9f-b81a-43b189557028`::Unexpected error
> jsonrpc.Executor/4::ERROR::2016-07-25
> 10:07:01,496::dispatcher::77::Storage.Dispatcher::(wrapper) {'status':
> {'message': 'Failed 

Re: [ovirt-users] Ovirt Hosted-Engine not installing ERROR: 'OVEHOSTED_NETWORK/host_name'

2016-07-28 Thread Simone Tiraboschi
On Thu, Jul 28, 2016 at 9:22 AM, Simone Tiraboschi <stira...@redhat.com> wrote:
> On Thu, Jul 28, 2016 at 7:50 AM, Yedidyah Bar David <d...@redhat.com> wrote:
>> On Wed, Jul 27, 2016 at 8:42 PM, Florian Nolden <f.nol...@xilloc.com> wrote:
>>> Hello,
>>>
>>> I try to install Ovirt 4.0.1-1 on a fresh installed CentOS 7.2 using a

Another thing, either the bugged version (2.0.1.2) and the fixed one
(2.0.1.3) are available just in the 4.0.2 Second Release Candidate
repo which has not still reached the GA status.
The latest release is Ovirt 4.0.1 so maybe you are also using the
wrong repo if you want that.

>>> replica 3 glusterfs. But I have trouble deploying the hosted engine.
>>>
>>> hosted-engine --deploy
>>>
>>> /usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/storage_backends.py:15:
>>> DeprecationWarning: vdscli uses xmlrpc. since ovirt 3.6 xmlrpc is
>>> deprecated, please use vdsm.jsonrpcvdscli
>>>   import vdsm.vdscli
>>>
>>> [ ERROR ] Failed to execute stage 'Environment customization':
>>> 'OVEHOSTED_NETWORK/host_name'
>
> The issue was caused by this patch
>  https://gerrit.ovirt.org/#/c/61078/
> yesterday we reverted it and built a new version (2.0.1.3) of
> hosted-engine-setup without that.
> It's already available:
> http://resources.ovirt.org/pub/ovirt-4.0-pre/rpm/el7/noarch/ovirt-hosted-engine-setup-2.0.1.3-1.el7.centos.noarch.rpm
>
>>> VDSM also did not create the ovirtmgmt bridge or the routing tables.
>>>
>>> I used the CentOS 7 minimal, and selected Infrastructure Server. I added the
>>> Puppet 4 repo and the Ovirt 4.0 Repo, no EPEL.
>>> I can reproduce it on 3 similar installed servers.
>>>
>>> Any Ideas?
>>
>> Please share the setup log. Thanks.
>>
>> Best,
>> --
>> Didi
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Safe to upgrade HE hosts from GUI?

2016-07-28 Thread Simone Tiraboschi
On Thu, Jul 28, 2016 at 10:41 AM, Wee Sritippho <we...@forest.go.th> wrote:

> On 21/7/2559 16:53, Simone Tiraboschi wrote:
>
> On Thu, Jul 21, 2016 at 11:43 AM, Wee Sritippho <we...@forest.go.th>
> wrote:
>
>
>> Can I just follow
>> http://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine
>> until step 3 and do everything else via GUI?
>>
> Yes, absolutely.
>
>
> Hi, I upgrade a host (host02) via GUI and now its score is 0. Restarted
> the services but the result is still the same. Kinda lost now. What should
> I do next?
>
>
Can you please attach ovirt-ha-agent logs?


> [root@host02 ~]# service vdsmd restart
> Redirecting to /bin/systemctl restart  vdsmd.service
> [root@host02 ~]# systemctl restart ovirt-ha-broker && systemctl restart
> ovirt-ha-agent
> [root@host02 ~]# systemctl status ovirt-ha-broker
> ● ovirt-ha-broker.service - oVirt Hosted Engine High Availability
> Communications Broker
>Loaded: loaded (/usr/lib/systemd/system/ovirt-ha-broker.service;
> enabled; vendor preset: disabled)
>Active: active (running) since Thu 2016-07-28 15:09:38 ICT; 20min ago
>  Main PID: 4614 (ovirt-ha-broker)
>CGroup: /system.slice/ovirt-ha-broker.service
>└─4614 /usr/bin/python
> /usr/share/ovirt-hosted-engine-ha/ovirt-ha-broker --no-daemon
>
> Jul 28 15:29:35 host02.ovirt.forest.go.th ovirt-ha-broker[4614]:
> INFO:ovirt_hosted_engine_ha.broker.listener.ConnectionHandler:Connection
> established
> Jul 28 15:29:35 host02.ovirt.forest.go.th ovirt-ha-broker[4614]:
> INFO:ovirt_hosted_engine_ha.broker.listener.ConnectionHandler:Connection
> closed
> Jul 28 15:29:35 host02.ovirt.forest.go.th ovirt-ha-broker[4614]:
> INFO:ovirt_hosted_engine_ha.broker.listener.ConnectionHandler:Connection
> established
> Jul 28 15:29:35 host02.ovirt.forest.go.th ovirt-ha-broker[4614]:
> INFO:ovirt_hosted_engine_ha.broker.listener.ConnectionHandler:Connection
> closed
> Jul 28 15:29:48 host02.ovirt.forest.go.th ovirt-ha-broker[4614]:
> INFO:ovirt_hosted_engine_ha.broker.listener.ConnectionHandler:Connection
> established
> Jul 28 15:29:48 host02.ovirt.forest.go.th ovirt-ha-broker[4614]:
> INFO:ovirt_hosted_engine_ha.broker.listener.ConnectionHandler:Connection
> closed
> Jul 28 15:29:48 host02.ovirt.forest.go.th ovirt-ha-broker[4614]:
> INFO:ovirt_hosted_engine_ha.broker.listener.ConnectionHandler:Connection
> established
> Jul 28 15:29:48 host02.ovirt.forest.go.th ovirt-ha-broker[4614]:
> INFO:ovirt_hosted_engine_ha.broker.listener.ConnectionHandler:Connection
> closed
> Jul 28 15:29:48 host02.ovirt.forest.go.th ovirt-ha-broker[4614]:
> INFO:ovirt_hosted_engine_ha.broker.listener.ConnectionHandler:Connection
> established
> Jul 28 15:29:48 host02.ovirt.forest.go.th ovirt-ha-broker[4614]:
> INFO:ovirt_hosted_engine_ha.broker.listener.ConnectionHandler:Connection
> closed
> [root@host02 ~]# systemctl status ovirt-ha-agent
> ● ovirt-ha-agent.service - oVirt Hosted Engine High Availability
> Monitoring Agent
>Loaded: loaded (/usr/lib/systemd/system/ovirt-ha-agent.service;
> enabled; vendor preset: disabled)
>Active: active (running) since Thu 2016-07-28 15:28:34 ICT; 1min 19s ago
>  Main PID: 11488 (ovirt-ha-agent)
>CGroup: /system.slice/ovirt-ha-agent.service
>└─11488 /usr/bin/python
> /usr/share/ovirt-hosted-engine-ha/ovirt-ha-agent --no-daemon
>
> Jul 28 15:29:52 host02.ovirt.forest.go.th ovirt-ha-agent[11488]:
> /usr/lib/python2.7/site-packages/yajsonrpc/stomp.py:352:
> DeprecationWarning: Dispatcher.pend...instead.
> Jul 28 15:29:52 host02.ovirt.forest.go.th ovirt-ha-agent[11488]: pending
> = getattr(dispatcher, 'pending', lambda: 0)
> Jul 28 15:29:53 host02.ovirt.forest.go.th ovirt-ha-agent[11488]:
> /usr/lib/python2.7/site-packages/yajsonrpc/stomp.py:352:
> DeprecationWarning: Dispatcher.pend...instead.
> Jul 28 15:29:53 host02.ovirt.forest.go.th ovirt-ha-agent[11488]: pending
> = getattr(dispatcher, 'pending', lambda: 0)
> Jul 28 15:29:53 host02.ovirt.forest.go.th ovirt-ha-agent[11488]:
> /usr/lib/python2.7/site-packages/yajsonrpc/stomp.py:352:
> DeprecationWarning: Dispatcher.pend...instead.
> Jul 28 15:29:53 host02.ovirt.forest.go.th ovirt-ha-agent[11488]: pending
> = getattr(dispatcher, 'pending', lambda: 0)
> Jul 28 15:29:53 host02.ovirt.forest.go.th ovirt-ha-agent[11488]:
> /usr/lib/python2.7/site-packages/yajsonrpc/stomp.py:352:
> DeprecationWarning: Dispatcher.pend...instead.
> Jul 28 15:29:53 host02.ovirt.forest.go.th ovirt-ha-agent[11488]: pending
> = getattr(dispatcher, 'pending', lambda: 0)
> Jul 28 15:29:53 host02.ovirt.forest.go.th ovirt-ha-agent[11488]:
> ovirt-ha-agent ovirt_hosted_engine_ha.agent.agent.Agent ERROR Error:
> 'Attemp

Re: [ovirt-users] 3.6 : Moving the hosted-engine to another storage

2016-07-28 Thread Simone Tiraboschi
On Thu, Jul 28, 2016 at 10:59 AM, Alexis HAUSER
 wrote:
>>Unfortunately we know that migrating from HE to HE is not as simple as
>>from physical to HE:
>>https://bugzilla.redhat.com/show_bug.cgi?id=1240466#c21
>>In general the issue is that the DB backup form the old hosted-engine
>>VM contains a lot of references to the previous hosted-engine env and
>>you cannot simply remove/edit them from the engine since they are lock
>>so you have to manually remove them from the DB which is quite
>>risky/error prone.
>
> This is a bit scaring. In case of issue with engine and trying to recover, it 
> could also happen.

The backup and restore procedure on the same env is well testes, the
issue is simply if you need to restore on a different environment
(it's basically a migration) cause you have to remove all the
references to the old env.

> What other way would you suggest for backing the engine VM and being sure to 
> be able to restore it as it was without errors ? Have you ever tried to 
> backup/restore from rsync ?
> If there are data in the DB written when you're performing it, do you think 
> it can cause issues ? If ovirt-engine service is stopped, is that problem 
> avoided ?

Postgres is transactional so if you do it in the proper way I don't
see issue but stopping the ovirt-engine service will of course help.
But the issue is not that you risk data corruption, the issue is that
when you import a backup of the engine DB to a different env, that
backup says the hosted-engine storage domain is still the old one
since it was that in the previous env and so on and you can simply
edit the hosted-engine strage domain location from the engine itself
since we are preventing it.

>>In the mean time I'd suggest, if feasible, to redeploy a new
>>hosted-engine env and reattach there your storage domains and your
>>hosts.
>>This will imply a downtime.
>
> Ok, I think I'll do that. A downtime isn't a problem right now, as I'm still 
> at a pre-production step. (preparing it for production soon)
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Ovirt Hosted-Engine not installing ERROR: 'OVEHOSTED_NETWORK/host_name'

2016-07-28 Thread Simone Tiraboschi
On Thu, Jul 28, 2016 at 7:50 AM, Yedidyah Bar David  wrote:
> On Wed, Jul 27, 2016 at 8:42 PM, Florian Nolden  wrote:
>> Hello,
>>
>> I try to install Ovirt 4.0.1-1 on a fresh installed CentOS 7.2 using a
>> replica 3 glusterfs. But I have trouble deploying the hosted engine.
>>
>> hosted-engine --deploy
>>
>> /usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/storage_backends.py:15:
>> DeprecationWarning: vdscli uses xmlrpc. since ovirt 3.6 xmlrpc is
>> deprecated, please use vdsm.jsonrpcvdscli
>>   import vdsm.vdscli
>>
>> [ ERROR ] Failed to execute stage 'Environment customization':
>> 'OVEHOSTED_NETWORK/host_name'

The issue was caused by this patch
 https://gerrit.ovirt.org/#/c/61078/
yesterday we reverted it and built a new version (2.0.1.3) of
hosted-engine-setup without that.
It's already available:
http://resources.ovirt.org/pub/ovirt-4.0-pre/rpm/el7/noarch/ovirt-hosted-engine-setup-2.0.1.3-1.el7.centos.noarch.rpm

>> VDSM also did not create the ovirtmgmt bridge or the routing tables.
>>
>> I used the CentOS 7 minimal, and selected Infrastructure Server. I added the
>> Puppet 4 repo and the Ovirt 4.0 Repo, no EPEL.
>> I can reproduce it on 3 similar installed servers.
>>
>> Any Ideas?
>
> Please share the setup log. Thanks.
>
> Best,
> --
> Didi
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] 3.6 : Moving the hosted-engine to another storage

2016-07-27 Thread Simone Tiraboschi
On Wed, Jul 27, 2016 at 5:38 PM, Nir Soffer  wrote:
> On Wed, Jul 27, 2016 at 3:38 PM, Alexis HAUSER
>  wrote:
>> Hi,
>>
>> I'm currently using an NFS storage for my hosted-engine. However, this NFS 
>> server will be removed soon. I'd like to move the hosted-engine to an iSCSI 
>> storage. How can I proceed ? The options for moving/copying VM disk don't 
>> seem to be available for the hosted engine in the web interface.
>
> Yes, hosted engine has some limitations at this point, mainly all the
> interesting
> features like live storage migration, live snapshots, hot plug/unplug disk are
> not available.
>
> Simone, do we have a simple procedure to move the engine to another storage?

Theoretically you could run engine-backup on the current engine VM,
redeploy from scratch on the new storage domain and before running
engine setup, inject there the backup.
Something like:
http://www.ovirt.org/develop/developer-guide/engine/migrate-to-hosted-engine/

Unfortunately we know that migrating from HE to HE is not as simple as
from physical to HE:
https://bugzilla.redhat.com/show_bug.cgi?id=1240466#c21
In general the issue is that the DB backup form the old hosted-engine
VM contains a lot of references to the previous hosted-engine env and
you cannot simply remove/edit them from the engine since they are lock
so you have to manually remove them from the DB which is quite
risky/error prone.

Improving engine-backup to automatically filter hosted-engine
references is an RFE for 4.1.

In the mean time I'd suggest, if feasible, to redeploy a new
hosted-engine env and reattach there your storage domains and your
hosts.
This will imply a downtime.

> Nir
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Ovirt Hosted-Engine not installing ERROR: 'OVEHOSTED_NETWORK/host_name'

2016-07-28 Thread Simone Tiraboschi
On Thu, Jul 28, 2016 at 1:07 PM, Florian Nolden <f.nol...@xilloc.com> wrote:
> Im using the the Ovirt 4.0 Release repo:
>
> http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm
>
> But the ovirt-4.0-dependencies.repo contains:
>
> [centos-ovirt40-candidate]
> name=CentOS-7 - oVirt 4.0
> baseurl=http://cbs.centos.org/repos/virt7-ovirt-40-candidate/x86_64/os/
> gpgcheck=0
> enabled=1
>
> I believe that shouldn't be there or?

Yes, you are absolutely right.
Thanks for reporting it.

> 2016-07-28 10:17 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>
>> On Thu, Jul 28, 2016 at 9:22 AM, Simone Tiraboschi <stira...@redhat.com>
>> wrote:
>> > On Thu, Jul 28, 2016 at 7:50 AM, Yedidyah Bar David <d...@redhat.com>
>> > wrote:
>> >> On Wed, Jul 27, 2016 at 8:42 PM, Florian Nolden <f.nol...@xilloc.com>
>> >> wrote:
>> >>> Hello,
>> >>>
>> >>> I try to install Ovirt 4.0.1-1 on a fresh installed CentOS 7.2 using a
>>
>> Another thing, either the bugged version (2.0.1.2) and the fixed one
>> (2.0.1.3) are available just in the 4.0.2 Second Release Candidate
>> repo which has not still reached the GA status.
>> The latest release is Ovirt 4.0.1 so maybe you are also using the
>> wrong repo if you want that.
>>
>> >>> replica 3 glusterfs. But I have trouble deploying the hosted engine.
>> >>>
>> >>> hosted-engine --deploy
>> >>>
>> >>>
>> >>> /usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/storage_backends.py:15:
>> >>> DeprecationWarning: vdscli uses xmlrpc. since ovirt 3.6 xmlrpc is
>> >>> deprecated, please use vdsm.jsonrpcvdscli
>> >>>   import vdsm.vdscli
>> >>>
>> >>> [ ERROR ] Failed to execute stage 'Environment customization':
>> >>> 'OVEHOSTED_NETWORK/host_name'
>> >
>> > The issue was caused by this patch
>> >  https://gerrit.ovirt.org/#/c/61078/
>> > yesterday we reverted it and built a new version (2.0.1.3) of
>> > hosted-engine-setup without that.
>> > It's already available:
>> >
>> > http://resources.ovirt.org/pub/ovirt-4.0-pre/rpm/el7/noarch/ovirt-hosted-engine-setup-2.0.1.3-1.el7.centos.noarch.rpm
>> >
>> >>> VDSM also did not create the ovirtmgmt bridge or the routing tables.
>> >>>
>> >>> I used the CentOS 7 minimal, and selected Infrastructure Server. I
>> >>> added the
>> >>> Puppet 4 repo and the Ovirt 4.0 Repo, no EPEL.
>> >>> I can reproduce it on 3 similar installed servers.
>> >>>
>> >>> Any Ideas?
>> >>
>> >> Please share the setup log. Thanks.
>> >>
>> >> Best,
>> >> --
>> >> Didi
>> >> ___
>> >> Users mailing list
>> >> Users@ovirt.org
>> >> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Solved: Re: 3.5 to 3.6 upgrade stuck

2016-07-22 Thread Simone Tiraboschi
On Fri, Jul 22, 2016 at 9:58 AM, Simone Tiraboschi <stira...@redhat.com> wrote:
> On Fri, Jul 22, 2016 at 4:11 AM, Robert Story <rst...@tislabs.com> wrote:
>> On Thu, 21 Jul 2016 16:04:41 -0400 Robert wrote:
>> RS> 
>> Thread-1::config::278::ovirt_hosted_engine_ha.broker.notifications.Notifications.config
>> RS>  ::(refresh_local_conf_file) local conf file was correctly written
>> RS>
>> RS> And then  nothing. It just hangs. Nothing more is logged Thread-1.
>>
>> So I started digging around the the python source, starting from
>> refresh_local_conf_file. I ended up in ./broker/notifications.py, in
>> send_email. I added some logging:
>>
>> def send_email(cfg, email_body):
>> """Send email."""
>>
>> logger = logging.getLogger("%s.Notifications" % __name__)
>>
>> try:
>> logger.debug(" setting up smtp 1")
>> server = smtplib.SMTP(cfg["smtp-server"], port=cfg["smtp-port"])
>> logger.debug(" setting up smtp 2")
>> ...
>>
>> Now the final messages are:
>>
>> Thread-1::DEBUG::2016-07-21 21:35:05,280::config::278::
>>   ovirt_hosted_engine_ha.broker.notifications.Notifications.config::
>>   (refresh_local_conf_file) local conf file was correctly written
>> Thread-1::DEBUG::2016-07-21 21:35:05,282::notifications::27::
>>   ovirt_hosted_engine_ha.broker.notifications.Notifications::
>>   (send_email)  setting up smtp 1
>>
>>
>> So the culprit is:
>>
>> server = smtplib.SMTP(cfg["smtp-server"], port=cfg["smtp-port"])
>>
>> Note that this does actually send the email - 2 minutes later.
>
> Thanks for time and your effort Robert!
> In general the agent shouldn't got stuck if the broker is not able to
> send a notification email within a certain amount of time.
> I'm open a bug to track this. Adding Martin here.

https://bugzilla.redhat.com/1359059

>> So I tried:
>>
>>   $ telnet localhost 25
>>   Trying ::1...
>>
>> which hung, and a little bell went off in my brain...
>>
>> After changing /etc/hosts from:
>>
>> 127.0.0.1   localhost localhost.localdomain localhost4 
>> localhost4.localdomain4
>> ::1 localhost localhost.localdomain localhost6 
>> localhost6.localdomain6
>>
>> to
>>
>> 127.0.0.1   localhost localhost.localdomain localhost4 
>> localhost4.localdomain4
>> ::1 localhost6 localhost6.localdomain6
>>
>> localhost resolves to 127.0.0.1, the delay is gone, and everything is fine.
>
> We are seeing similar reports regarding ip4/ip6 issues also migrating on 4.0
> See also http://lists.ovirt.org/pipermail/users/2016-June/040578.html and
> https://bugzilla.redhat.com/show_bug.cgi?id=1358530
>
> Adding Oved here.
>
>> I don't want to update /etc/hosts on each host. Is there somewhere I can
>> edit the broker config for mail?
>
> The shortest option is to edit broker.conf inside the configuration
> volume on the hosted-engine storage domain but it's a bit tricky and
> also potentially dangerous if not well done.
> We have an RFE about letting you reconfigure it from the engine, for
> now, if you are brave enough, please try something like this.
>
> dir=`mktemp -d` && cd $dir
> mnt_point=/rhev/data-center/mnt/192.168.1.115:_Virtual_ext35u36 #
> pleace with your local mount point
> systemctl stop ovirt-ha-broker # on all the hosts!
> sdUUID_line=$(grep sdUUID /etc/ovirt-hosted-engine/hosted-engine.conf)
> sdUUID=${sdUUID_line:7:36}
> conf_volume_UUID_line=$(grep conf_volume_UUID
> /etc/ovirt-hosted-engine/hosted-engine.conf)
> conf_volume_UUID=${conf_volume_UUID_line:17:36}
> conf_image_UUID_line=$(grep conf_image_UUID
> /etc/ovirt-hosted-engine/hosted-engine.conf)
> conf_image_UUID=${conf_image_UUID_line:16:36}
> sudo -u vdsm dd
> if=$mnt_point/$sdUUID/images/$conf_image_UUID/$conf_volume_UUID
> 2>/dev/null| tar -xvf -
> # here you have to edit the locally extracted broker.conf
> tar -cO * | sudo -u vdsm dd
> of=$mnt_point/$sdUUID/images/$conf_image_UUID/$conf_volume_UUID
> systemctl restart ovirt-ha-agent # on all the hosts
>
> I strongly advice to take a backup before editing.
>
>> Robert
>>
>> --
>> Senior Software Engineer @ Parsons
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Safe to upgrade HE hosts from GUI?

2016-07-29 Thread Simone Tiraboschi
On Fri, Jul 29, 2016 at 11:35 AM, Wee Sritippho <we...@forest.go.th> wrote:
> On 29/7/2559 15:50, Simone Tiraboschi wrote:
>>
>> On Fri, Jul 29, 2016 at 6:31 AM, Wee Sritippho <we...@forest.go.th> wrote:
>>>
>>> On 28/7/2559 15:54, Simone Tiraboschi wrote:
>>>
>>> On Thu, Jul 28, 2016 at 10:41 AM, Wee Sritippho <we...@forest.go.th>
>>> wrote:
>>>>
>>>> On 21/7/2559 16:53, Simone Tiraboschi wrote:
>>>>
>>>> On Thu, Jul 21, 2016 at 11:43 AM, Wee Sritippho <we...@forest.go.th>
>>>> wrote:
>>>>
>>>>> Can I just follow
>>>>>
>>>>> http://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine
>>>>> until step 3 and do everything else via GUI?
>>>>
>>>> Yes, absolutely.
>>>>
>>>>
>>>> Hi, I upgrade a host (host02) via GUI and now its score is 0. Restarted
>>>> the services but the result is still the same. Kinda lost now. What
>>>> should I
>>>> do next?
>>>>
>>> Can you please attach ovirt-ha-agent logs?
>>>
>>>
>>> Yes, here are the logs:
>>> https://app.box.com/s/b4urjty8dsuj98n3ywygpk3oh5o7pbsh
>>
>> Thanks Wee,
>> your issue is here:
>> MainThread::ERROR::2016-07-17
>>
>> 14:32:45,586::storage_server::143::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(_validate_pre_connected_path)
>> The hosted-engine storage domain is already mounted on
>>
>> '/rhev/data-center/mnt/glusterSD/host02.ovirt.forest.go.th:_hosted__engine/639e689c-8493-479b-a6eb-cc92b6fc4cf4'
>> with a path that is not supported anymore: the right path should be
>>
>> '/rhev/data-center/mnt/glusterSD/host01.ovirt.forest.go.th:_hosted__engine/639e689c-8493-479b-a6eb-cc92b6fc4cf4'.
>>
>> Did you manually tried to avoid the issue of a single entry point for
>> the gluster FS volume using host01.ovirt.forest.go.th:_hosted__engine
>> and host02.ovirt.forest.go.th:_hosted__engine there?
>> This could cause a lot of confusion since the code could not detect
>> that the storage domain is the same and you can end with it mounted
>> twice into different locations and a lot of issues.
>> The correct solution of that issue was this one:
>> https://bugzilla.redhat.com/show_bug.cgi?id=1298693#c20
>>
>> Now, to have it fixed on your env you have to hack a bit.
>> First step, you have to edit
>> /etc/ovirt-hosted-engine/hosted-engine.conf on all your hosted-engine
>> hosts to ensure that the storage field always point to the same entry
>> point (host01 for instance)
>> Then on each host you can add something like:
>>
>> mnt_options=backupvolfile-server=host02.ovirt.forest.go.th:host03.ovirt.forest.go.th,fetch-attempts=2,log-level=WARNING,log-file=/var/log/engine_domain.log
>>
>> Then check the representation of your storage connection in the table
>> storage_server_connections of the engine DB and make sure that
>> connection refers to the entry point you used in hosted-engine.conf on
>> all your hosts, you have lastly to set the value of mount_options also
>> here.
>
> Weird. The configuration in all hosts are already referring to host01.

but for sure you have a connection pointing to host02 somewhere, did
you try to manually deploy from CLI connecting the gluster volume on
host02?

> Also, in the storage_server_connections table:
>
> engine=> SELECT * FROM storage_server_connections;
>   id  | connection|
> user_name | password | iqn | port | portal | storage_type | mount_options |
> vfs_type
>  | nfs_version | nfs_timeo | nfs_retrans
> --+--+---+--+-+--++--+---+--
> -+-+---+-
>  bd78d299-c8ff-4251-8aab-432ce6443ae8 |
> host01.ovirt.forest.go.th:/hosted_engine |   | | |  | 1
> |7 |   | glusterfs
>  | |   |
> (1 row)
>
>
>>
>> Please tune also the value of network.ping-timeout for your glusterFS
>> volume to avoid this:
>>   https://bugzilla.redhat.com/show_bug.cgi?id=1319657#c17
>
>
> --
> Wee
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Safe to upgrade HE hosts from GUI?

2016-07-29 Thread Simone Tiraboschi
On Fri, Jul 29, 2016 at 6:31 AM, Wee Sritippho <we...@forest.go.th> wrote:
> On 28/7/2559 15:54, Simone Tiraboschi wrote:
>
> On Thu, Jul 28, 2016 at 10:41 AM, Wee Sritippho <we...@forest.go.th> wrote:
>>
>> On 21/7/2559 16:53, Simone Tiraboschi wrote:
>>
>> On Thu, Jul 21, 2016 at 11:43 AM, Wee Sritippho <we...@forest.go.th>
>> wrote:
>>
>>>
>>> Can I just follow
>>> http://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine
>>> until step 3 and do everything else via GUI?
>>
>> Yes, absolutely.
>>
>>
>> Hi, I upgrade a host (host02) via GUI and now its score is 0. Restarted
>> the services but the result is still the same. Kinda lost now. What should I
>> do next?
>>
>
> Can you please attach ovirt-ha-agent logs?
>
>
> Yes, here are the logs:
> https://app.box.com/s/b4urjty8dsuj98n3ywygpk3oh5o7pbsh

Thanks Wee,
your issue is here:
MainThread::ERROR::2016-07-17
14:32:45,586::storage_server::143::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(_validate_pre_connected_path)
The hosted-engine storage domain is already mounted on
'/rhev/data-center/mnt/glusterSD/host02.ovirt.forest.go.th:_hosted__engine/639e689c-8493-479b-a6eb-cc92b6fc4cf4'
with a path that is not supported anymore: the right path should be
'/rhev/data-center/mnt/glusterSD/host01.ovirt.forest.go.th:_hosted__engine/639e689c-8493-479b-a6eb-cc92b6fc4cf4'.

Did you manually tried to avoid the issue of a single entry point for
the gluster FS volume using host01.ovirt.forest.go.th:_hosted__engine
and host02.ovirt.forest.go.th:_hosted__engine there?
This could cause a lot of confusion since the code could not detect
that the storage domain is the same and you can end with it mounted
twice into different locations and a lot of issues.
The correct solution of that issue was this one:
https://bugzilla.redhat.com/show_bug.cgi?id=1298693#c20

Now, to have it fixed on your env you have to hack a bit.
First step, you have to edit
/etc/ovirt-hosted-engine/hosted-engine.conf on all your hosted-engine
hosts to ensure that the storage field always point to the same entry
point (host01 for instance)
Then on each host you can add something like:
mnt_options=backupvolfile-server=host02.ovirt.forest.go.th:host03.ovirt.forest.go.th,fetch-attempts=2,log-level=WARNING,log-file=/var/log/engine_domain.log

Then check the representation of your storage connection in the table
storage_server_connections of the engine DB and make sure that
connection refers to the entry point you used in hosted-engine.conf on
all your hosts, you have lastly to set the value of mount_options also
here.

Please tune also the value of network.ping-timeout for your glusterFS
volume to avoid this:
 https://bugzilla.redhat.com/show_bug.cgi?id=1319657#c17

> --
> Wee
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Safe to upgrade HE hosts from GUI?

2016-08-01 Thread Simone Tiraboschi
On Mon, Aug 1, 2016 at 3:54 AM, Wee Sritippho <we...@forest.go.th> wrote:
> On 29/7/2559 17:07, Simone Tiraboschi wrote:
>>
>> On Fri, Jul 29, 2016 at 11:35 AM, Wee Sritippho <we...@forest.go.th>
>> wrote:
>>>
>>> On 29/7/2559 15:50, Simone Tiraboschi wrote:
>>>>
>>>> On Fri, Jul 29, 2016 at 6:31 AM, Wee Sritippho <we...@forest.go.th>
>>>> wrote:
>>>>>
>>>>> On 28/7/2559 15:54, Simone Tiraboschi wrote:
>>>>>
>>>>> On Thu, Jul 28, 2016 at 10:41 AM, Wee Sritippho <we...@forest.go.th>
>>>>> wrote:
>>>>>>
>>>>>> On 21/7/2559 16:53, Simone Tiraboschi wrote:
>>>>>>
>>>>>> On Thu, Jul 21, 2016 at 11:43 AM, Wee Sritippho <we...@forest.go.th>
>>>>>> wrote:
>>>>>>
>>>>>>> Can I just follow
>>>>>>>
>>>>>>>
>>>>>>> http://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine
>>>>>>> until step 3 and do everything else via GUI?
>>>>>>
>>>>>> Yes, absolutely.
>>>>>>
>>>>>>
>>>>>> Hi, I upgrade a host (host02) via GUI and now its score is 0.
>>>>>> Restarted
>>>>>> the services but the result is still the same. Kinda lost now. What
>>>>>> should I
>>>>>> do next?
>>>>>>
>>>>> Can you please attach ovirt-ha-agent logs?
>>>>>
>>>>>
>>>>> Yes, here are the logs:
>>>>> https://app.box.com/s/b4urjty8dsuj98n3ywygpk3oh5o7pbsh
>>>>
>>>> Thanks Wee,
>>>> your issue is here:
>>>> MainThread::ERROR::2016-07-17
>>>>
>>>>
>>>> 14:32:45,586::storage_server::143::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(_validate_pre_connected_path)
>>>> The hosted-engine storage domain is already mounted on
>>>>
>>>>
>>>> '/rhev/data-center/mnt/glusterSD/host02.ovirt.forest.go.th:_hosted__engine/639e689c-8493-479b-a6eb-cc92b6fc4cf4'
>>>> with a path that is not supported anymore: the right path should be
>>>>
>>>>
>>>> '/rhev/data-center/mnt/glusterSD/host01.ovirt.forest.go.th:_hosted__engine/639e689c-8493-479b-a6eb-cc92b6fc4cf4'.
>>>>
>>>> Did you manually tried to avoid the issue of a single entry point for
>>>> the gluster FS volume using host01.ovirt.forest.go.th:_hosted__engine
>>>> and host02.ovirt.forest.go.th:_hosted__engine there?
>>>> This could cause a lot of confusion since the code could not detect
>>>> that the storage domain is the same and you can end with it mounted
>>>> twice into different locations and a lot of issues.
>>>> The correct solution of that issue was this one:
>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1298693#c20
>>>>
>>>> Now, to have it fixed on your env you have to hack a bit.
>>>> First step, you have to edit
>>>> /etc/ovirt-hosted-engine/hosted-engine.conf on all your hosted-engine
>>>> hosts to ensure that the storage field always point to the same entry
>>>> point (host01 for instance)
>>>> Then on each host you can add something like:
>>>>
>>>>
>>>> mnt_options=backupvolfile-server=host02.ovirt.forest.go.th:host03.ovirt.forest.go.th,fetch-attempts=2,log-level=WARNING,log-file=/var/log/engine_domain.log
>>>>
>>>> Then check the representation of your storage connection in the table
>>>> storage_server_connections of the engine DB and make sure that
>>>> connection refers to the entry point you used in hosted-engine.conf on
>>>> all your hosts, you have lastly to set the value of mount_options also
>>>> here.
>>>
>>> Weird. The configuration in all hosts are already referring to host01.
>>
>> but for sure you have a connection pointing to host02 somewhere, did
>> you try to manually deploy from CLI connecting the gluster volume on
>> host02?
>
> If I recall correctly, yes.

Ok, so please reboot your host before trying again to make sure that
every reference get cleaned.

>>> Also, in the storage_server_connections table:
>>>
>>> engine=> SELECT * FROM storage_server_connections;
>>>id  | connection|
>>> user_name | password | iqn | port | portal | storage_type | mount_options
>>> |
>>> vfs_type
>>>   | nfs_version | nfs_timeo | nfs_retrans
>>>
>>> --+--+---+--+-+--++--+---+--
>>> -+-+---+-
>>>   bd78d299-c8ff-4251-8aab-432ce6443ae8 |
>>> host01.ovirt.forest.go.th:/hosted_engine |   | | |  | 1
>>> |7 |   | glusterfs
>>>   | |   |
>>> (1 row)
>>>
>>>
>>>> Please tune also the value of network.ping-timeout for your glusterFS
>>>> volume to avoid this:
>>>>https://bugzilla.redhat.com/show_bug.cgi?id=1319657#c17
>>>
>>>
>>> --
>>> Wee
>>>
>
> --
> Wee
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] ovirt-ha-agent and too many open files error

2016-08-09 Thread Simone Tiraboschi
On Tue, Aug 9, 2016 at 4:59 PM, Gianluca Cecchi
 wrote:
> Hello,
> I have a 4.0 test environment (single host with self hosted engine) where I
> have 6 VMs defined (5 running) and no much activity.
>
> I do't monitor this system very much.
>
> Now I have connected to it to evaluate upgrade to 4.0.1 and see that about
> 15 days ago the ovirt-ha-agent died because of too many open files
>
> [root@ractor ovirt-hosted-engine-ha]# systemctl status ovirt-ha-agent -l
> ● ovirt-ha-agent.service - oVirt Hosted Engine High Availability Monitoring
> Agent
>Loaded: loaded (/usr/lib/systemd/system/ovirt-ha-agent.service; enabled;
> vendor preset: disabled)
>Active: inactive (dead) since Fri 2016-07-22 16:39:49 CEST; 2 weeks 4
> days ago
>  Main PID: 72795 (code=exited, status=0/SUCCESS)
>
> Jul 22 16:39:47 ractor.mydomain ovirt-ha-agent[72795]: self.set_file(fd)
> Jul 22 16:39:47 ractor.mydomain ovirt-ha-agent[72795]: File
> "/usr/lib64/python2.7/asyncore.py", line 657, in set_file
> Jul 22 16:39:47 ractor.mydomain ovirt-ha-agent[72795]: self.socket =
> file_wrapper(fd)
> Jul 22 16:39:47 ractor.mydomain ovirt-ha-agent[72795]: File
> "/usr/lib64/python2.7/asyncore.py", line 616, in __init__
> Jul 22 16:39:47 ractor.mydomain ovirt-ha-agent[72795]: self.fd = os.dup(fd)
> Jul 22 16:39:47 ractor.mydomain ovirt-ha-agent[72795]: OSError: [Errno 24]
> Too many open files
> Jul 22 16:39:47 ractor.mydomain ovirt-ha-agent[72795]: ovirt-ha-agent
> ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine ERROR Shutting down
> the agent because of 3 failures in a row!
> Jul 22 16:39:47 ractor.mydomain ovirt-ha-agent[72795]:
> ERROR:ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine:Shutting down
> the agent because of 3 failures in a row!
> Jul 22 16:39:49 ractor.mydomain ovirt-ha-agent[72795]:
> WARNING:ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine:The VM is
> running locally or we have no data, keeping the domain monitor.
> Jul 22 16:39:49 ractor.mydomain ovirt-ha-agent[72795]:
> INFO:ovirt_hosted_engine_ha.agent.agent.Agent:Agent shutting down
>
> Is this sort of known problem or any reason to investigate?
> It seems very strange to have reached this limit
>
> I presume the agent runs as vdsm user and that the oVirt installation
> creates the file
> /etc/security/limits.d/99-vdsm.conf
>
> with
> # This limits are intended for medium VDSM hosts, for large hosts scale
> these
> # numbers appropriately.
>
> # nproc should be the maximum amount of storage operations usage.
> # VMs run by "qemu" user, vm processes are not relavent to "vdsm" user
> limits.
> vdsm - nproc 4096
>
> # nofile should be at least 3(stdin,stdour,stderr) * each external process.
> # 3 * 4096 = 12288
> vdsm - nofile 12288
>
> As a rough estimation (over estimation actually , due to many duplicates) I
> have now:
> # lsof -u vdsm | wc -l
> 488
>
> Anything else to check?

Ciao Gianluca,
can you please report which vdsm version are using there?
we had a similar issue in the past but it should be already solved:
https://bugzilla.redhat.com/1343005

> Gianluca
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt upgrade

2016-08-10 Thread Simone Tiraboschi
On Wed, Aug 10, 2016 at 7:53 AM, Yedidyah Bar David  wrote:
> On Tue, Aug 9, 2016 at 8:39 PM, Budur Nagaraju  wrote:
>> HI
>>
>> How to upgrade oVirt3.5 to 4 ? am using centos6.7.
>
> First upgrade or reinstall your hosts with el7.
> oVirt 3.6+ does not provide packages for el6. el6 hosts will
> still work, but easiest to upgrade to el7 in 3.5.
>
> If you are using hosted-engine, you can follow [1].
>
> Then upgrade to 3.6 using regular docs [2].
>
> In 4.0, the engine is also el7 only. To upgrade, use engine-backup
> to backup, then reinstall the machine with el7, then restore. I do
> not think we have a special page for this, but you can follow the
> doc text of [3], which is also copied in the 4.0 release notes [4].
> Please review [5] for more information about engine backup and restore.
> If you using hosted-engine, you can use the new upgrade tool to upgrade
> the engine vm using the engine appliance. Not sure that's properly
> documented either, see also [6]. Also note that this bug is for 4.0.2,
> to be released soon.

We have this:
https://www.ovirt.org/develop/release-management/features/hosted-engine-migration-to-4-0/

>
> As always, if at all possible, better try first on a test system,
> especially if it's in production, and backup everything beforehand,
> including your VMs etc.
>
> Best regards,
>
> [1] https://www.ovirt.org/documentation/how-to/hosted-engine-host-OS-upgrade/
> [2] http://www.ovirt.org/develop/release-management/releases/3.6/
> [3] https://bugzilla.redhat.com/show_bug.cgi?id=1318580
> [4] https://www.ovirt.org/release/4.0.0/
> [5] 
> http://www.ovirt.org/develop/release-management/features/engine/engine-backup/
> [6] https://bugzilla.redhat.com/show_bug.cgi?id=1319457
> --
> Didi
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] engine-backup and messages in web admin events pane

2016-08-10 Thread Simone Tiraboschi
On Tue, Aug 9, 2016 at 4:50 PM, Gianluca Cecchi
 wrote:
> Hello,
> having oVirt Engine Version: 4.0.0.6-1.el7.centos, when I run on engine VM a
> command like
>
> # engine-backup --mode=backup --scope=all
> --file=/root/engine-backup.archive_$(date +%d-%m-%Y)
> --log=/root/engine-backup_$(date +%d-%m-%Y).log
> Backing up:
> Notifying engine
> - Files
> - Engine database 'engine'
> - DWH database 'ovirt_engine_history'
> Packing into file '/root/engine-backup.archive_09-08-2016'
> Notifying engine
> Done.
>
> I see then that in engine web admin events pane I only have rows of kind:
> Finished
> Started
>
> without any reference to what...
> See here:
> https://drive.google.com/file/d/0BwoPbcrMv8mvYXNLZkZtQ1FFRUU/view?usp=sharing
>
> Is it me running backup in the wrong way or was it improved in 4.0.1?

Your backup it's ok,
I just reproduced it and it seams still like that and I agree with you
that logging there just 'Finished'/'Started' without any reference to
the subject it's a bit confusing.
Can you please open a bug ticket to get it fixed?
thanks

> Thanks,
> Gianluca
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] 3.6 : iSCSI LUN not detected

2016-08-10 Thread Simone Tiraboschi
On Wed, Aug 10, 2016 at 10:38 AM, Alexis HAUSER
 wrote:
> Hi,
>
> I am reinstalling a new Node with a new hosted-engine and I would like to 
> import an iSCSI storage from a previous ovirt installation.
> However, I can see all LUN present on that iSCSI but the one I want... I 
> checked from the iSCSI array and this disk still exists, it's just not 
> detected from Ovirt (3.6)...
> I tried to make a new data domain and chosed that same iSCSI and it's also 
> not detected.
>
> Any ideas ?

Did you deployed the engine from scratch or did your restored there a backup?
In 3.5 we were registering an hosted-engine iSCSI storage domain as a
(fake) direct LUN into the engine to prevent any misuse.

> I didn't remove the storage from the Engine interface on the previous 
> installation, just just turned off all VMs accessing the iSCSI and unplugged 
> the cable. Is it possible there is still a lock file or something from the 
> previous hypervisor ?
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] New host HE 4.0 -> [ ERROR ] Failed to execute stage 'Environment customization': Specified gateway is not pingable

2016-07-21 Thread Simone Tiraboschi
On Thu, Jul 21, 2016 at 9:27 AM, Matt . <yamakasi@gmail.com> wrote:
> Even when I change that on each host before I do a new deploy it still
> uses the old GW address.

If you deploy additional hosts with the CLI utility, each new host
will be deployed consuming the answerfile generated by the first host
and saved on the shared storage.
You have to manually tweak there if you to avoid.

Another smarted option is to deploy additional HE host from the
webadmin interface.


> 2016-07-21 9:24 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>> On Wed, Jul 20, 2016 at 11:57 PM, Matt . <yamakasi@gmail.com> wrote:
>>> OK, I tried to edit the answerfiles on the other hosts to use another
>>> IP but it doesn't ping and still uses the old GW address.
>>>
>>> Any idea ?
>>
>> The gateway address is saved under
>> /etc/ovirt-hosted-engine/hosted-engine.conf
>> on each host.
>>
>>> 2016-07-20 22:27 GMT+02:00 Matt . <yamakasi@gmail.com>:
>>>> Maybe it's a good idea to be able to add multiple addresses it should
>>>> ping in order of response.
>>>>
>>>> 2016-07-20 22:13 GMT+02:00 Sandro Bonazzola <sbona...@redhat.com>:
>>>>> Il 20/Lug/2016 10:06 PM, "Matt ." <yamakasi@gmail.com> ha scritto:
>>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> I'm trying to add a host to a HE installation but my gateway is not
>>>>>> pingable. It's a floating IP which is usable but not pingable in this
>>>>>> case.
>>>>>>
>>>>>> During the first install of the HE I could give any IP to ping and I
>>>>>> used a source IP for the floating GW address.
>>>>>>
>>>>>> What can I do to get around this as the network is OK, even when the
>>>>>> GW is not pingable.
>>>>>
>>>>> I think that gw not pingable will be interpreted as network failure by the
>>>>> agent and will lower the score of the hosts.
>>>>>
>>>>>>
>>>>>> Any idea is welcome!
>>>>>>
>>>>>> Thanks!
>>>>>> ___
>>>>>> Users mailing list
>>>>>> Users@ovirt.org
>>>>>> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] New host HE 4.0 -> [ ERROR ] Failed to execute stage 'Environment customization': Specified gateway is not pingable

2016-07-21 Thread Simone Tiraboschi
On Thu, Jul 21, 2016 at 10:03 AM, Matt . <yamakasi@gmail.com> wrote:
> OK, I will check this laster on, as I was having issues with the
> deploy from the GUI before as nothing really happened, I thought it
> was a bug on the first 4.0 release ?
>
> I think this should be investigated, the whole deployment as it would
> be good to have the possibility to list some GW's in case you use
> failover.
>
> I was even thinking of setting the engine IP there, should be working
> as good as also.

Not really since the agent is pinging also when the engine VM is down
due to maintenance.

> 2016-07-21 9:53 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>> On Thu, Jul 21, 2016 at 9:47 AM, Matt . <yamakasi@gmail.com> wrote:
>>> OK, I already thought that but where is this saved as I don't have any
>>> rights as root to the ha_agent folder on the HE storage ?
>>
>> It's not a folder: it's a tar file directly saved over a VDSM volume.
>> Editing it's a bit tricky.
>>
>>> In worst case scenario, how would i be able to deploy a new hosted
>>> engine from a hosted engine ? This sounds undoable to me and also
>>> unwanted as the HE is OK so far.
>>
>> With a 4.0 engine, when you choose to deploy a new host from the
>> webadmin or from the rest API you can also choose to deploy it as an
>> hosted-engine host if the engine has been deployed with hosted-engine.
>> In this case, if I'm not wrong, you can also specify a different gateway 
>> value.
>>
>>
>>> 2016-07-21 9:41 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>>> On Thu, Jul 21, 2016 at 9:27 AM, Matt . <yamakasi@gmail.com> wrote:
>>>>> Even when I change that on each host before I do a new deploy it still
>>>>> uses the old GW address.
>>>>
>>>> If you deploy additional hosts with the CLI utility, each new host
>>>> will be deployed consuming the answerfile generated by the first host
>>>> and saved on the shared storage.
>>>> You have to manually tweak there if you to avoid.
>>>>
>>>> Another smarted option is to deploy additional HE host from the
>>>> webadmin interface.
>>>>
>>>>
>>>>> 2016-07-21 9:24 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>>>>> On Wed, Jul 20, 2016 at 11:57 PM, Matt . <yamakasi@gmail.com> wrote:
>>>>>>> OK, I tried to edit the answerfiles on the other hosts to use another
>>>>>>> IP but it doesn't ping and still uses the old GW address.
>>>>>>>
>>>>>>> Any idea ?
>>>>>>
>>>>>> The gateway address is saved under
>>>>>> /etc/ovirt-hosted-engine/hosted-engine.conf
>>>>>> on each host.
>>>>>>
>>>>>>> 2016-07-20 22:27 GMT+02:00 Matt . <yamakasi@gmail.com>:
>>>>>>>> Maybe it's a good idea to be able to add multiple addresses it should
>>>>>>>> ping in order of response.
>>>>>>>>
>>>>>>>> 2016-07-20 22:13 GMT+02:00 Sandro Bonazzola <sbona...@redhat.com>:
>>>>>>>>> Il 20/Lug/2016 10:06 PM, "Matt ." <yamakasi@gmail.com> ha scritto:
>>>>>>>>>>
>>>>>>>>>> Hi,
>>>>>>>>>>
>>>>>>>>>> I'm trying to add a host to a HE installation but my gateway is not
>>>>>>>>>> pingable. It's a floating IP which is usable but not pingable in this
>>>>>>>>>> case.
>>>>>>>>>>
>>>>>>>>>> During the first install of the HE I could give any IP to ping and I
>>>>>>>>>> used a source IP for the floating GW address.
>>>>>>>>>>
>>>>>>>>>> What can I do to get around this as the network is OK, even when the
>>>>>>>>>> GW is not pingable.
>>>>>>>>>
>>>>>>>>> I think that gw not pingable will be interpreted as network failure 
>>>>>>>>> by the
>>>>>>>>> agent and will lower the score of the hosts.
>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Any idea is welcome!
>>>>>>>>>>
>>>>>>>>>> Thanks!
>>>>>>>>>> ___
>>>>>>>>>> Users mailing list
>>>>>>>>>> Users@ovirt.org
>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] New host HE 4.0 -> [ ERROR ] Failed to execute stage 'Environment customization': Specified gateway is not pingable

2016-07-21 Thread Simone Tiraboschi
On Wed, Jul 20, 2016 at 11:57 PM, Matt .  wrote:
> OK, I tried to edit the answerfiles on the other hosts to use another
> IP but it doesn't ping and still uses the old GW address.
>
> Any idea ?

The gateway address is saved under
/etc/ovirt-hosted-engine/hosted-engine.conf
on each host.

> 2016-07-20 22:27 GMT+02:00 Matt . :
>> Maybe it's a good idea to be able to add multiple addresses it should
>> ping in order of response.
>>
>> 2016-07-20 22:13 GMT+02:00 Sandro Bonazzola :
>>> Il 20/Lug/2016 10:06 PM, "Matt ."  ha scritto:

 Hi,

 I'm trying to add a host to a HE installation but my gateway is not
 pingable. It's a floating IP which is usable but not pingable in this
 case.

 During the first install of the HE I could give any IP to ping and I
 used a source IP for the floating GW address.

 What can I do to get around this as the network is OK, even when the
 GW is not pingable.
>>>
>>> I think that gw not pingable will be interpreted as network failure by the
>>> agent and will lower the score of the hosts.
>>>

 Any idea is welcome!

 Thanks!
 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] trying to upgrade host engine 4.0 to 4.1

2016-07-21 Thread Simone Tiraboschi
On Wed, Jul 20, 2016 at 10:26 PM, Rafael Almeida
 wrote:
> Hello, i'm trying to upgrade my ovirt engine host  4.0.0.6-1 in my centos
> 7.2 x64 to 4.1, using:

We just released 4.0.1, not 4.1: https://www.ovirt.org/release/4.0.1/

> yum update "ovirt-engine-setup*"
>
> but not work:
>
> No packages marked for update
> my repos are:
>
> CentOS-Base.repo   CentOS-Media.repoepel-testing.repo
> CentOS-CR.repo CentOS-Sources.repo  ovirt-4.0-dependencies.repo
> CentOS-Debuginfo.repo  CentOS-Vault.repoovirt-4.0.repo
> CentOS-fasttrack.repo  epel.repo
> if I try with the command
>
> yum install vdsm*
>
> list the new version (4.1)
>
> vdsm  x86_64 4.18.4.1-0.el7.centos  ovirt-4.0
>
> what is the problem? please helpme
>
> Thnx
>
> Rafael Almeida Orellana
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] 3.5 to 3.6 upgrade stuck

2016-07-21 Thread Simone Tiraboschi
On Thu, Jul 21, 2016 at 7:14 AM, Robert Story  wrote:
> I have a 3.5 hosted-engine with 5 el7 nodes. Today I tried upgrading to 3.6.
> The engine upgrade went great, no problems.
>
> I had a host in maintenance mode, so I added the 3.6 repos and ran yum
> update. I waited for the upgrade successful message. I checked the score
> for the node, and it was still 2400, not 3400. Tried rebooting, but no
> luck. So I put another host in maintenance mode, and had the same result.

MainThread::INFO::2016-07-20
23:44:30,352::upgrade::1031::ovirt_hosted_engine_ha.lib.upgrade.StorageServer::(upgrade_35_36)
Successfully upgraded

Everything seams OK on the upgrade path.


> Both nodes are getting this error:
>
> MainThread::ERROR::2016-07-21 
> 01:05:04,187::brokerlink::279::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(_communicate)
>  Connection closed: Connection timed out
> MainThread::ERROR::2016-07-21 
> 01:05:04,188::agent::205::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
>  Error: 'Failed to start monitor , options {'hostname': 
> 'poseidon.netsec'}: Connection timed out' - trying to restart agent

Can you please attach also broker.log? maybe the issue is somewhere else.

> I've attached logs from the second host coming up after a reboot, along
> with engine log from the same timeframe.
>
> Any suggestions on a way forward would be greatly appreciated.
>
>
> Robert
>
> --
> Senior Software Engineer @ Parsons
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] ovirt-3.6 : Hosted-engine crashed and can't restart

2016-07-21 Thread Simone Tiraboschi
On Wed, Jul 20, 2016 at 5:01 PM, Alexis HAUSER
 wrote:
> After assigning an IP adress to a VLAN network (it was using DHCP by default) 
> that was on the same NIC than ovirtmgmt, my hosted-engine crashed and can't 
> start again...I have no idea how to fix this.
> I had a similar issue some months ago but with a different error. I tried to 
> restart the ha agent that seems to be linked with this error, also restarted 
> the host. I also tried to remove the _DIRECT_IO_ lockfile on the engine 
> storage as it fixed my problem last time but it didn't help...
>
> Any ideas ? Do you think editing manually the logical network in the host and 
> reverting them at it was before crash can help ?
>
>
>
>
>
>
> hosted-engine --vm-status
> Traceback (most recent call last):
>   File "/usr/lib64/python2.7/runpy.py", line 162, in _run_module_as_main
> "__main__", fname, loader, pkg_name)
>   File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code
> exec code in run_globals
>   File 
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_setup/vm_status.py", 
> line 117, in 
> if not status_checker.print_status():
>   File 
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_setup/vm_status.py", 
> line 60, in print_status
> all_host_stats = ha_cli.get_all_host_stats()
>   File 
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py", 
> line 160, in get_all_host_stats
> return self.get_all_stats(self.StatModes.HOST)
>   File 
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py", 
> line 103, in get_all_stats
> self._configure_broker_conn(broker)
>   File 
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py", 
> line 180, in _configure_broker_conn
> dom_type=dom_type)
>   File 
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", 
> line 176, in set_storage_domain
> .format(sd_type, options, e))
> ovirt_hosted_engine_ha.lib.exceptions.RequestError: Failed to set storage 
> domain FilesystemBackend, options {'dom_type': 'nfs3', 'sd_uuid': 
> 'e41807e5-ee68-40a2-a642-cc226ba0e82d'}: Request failed:  'ovirt_hosted_engine_ha.lib.storage_backends.BackendFailureException'>
>
>
> vdsClient -s 0 list
>
> 16450089-911e-4bad-a8b7-98e84a79ef3a
> Status = Down
> nicModel = rtl8139,pv
> statusTime = 4295559350
> exitMessage = Unable to get volume size for domain 
> e41807e5-ee68-40a2-a642-cc226ba0e82d volume 
> 053df3a6-db18-445a-8f75-61c630ab0003
> emulatedMachine = rhel6.5.0
> pid = 0
> vmName = HostedEngine
> devices = [{'index': '0', 'iface': 'virtio', 'format': 'raw', 
> 'bootOrder': '1', 'address': {'slot': '0x06', 'bus': '0x00', 'domain': 
> '0x', 'type': 'pci', 'function': '0x0'}, 'volumeID': 
> '053df3a6-db18-445a-8f75-61c630ab0003', 'imageID': 
> 'b6daa50d-adad-46a5-8f5f-accfb155a1e1', 'readonly': 'false', 'domainID': 
> 'e41807e5-ee68-40a2-a642-cc226ba0e82d', 'deviceId': 
> 'b6daa50d-adad-46a5-8f5f-accfb155a1e1', 'poolID': 
> '----', 'device': 'disk', 'shared': 
> 'exclusive', 'propagateErrors': 'off', 'type': 'disk'}, {'nicModel': 'pv', 
> 'macAddr': '00:16:3e:1c:4b:81', 'linkActive': 'true', 'network': 'ovirtmgmt', 
> 'deviceId': '0aeaea2f-a419-43cc-92d7-8422f6aa9223', 'address': 'None', 
> 'device': 'bridge', 'type': 'interface'}, {'index': '2', 'iface': 'ide', 
> 'readonly': 'true', 'deviceId': '8c3179ac-b322-4f5c-9449-c52e3665e0ae', 
> 'address': {'bus': '1', 'controller': '0', 'type': 'drive', 'target': '0', 
> 'unit': '0'}, 'device': 'cdrom', 'shared': 'false', 'path': '', 'type': 
> 'disk'}, {'device': 'scsi', 'model': 'virtio-scsi', 'type': 'controller', 
> 'deviceId': '21db0c6e-071c-48ff-b905-95478b37c384', 'address': {'slot': 
> '0x04', 'bus': '0x00', 'domain': '0x', 'type': 'pci', 'function': 
> '0x0'}}, {'device': 'usb', 'type': 'controller', 'deviceId': 
> 'c0384f68-d0c9-4ebb-a779-8dc9911ce2f8', 'address': {'slot': '0x01', 'bus': 
> '0x00', 'domain': '0x', 'type': 'pci', 'function': '0x2'}}, {'device': 
> 'ide', 'type': 'controller', 'deviceId': 
> 'd5a2dd13-138a-482b-9bc3-994b10ec4100', 'address': {'slot': '0x01', 'bus': 
> '0x00', 'domain': '0x', 'type': 'pci', 'function': '0x1'}}, {'device': 
> 'virtio-serial', 'type': 'controller', 'deviceId': 
> '9e695172-c9b0-47df-bc76-8170219dec28', 'address': {'slot': '0x05', 'bus': 
> '0x00', 'domain': '0x', 'type': 'pci', 'function': '0x0'}}]
> guestDiskMapping = {}
> vmType = kvm
> displaySecurePort = -1
> exitReason = 1
> memSize = 6000
> displayPort = -1
> clientIp =
> spiceSecureChannels = 
> smain,sdisplay,sinputs,scursor,splayback,srecord,ssmartcard,susbredir
> smp = 4
> displayIp = 0
> display = vnc
> exitCode = 1
>
>
> systemctl status ovirt-ha-agent.service -l
> ● 

Re: [ovirt-users] New host HE 4.0 -> [ ERROR ] Failed to execute stage 'Environment customization': Specified gateway is not pingable

2016-07-21 Thread Simone Tiraboschi
On Thu, Jul 21, 2016 at 9:47 AM, Matt . <yamakasi@gmail.com> wrote:
> OK, I already thought that but where is this saved as I don't have any
> rights as root to the ha_agent folder on the HE storage ?

It's not a folder: it's a tar file directly saved over a VDSM volume.
Editing it's a bit tricky.

> In worst case scenario, how would i be able to deploy a new hosted
> engine from a hosted engine ? This sounds undoable to me and also
> unwanted as the HE is OK so far.

With a 4.0 engine, when you choose to deploy a new host from the
webadmin or from the rest API you can also choose to deploy it as an
hosted-engine host if the engine has been deployed with hosted-engine.
In this case, if I'm not wrong, you can also specify a different gateway value.


> 2016-07-21 9:41 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>> On Thu, Jul 21, 2016 at 9:27 AM, Matt . <yamakasi@gmail.com> wrote:
>>> Even when I change that on each host before I do a new deploy it still
>>> uses the old GW address.
>>
>> If you deploy additional hosts with the CLI utility, each new host
>> will be deployed consuming the answerfile generated by the first host
>> and saved on the shared storage.
>> You have to manually tweak there if you to avoid.
>>
>> Another smarted option is to deploy additional HE host from the
>> webadmin interface.
>>
>>
>>> 2016-07-21 9:24 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>>> On Wed, Jul 20, 2016 at 11:57 PM, Matt . <yamakasi@gmail.com> wrote:
>>>>> OK, I tried to edit the answerfiles on the other hosts to use another
>>>>> IP but it doesn't ping and still uses the old GW address.
>>>>>
>>>>> Any idea ?
>>>>
>>>> The gateway address is saved under
>>>> /etc/ovirt-hosted-engine/hosted-engine.conf
>>>> on each host.
>>>>
>>>>> 2016-07-20 22:27 GMT+02:00 Matt . <yamakasi@gmail.com>:
>>>>>> Maybe it's a good idea to be able to add multiple addresses it should
>>>>>> ping in order of response.
>>>>>>
>>>>>> 2016-07-20 22:13 GMT+02:00 Sandro Bonazzola <sbona...@redhat.com>:
>>>>>>> Il 20/Lug/2016 10:06 PM, "Matt ." <yamakasi@gmail.com> ha scritto:
>>>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> I'm trying to add a host to a HE installation but my gateway is not
>>>>>>>> pingable. It's a floating IP which is usable but not pingable in this
>>>>>>>> case.
>>>>>>>>
>>>>>>>> During the first install of the HE I could give any IP to ping and I
>>>>>>>> used a source IP for the floating GW address.
>>>>>>>>
>>>>>>>> What can I do to get around this as the network is OK, even when the
>>>>>>>> GW is not pingable.
>>>>>>>
>>>>>>> I think that gw not pingable will be interpreted as network failure by 
>>>>>>> the
>>>>>>> agent and will lower the score of the hosts.
>>>>>>>
>>>>>>>>
>>>>>>>> Any idea is welcome!
>>>>>>>>
>>>>>>>> Thanks!
>>>>>>>> ___
>>>>>>>> Users mailing list
>>>>>>>> Users@ovirt.org
>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] New host HE 4.0 -> [ ERROR ] Failed to execute stage 'Environment customization': Specified gateway is not pingable

2016-07-21 Thread Simone Tiraboschi
On Thu, Jul 21, 2016 at 12:28 PM, Matt . <yamakasi@gmail.com> wrote:
> No problem I checked it out and applied it, no solution it seems.

Do you mean that it's still referring to your initial gateway or that
the deploy failed?

> 2016-07-21 12:14 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>> On Thu, Jul 21, 2016 at 12:09 PM, Matt . <yamakasi@gmail.com> wrote:
>>> OK, I did a deploy from the engine and the Hosted is "Not Active" and
>>
>> Yes, sorry, I forgot to mention that.
>> We have an open bug about re-activating an hosted-engine host from the
>> engine: https://bugzilla.redhat.com/show_bug.cgi?id=1353600
>> Connecting to the host and exiting the maintenance mode from there is
>> the temporary solution.
>>
>>> Engine HA is empty on 4.0.1
>>>
>>> 2016-07-21 11:58 GMT+02:00 Matt . <yamakasi@gmail.com>:
>>>> True, but you can up an IP on any device in that case because you know
>>>> what you are doing.
>>>>
>>>> I think that this can be finetuned, or at least multiple IPs would be
>>>> nice as I now need to add one IP from my vrrp setup, so if there is a
>>>> failover or I need to take down the second switch I'm also done.
>>>>
>>>> 2016-07-21 10:23 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>>>> On Thu, Jul 21, 2016 at 10:03 AM, Matt . <yamakasi@gmail.com> wrote:
>>>>>> OK, I will check this laster on, as I was having issues with the
>>>>>> deploy from the GUI before as nothing really happened, I thought it
>>>>>> was a bug on the first 4.0 release ?
>>>>>>
>>>>>> I think this should be investigated, the whole deployment as it would
>>>>>> be good to have the possibility to list some GW's in case you use
>>>>>> failover.
>>>>>>
>>>>>> I was even thinking of setting the engine IP there, should be working
>>>>>> as good as also.
>>>>>
>>>>> Not really since the agent is pinging also when the engine VM is down
>>>>> due to maintenance.
>>>>>
>>>>>> 2016-07-21 9:53 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>>>>>> On Thu, Jul 21, 2016 at 9:47 AM, Matt . <yamakasi@gmail.com> wrote:
>>>>>>>> OK, I already thought that but where is this saved as I don't have any
>>>>>>>> rights as root to the ha_agent folder on the HE storage ?
>>>>>>>
>>>>>>> It's not a folder: it's a tar file directly saved over a VDSM volume.
>>>>>>> Editing it's a bit tricky.
>>>>>>>
>>>>>>>> In worst case scenario, how would i be able to deploy a new hosted
>>>>>>>> engine from a hosted engine ? This sounds undoable to me and also
>>>>>>>> unwanted as the HE is OK so far.
>>>>>>>
>>>>>>> With a 4.0 engine, when you choose to deploy a new host from the
>>>>>>> webadmin or from the rest API you can also choose to deploy it as an
>>>>>>> hosted-engine host if the engine has been deployed with hosted-engine.
>>>>>>> In this case, if I'm not wrong, you can also specify a different 
>>>>>>> gateway value.
>>>>>>>
>>>>>>>
>>>>>>>> 2016-07-21 9:41 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>>>>>>>> On Thu, Jul 21, 2016 at 9:27 AM, Matt . <yamakasi@gmail.com> 
>>>>>>>>> wrote:
>>>>>>>>>> Even when I change that on each host before I do a new deploy it 
>>>>>>>>>> still
>>>>>>>>>> uses the old GW address.
>>>>>>>>>
>>>>>>>>> If you deploy additional hosts with the CLI utility, each new host
>>>>>>>>> will be deployed consuming the answerfile generated by the first host
>>>>>>>>> and saved on the shared storage.
>>>>>>>>> You have to manually tweak there if you to avoid.
>>>>>>>>>
>>>>>>>>> Another smarted option is to deploy additional HE host from the
>>>>>>>>> webadmin interface.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>>

Re: [ovirt-users] New host HE 4.0 -> [ ERROR ] Failed to execute stage 'Environment customization': Specified gateway is not pingable

2016-07-21 Thread Simone Tiraboschi
On Thu, Jul 21, 2016 at 12:38 PM, Matt . <yamakasi@gmail.com> wrote:
> The deploy went OK from the GUI but it's still not "active" as Hosted
> engine Host. The install also wen very fast so it looked like it
> didn't do anything witl the deploy setting in the GUI @ the hosted
> engine tab.

The difference with a standard host is just that it should create
/etc/ovirt-hosted-engine/hosted-engine.conf and start ovirt-ha-agent
and ovirt-ha-broker so it shouldn't be that longer.

At the end, due to the issue with syncing the local maintenance mode
from the engine, you have to connect to the host and run
'hosted-engine -set-maintenance --mode=none' from there.

Can you please post the result of
hosted-engine --vm-status and the status of ovirt-ha-agent?

> Or should I re-deploy again on the commandline after adding the host
> through the GUI ?
>
> About the initial GW, I didn't see that in the GUI.
>
> 2016-07-21 12:31 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>> On Thu, Jul 21, 2016 at 12:28 PM, Matt . <yamakasi@gmail.com> wrote:
>>> No problem I checked it out and applied it, no solution it seems.
>>
>> Do you mean that it's still referring to your initial gateway or that
>> the deploy failed?
>>
>>> 2016-07-21 12:14 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>>> On Thu, Jul 21, 2016 at 12:09 PM, Matt . <yamakasi@gmail.com> wrote:
>>>>> OK, I did a deploy from the engine and the Hosted is "Not Active" and
>>>>
>>>> Yes, sorry, I forgot to mention that.
>>>> We have an open bug about re-activating an hosted-engine host from the
>>>> engine: https://bugzilla.redhat.com/show_bug.cgi?id=1353600
>>>> Connecting to the host and exiting the maintenance mode from there is
>>>> the temporary solution.
>>>>
>>>>> Engine HA is empty on 4.0.1
>>>>>
>>>>> 2016-07-21 11:58 GMT+02:00 Matt . <yamakasi@gmail.com>:
>>>>>> True, but you can up an IP on any device in that case because you know
>>>>>> what you are doing.
>>>>>>
>>>>>> I think that this can be finetuned, or at least multiple IPs would be
>>>>>> nice as I now need to add one IP from my vrrp setup, so if there is a
>>>>>> failover or I need to take down the second switch I'm also done.
>>>>>>
>>>>>> 2016-07-21 10:23 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>>>>>> On Thu, Jul 21, 2016 at 10:03 AM, Matt . <yamakasi@gmail.com> wrote:
>>>>>>>> OK, I will check this laster on, as I was having issues with the
>>>>>>>> deploy from the GUI before as nothing really happened, I thought it
>>>>>>>> was a bug on the first 4.0 release ?
>>>>>>>>
>>>>>>>> I think this should be investigated, the whole deployment as it would
>>>>>>>> be good to have the possibility to list some GW's in case you use
>>>>>>>> failover.
>>>>>>>>
>>>>>>>> I was even thinking of setting the engine IP there, should be working
>>>>>>>> as good as also.
>>>>>>>
>>>>>>> Not really since the agent is pinging also when the engine VM is down
>>>>>>> due to maintenance.
>>>>>>>
>>>>>>>> 2016-07-21 9:53 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>>>>>>>> On Thu, Jul 21, 2016 at 9:47 AM, Matt . <yamakasi@gmail.com> 
>>>>>>>>> wrote:
>>>>>>>>>> OK, I already thought that but where is this saved as I don't have 
>>>>>>>>>> any
>>>>>>>>>> rights as root to the ha_agent folder on the HE storage ?
>>>>>>>>>
>>>>>>>>> It's not a folder: it's a tar file directly saved over a VDSM volume.
>>>>>>>>> Editing it's a bit tricky.
>>>>>>>>>
>>>>>>>>>> In worst case scenario, how would i be able to deploy a new hosted
>>>>>>>>>> engine from a hosted engine ? This sounds undoable to me and also
>>>>>>>>>> unwanted as the HE is OK so far.
>>>>>>>>>
>>>>>>>>> With a 4.0 engine, when you choose to deploy a new host from the
>>>>>>>>> w

Re: [ovirt-users] Safe to upgrade HE hosts from GUI?

2016-07-21 Thread Simone Tiraboschi
On Thu, Jul 21, 2016 at 11:43 AM, Wee Sritippho  wrote:

> Hi,
>
> I used to follow
> http://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine
> when upgrading Hosted Engine (HE) but always fail to make the engine VM
> migrate to the fresh upgraded host as described in step 7. Furthermore, the
> update available icon never disappeared from the GUI.
>
Yes, you are right on that: it will happen only upgrading from 3.5, where
maximum score for an hosted-engine hosts was 2400 points, to 3.6 where the
maximum score is 3400.
On 3.6.z upgrades all the hosts are already at 3400 points and so the VM
will not migrate fro that.


> So I though using the GUI might be better for an amateur like me.
>
> Can I just follow
> http://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine
> until step 3 and do everything else via GUI?
>
Yes, absolutely.


> Thank you,
>
> --
> Wee
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] New host HE 4.0 -> [ ERROR ] Failed to execute stage 'Environment customization': Specified gateway is not pingable

2016-07-21 Thread Simone Tiraboschi
On Thu, Jul 21, 2016 at 12:09 PM, Matt . <yamakasi@gmail.com> wrote:
> OK, I did a deploy from the engine and the Hosted is "Not Active" and

Yes, sorry, I forgot to mention that.
We have an open bug about re-activating an hosted-engine host from the
engine: https://bugzilla.redhat.com/show_bug.cgi?id=1353600
Connecting to the host and exiting the maintenance mode from there is
the temporary solution.

> Engine HA is empty on 4.0.1
>
> 2016-07-21 11:58 GMT+02:00 Matt . <yamakasi@gmail.com>:
>> True, but you can up an IP on any device in that case because you know
>> what you are doing.
>>
>> I think that this can be finetuned, or at least multiple IPs would be
>> nice as I now need to add one IP from my vrrp setup, so if there is a
>> failover or I need to take down the second switch I'm also done.
>>
>> 2016-07-21 10:23 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>> On Thu, Jul 21, 2016 at 10:03 AM, Matt . <yamakasi@gmail.com> wrote:
>>>> OK, I will check this laster on, as I was having issues with the
>>>> deploy from the GUI before as nothing really happened, I thought it
>>>> was a bug on the first 4.0 release ?
>>>>
>>>> I think this should be investigated, the whole deployment as it would
>>>> be good to have the possibility to list some GW's in case you use
>>>> failover.
>>>>
>>>> I was even thinking of setting the engine IP there, should be working
>>>> as good as also.
>>>
>>> Not really since the agent is pinging also when the engine VM is down
>>> due to maintenance.
>>>
>>>> 2016-07-21 9:53 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>>>> On Thu, Jul 21, 2016 at 9:47 AM, Matt . <yamakasi@gmail.com> wrote:
>>>>>> OK, I already thought that but where is this saved as I don't have any
>>>>>> rights as root to the ha_agent folder on the HE storage ?
>>>>>
>>>>> It's not a folder: it's a tar file directly saved over a VDSM volume.
>>>>> Editing it's a bit tricky.
>>>>>
>>>>>> In worst case scenario, how would i be able to deploy a new hosted
>>>>>> engine from a hosted engine ? This sounds undoable to me and also
>>>>>> unwanted as the HE is OK so far.
>>>>>
>>>>> With a 4.0 engine, when you choose to deploy a new host from the
>>>>> webadmin or from the rest API you can also choose to deploy it as an
>>>>> hosted-engine host if the engine has been deployed with hosted-engine.
>>>>> In this case, if I'm not wrong, you can also specify a different gateway 
>>>>> value.
>>>>>
>>>>>
>>>>>> 2016-07-21 9:41 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>>>>>> On Thu, Jul 21, 2016 at 9:27 AM, Matt . <yamakasi@gmail.com> wrote:
>>>>>>>> Even when I change that on each host before I do a new deploy it still
>>>>>>>> uses the old GW address.
>>>>>>>
>>>>>>> If you deploy additional hosts with the CLI utility, each new host
>>>>>>> will be deployed consuming the answerfile generated by the first host
>>>>>>> and saved on the shared storage.
>>>>>>> You have to manually tweak there if you to avoid.
>>>>>>>
>>>>>>> Another smarted option is to deploy additional HE host from the
>>>>>>> webadmin interface.
>>>>>>>
>>>>>>>
>>>>>>>> 2016-07-21 9:24 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>>>>>>>> On Wed, Jul 20, 2016 at 11:57 PM, Matt . <yamakasi@gmail.com> 
>>>>>>>>> wrote:
>>>>>>>>>> OK, I tried to edit the answerfiles on the other hosts to use another
>>>>>>>>>> IP but it doesn't ping and still uses the old GW address.
>>>>>>>>>>
>>>>>>>>>> Any idea ?
>>>>>>>>>
>>>>>>>>> The gateway address is saved under
>>>>>>>>> /etc/ovirt-hosted-engine/hosted-engine.conf
>>>>>>>>> on each host.
>>>>>>>>>
>>>>>>>>>> 2016-07-20 22:27 GMT+02:00 Matt . <yamakasi@gmail.com>:
>>>>>>>>>>

Re: [ovirt-users] Hosted Engine oVirt 3.6 iscsi multipath

2016-07-20 Thread Simone Tiraboschi
On Wed, Jul 20, 2016 at 7:09 AM, Julian De Marchi
 wrote:
> heya--
>
> i have an ovirt 3.6 installation. all my iscsi LUNs are mulipath, however
> the hostedengine storage is not.
>
> i tried to import the hostedengine storage domain into my cluster but got a
> warning it is controlled via another oVirt cluster. i accepted the warning
> and imported anyway, but it crashed my hosted engine and hosted-engine
> broker restarted as expected.

You don't have to try manually importing it: the engine will
automatically import once you correctly added your first regular
storage domain.

> i am curious of how to configure multipath for the hosted engine storage
> domain ?

We have an open RFE for that, unfortunately it's still not ready.

> many thanks!
>
> --julian
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Failed to start self hosted engine after upgrading oVirt to 4.0

2016-06-28 Thread Simone Tiraboschi
On Tue, Jun 28, 2016 at 5:24 PM, Dan Kenigsberg <dan...@redhat.com> wrote:
> On Tue, Jun 28, 2016 at 04:08:51PM +0200, Simone Tiraboschi wrote:
>> On Tue, Jun 28, 2016 at 3:02 PM, Dan Kenigsberg <dan...@redhat.com> wrote:
>> > On Mon, Jun 27, 2016 at 10:08:33AM +0200, Stefano Danzi wrote:
>> >> Hi!
>> >
>> >
>> > Thanks for the detailed logging!
>> >
>> >>
>> >> The broker error is:
>> >>
>> >> ==> /var/log/ovirt-hosted-engine-ha/agent.log <==
>> >> MainThread::INFO::2016-06-27 
>> >> 09:27:03,311::brokerlink::140::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(start_monitor)
>> >> Success, id 140293563619152
>> >>
>> >> ==> /var/log/ovirt-hosted-engine-ha/broker.log <==
>> >> Thread-25::ERROR::2016-06-27 
>> >> 09:27:03,314::listener::182::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(handle)
>> >> Error while serving connection
>> >> Traceback (most recent call last):
>> >>   File
>> >> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/listener.py",
>> >> line 166, in handle
>> >> data)
>> >>   File
>> >> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/listener.py",
>> >> line 299, in _dispatch
>> >> .set_storage_domain(client, sd_type, **options)
>> >>   File 
>> >> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/storage_broker.py",
>> >> line 66, in set_storage_domain
>> >> self._backends[client].connect()
>> >>   File 
>> >> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/storage_backends.py",
>> >> line 400, in connect
>> >> volUUID=volume.volume_uuid
>> >>   File 
>> >> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/storage_backends.py",
>> >> line 245, in _get_volume_path
>> >> volUUID
>> >
>> > We have two issues here. First is that
>> > https://gerrit.ovirt.org/gitweb?p=ovirt-hosted-engine-ha.git;a=blob;f=ovirt_hosted_engine_ha/lib/storage_backends.py;h=f2fbdc43d0e4afd7539a3a1de75de0cb07bdca9d;hb=HEAD#l271
>> > is still using vdscli to contact vdsm, instead of the preferred
>> > jsonrpccli.
>>
>> We already have this one:
>> https://bugzilla.redhat.com/show_bug.cgi?id=1101554
>>
> Bug 1101554 - [RFE] HE-setup: use vdsm api instead of vdsClient
>
> It does not seem to be the same thing; here I'm referring to
> ovirt-hosted-engine-ha (not -setup). -ha *is* using vdsm API, but in one
> condition it's the soon-to-be-deprecated xmlrpc.

ovirt-hosted-engine-setup is already using just API via jsonrpc.
The issue is that ovirt-hosted-engine-ha somewhere uses vdsClient
which AFAIK internally still uses xmlrpc
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Failed to start self hosted engine after upgrading oVirt to 4.0

2016-06-28 Thread Simone Tiraboschi
On Tue, Jun 28, 2016 at 3:02 PM, Dan Kenigsberg  wrote:
> On Mon, Jun 27, 2016 at 10:08:33AM +0200, Stefano Danzi wrote:
>> Hi!
>
>
> Thanks for the detailed logging!
>
>>
>> The broker error is:
>>
>> ==> /var/log/ovirt-hosted-engine-ha/agent.log <==
>> MainThread::INFO::2016-06-27 
>> 09:27:03,311::brokerlink::140::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(start_monitor)
>> Success, id 140293563619152
>>
>> ==> /var/log/ovirt-hosted-engine-ha/broker.log <==
>> Thread-25::ERROR::2016-06-27 
>> 09:27:03,314::listener::182::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(handle)
>> Error while serving connection
>> Traceback (most recent call last):
>>   File
>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/listener.py",
>> line 166, in handle
>> data)
>>   File
>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/listener.py",
>> line 299, in _dispatch
>> .set_storage_domain(client, sd_type, **options)
>>   File 
>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/storage_broker.py",
>> line 66, in set_storage_domain
>> self._backends[client].connect()
>>   File 
>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/storage_backends.py",
>> line 400, in connect
>> volUUID=volume.volume_uuid
>>   File 
>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/storage_backends.py",
>> line 245, in _get_volume_path
>> volUUID
>
> We have two issues here. First is that
> https://gerrit.ovirt.org/gitweb?p=ovirt-hosted-engine-ha.git;a=blob;f=ovirt_hosted_engine_ha/lib/storage_backends.py;h=f2fbdc43d0e4afd7539a3a1de75de0cb07bdca9d;hb=HEAD#l271
> is still using vdscli to contact vdsm, instead of the preferred
> jsonrpccli.

We already have this one:
https://bugzilla.redhat.com/show_bug.cgi?id=1101554

> The second is that vdscli.connect's heuristic ends up reading the local
> server address from vdsm config, where it finds the default ipv6-local
> address of "::".
>
> Please try setting
>
> [addresses]
> management_ip='0.0.0.0'
>
> in your /etc/vdsm/vdsm.conf instead of the crontab hacks.
>
>
> Would you please open a bug about the two issues
> (ovirt-hosted-engine-ha and vdsm networking)?
>
> Would you report the output of `netstats -nltp` on your host, as I do
> not completely understand why no interface (not even the loopback one)
> was listening on ipv6?
>
> Regards,
> Dan.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Move hosted engine to another storage

2016-07-07 Thread Simone Tiraboschi
On Thu, Jul 7, 2016 at 6:24 PM, Cristian Mammoli  wrote:
> Hi list, need to move the hosted engine from a storage to another. Both the
> source and the destination are iSCSI LUN.
> What is correct procedure (if any)?
> I didn't find anything in the docs

Ciao Christian,
theoretically you could run engine-backup on the current engine VM,
redeploy from scratch on the new storage domain and before running
engine setup, inject there the backup.

Unfortunately we know that is not that simple:
https://bugzilla.redhat.com/show_bug.cgi?id=1240466#c21
and you have to manually remove any reference to the previous
hosted-engine storage domain and hosted-engine VM from your restored
DB since they are locked in the engine and so you couldn't remove from
there.

Improving engine-backup to automatically filter hosted-engine
references is an RFE for 4.1.


> Thanks
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Fail to setup Hyperconverged Infrastructure using oVirt and Gluster

2016-07-08 Thread Simone Tiraboschi
On Thu, Jul 7, 2016 at 5:44 PM, Dewey Du  wrote:
>  "PermitRootLogin without-password" is disabled. Is this a MUST? I am
> following the steps from the link below, and there is no steps needed to
> enable "PermitRootLogin without-password".

No, it's not.
The engine will connect using a key deployed on the host by hosted-engine-setup.

> http://blogs-ramesh.blogspot.jp/2016/01/ovirt-and-gluster-hyperconvergence.html
>
> Do you have anyother instructions to setup
> ovirt-and-gluster-hyperconvergence?
>
> By the way, the "localhost.localdomain"  in the engine.log "Failed to
> authenticate session with host 'hosted_engine_1': SSH authentication to
> 'root@localhost.localdomain' failed." means the engine VM or Host-01?

It means the engine and the issue is just there.
Can you please check the hostname of your host?
We had a bug in past about that but now hosted-engine-setup should
refuse to deploy, which version are you using?

> On Thu, Jul 7, 2016 at 4:11 PM, Ramesh Nachimuthu 
> wrote:
>>
>> Hi Dewey,
>>
>> Looks like SSH login from Engine VM to 'Host-01' is failing. Can you
>> confirm that "PermitRootLogin without-password" is enabled on 'Host-01'
>> where you are running " hosted-engine --deploy "
>>
>>
>> Regards,
>> Ramesh
>>
>> - Original Message -
>> > From: "Dewey Du" 
>> > To: "Scott" 
>> > Cc: "users" 
>> > Sent: Thursday, July 7, 2016 1:11:14 PM
>> > Subject: Re: [ovirt-users] Fail to setup Hyperconverged Infrastructure
>> > using oVirt and Gluster
>> >
>> > On Engine VM, the command " engine-setup " is executed successfully.
>> >
>> > On Host-01 which running the command " hosted-engine --deploy "
>> > always display the following message:
>> >
>> > Make a selection from the options below:
>> > (1) Continue setup - oVirt-Engine installation is ready and ovirt-engine
>> > service is up
>> > (2) Abort setup
>> > (3) Power off and restart the VM
>> > (4) Destroy VM and abort setup
>> > (1, 2, 3, 4)[1]: 1
>> >
>> > My engine.log on Engine VM is attached below
>> >
>> > 2016-07-05 15:55:54,086 ERROR
>> > [org.ovirt.engine.core.bll.hostdeploy.AddVdsCommand] (default task-17)
>> > [77e52cbc] Failed to authenticate session with host 'hosted_engine_1':
>> > SSH
>> > authentication to 'root@localhost.localdomain' failed. Please verify
>> > provided credentials. Make sure key is authorized at host
>> > 2016-07-05 15:55:54,086 WARN
>> > [org.ovirt.engine.core.bll.hostdeploy.AddVdsCommand] (default task-17)
>> > [77e52cbc] CanDoAction of action 'AddVds' failed for user
>> > admin@internal.
>> > Reasons: VAR__ACTION__ADD,VAR__TYPE__HOST,$server
>> > localhost.localdomain,VDS_CANNOT_AUTHENTICATE_TO_SERVER
>> > 2016-07-05 15:55:54,098 ERROR
>> > [org.ovirt.engine.api.restapi.resource.AbstractBackendResource] (default
>> > task-17) [] Operation Failed: [Cannot add Host. SSH authentication
>> > failed,
>> > verify authentication parameters are correct (Username/Password,
>> > public-key
>> > etc.) You may refer to the engine.log file for further details.]
>> >
>> > On Sun, Jul 3, 2016 at 11:27 PM, Scott < romra...@gmail.com > wrote:
>> >
>> >
>> >
>> > Do you have root logins disabled in SSH? If I remember right, oVirt will
>> > use
>> > SSH keys once configured so you need "PermitRootLogin without-password"
>> > at a
>> > minimum.
>> >
>> > The engine log and your auth/secure log on the host should probably give
>> > you
>> > some idea of what happened.
>> >
>> > Scott
>> >
>> > On Sun, Jul 3, 2016 at 10:13 AM Dewey Du < dewe...@gmail.com > wrote:
>> >
>> >
>> >
>> > oVirt 3.6
>> >
>> > # hosted-engine --deploy
>> >
>> > [ ERROR ] Cannot automatically add the host to cluster Default: Cannot
>> > add
>> > Host. SSH authentication failed, verify authentication parameters are
>> > correct (Username/Password, public-key etc.) You may refer to the
>> > engine.log
>> > file for further details.
>> >
>> > Has anyone encountered this issue before? Thx.
>> > ___
>> > Users mailing list
>> > Users@ovirt.org
>> > http://lists.ovirt.org/mailman/listinfo/users
>> >
>> >
>> > ___
>> > Users mailing list
>> > Users@ovirt.org
>> > http://lists.ovirt.org/mailman/listinfo/users
>> >
>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Fail to setup Hyperconverged Infrastructure using oVirt and Gluster

2016-07-08 Thread Simone Tiraboschi
On Fri, Jul 8, 2016 at 4:41 PM, Dewey Du <dewe...@gmail.com> wrote:
> on Host-01
>
> [root@localhost ~]# hostname
> localhost.localdomain

OK, the issue is here ^^^
Please use an uniquely resolvable hostname for your host.


> on Engine VM
>
> [root@engine ~]# hostname
> engine.ovirt.hscloud
>
> [root@engine ~]# engine-setup
>
>
>   --== CONFIGURATION PREVIEW ==--
>
>   Bridge interface   : em1
>   Engine FQDN: engine.ovirt.hscloud
>   Bridge name: ovirtmgmt
>   Host address   : localhost.localdomain
>   SSH daemon port: 22
>   Gateway address: 10.0.100.2
>   Host name for web application  : hosted_engine_1
>   Storage Domain type: glusterfs
>   Host ID: 1
>   Image size GB  : 10
>   GlusterFS Share Name   : hosted_engine_glusterfs
>   GlusterFS Brick Provisioning   : False
>   Storage connection : 10.0.100.17:/ovirt-engine
>   Console type   : vnc
>   Memory size MB : 16384
>   MAC address: 00:16:3e:1b:0f:e2
>   Boot type  : disk
>   Number of CPUs : 4
>   OVF archive (for disk boot):
> /usr/share/ovirt-engine-appliance/ovirt-engine-appliance-3.6-20160623.1.el7.centos.ova
>       CPU Type   : model_Haswell-noTSX
>
>
>
> On Fri, Jul 8, 2016 at 8:36 PM, Simone Tiraboschi <stira...@redhat.com>
> wrote:
>>
>> On Thu, Jul 7, 2016 at 5:44 PM, Dewey Du <dewe...@gmail.com> wrote:
>> >  "PermitRootLogin without-password" is disabled. Is this a MUST? I am
>> > following the steps from the link below, and there is no steps needed to
>> > enable "PermitRootLogin without-password".
>>
>> No, it's not.
>> The engine will connect using a key deployed on the host by
>> hosted-engine-setup.
>>
>> >
>> > http://blogs-ramesh.blogspot.jp/2016/01/ovirt-and-gluster-hyperconvergence.html
>> >
>> > Do you have anyother instructions to setup
>> > ovirt-and-gluster-hyperconvergence?
>> >
>> > By the way, the "localhost.localdomain"  in the engine.log "Failed to
>> > authenticate session with host 'hosted_engine_1': SSH authentication to
>> > 'root@localhost.localdomain' failed." means the engine VM or Host-01?
>>
>> It means the engine and the issue is just there.
>> Can you please check the hostname of your host?
>> We had a bug in past about that but now hosted-engine-setup should
>> refuse to deploy, which version are you using?
>>
>> > On Thu, Jul 7, 2016 at 4:11 PM, Ramesh Nachimuthu <rnach...@redhat.com>
>> > wrote:
>> >>
>> >> Hi Dewey,
>> >>
>> >> Looks like SSH login from Engine VM to 'Host-01' is failing. Can you
>> >> confirm that "PermitRootLogin without-password" is enabled on 'Host-01'
>> >> where you are running " hosted-engine --deploy "
>> >>
>> >>
>> >> Regards,
>> >> Ramesh
>> >>
>> >> - Original Message -
>> >> > From: "Dewey Du" <dewe...@gmail.com>
>> >> > To: "Scott" <romra...@gmail.com>
>> >> > Cc: "users" <users@ovirt.org>
>> >> > Sent: Thursday, July 7, 2016 1:11:14 PM
>> >> > Subject: Re: [ovirt-users] Fail to setup Hyperconverged
>> >> > Infrastructure
>> >> > using oVirt and Gluster
>> >> >
>> >> > On Engine VM, the command " engine-setup " is executed successfully.
>> >> >
>> >> > On Host-01 which running the command " hosted-engine --deploy "
>> >> > always display the following message:
>> >> >
>> >> > Make a selection from the options below:
>> >> > (1) Continue setup - oVirt-Engine installation is ready and
>> >> > ovirt-engine
>> >> > service is up
>> >> > (2) Abort setup
>> >> > (3) Power off and restart the VM
>> >> > (4) Destroy VM and abort setup
>> >> > (1, 2, 3, 4)[1]: 1
>> >> >
>> >> > My engine.log on Engine

Re: [ovirt-users] oVirt 4.0 hosted-engine deploy fail on fc domain

2016-07-08 Thread Simone Tiraboschi
On Fri, Jul 8, 2016 at 12:17 PM, NUNIN Roberto  wrote:
> Hello
>
> I’m in trouble deploying hosted-engine on a fresh-installed Centos7.2
> server, chosing fc domain:
>
>
>
> [ ERROR ] Failed to execute stage 'Environment customization': 'devList'
>
>
>
> Installation ends with error, following the versions of ovirt installed
> software:
>
>
>
> ovirt-hosted-engine-setup-2.0.0.2-1.el7.centos.noarch
>
> ovirt-engine-sdk-python-3.6.7.0-1.el7.centos.noarch
>
> ovirt-setup-lib-1.0.2-1.el7.centos.noarch
>
> ovirt-host-deploy-1.5.0-1.el7.centos.noarch
>
> ovirt-imageio-daemon-0.3.0-0.201606191345.git9f3d6d4.el7.centos.noarch
>
> ovirt-vmconsole-host-1.0.3-1.el7.centos.noarch
>
> ovirt-release40-4.0.0-5.noarch
>
> ovirt-vmconsole-1.0.3-1.el7.centos.noarch
>
> ovirt-hosted-engine-ha-2.0.0-1.el7.centos.noarch
>
> libgovirt-0.3.3-1.el7_2.1.x86_64
>
> ovirt-imageio-common-0.3.0-0.201606191345.git9f3d6d4.el7.centos.noarch
>
>
>
> Here part of the log:
>
>
>
> 2016-07-08 11:46:08 DEBUG otopi.plugins.otopi.dialog.human
> dialog.__logString:204 DIALOG:SEND Please specify the
> storage you would like to use (glusterfs, iscsi, fc, nfs3, nfs4)[nfs3]:
>
> 2016-07-08 11:46:11 DEBUG otopi.plugins.otopi.dialog.human
> dialog.__logString:204 DIALOG:RECEIVEfc
>
> 2016-07-08 11:46:11 DEBUG otopi.context context.dumpEnvironment:760
> ENVIRONMENT DUMP - BEGIN
>
> 2016-07-08 11:46:11 DEBUG otopi.context context.dumpEnvironment:770 ENV
> OVEHOSTED_STORAGE/domainType=str:'fc'
>
> 2016-07-08 11:46:11 DEBUG otopi.context context.dumpEnvironment:774
> ENVIRONMENT DUMP - END
>
> 2016-07-08 11:46:11 DEBUG otopi.context context._executeMethod:128 Stage
> customization METHOD
> otopi.plugins.gr_he_setup.storage.blockd.Plugin._customization
>
> 2016-07-08 11:46:11 DEBUG otopi.plugins.gr_he_setup.storage.blockd
> blockd._fc_get_lun_list:404 {'status': {'message': 'Done', 'code': 0},
> 'items': [{u'status': u'free', u'vendorID': u'3PARdata', u'capacity':
> u'1099511627776', u'fwrev': u'3122', u'vgUUID': u'', u'pvsize': u'',
> u'pathlist': [], u'logicalblocksize': u'512', u'pathstatus': [{u'capacity':
> u'1099511627776', u'physdev': u'sdb', u'type': u'FCP', u'state': u'active',
> u'lun': u'1'}, {u'capacity': u'1099511627776', u'physdev': u'sdc', u'type':
> u'FCP', u'state': u'active', u'lun': u'1'}, {u'capacity': u'1099511627776',
> u'physdev': u'sdd', u'type': u'FCP', u'state': u'active', u'lun': u'1'},
> {u'capacity': u'1099511627776', u'physdev': u'sde', u'type': u'FCP',
> u'state': u'active', u'lun': u'1'}, {u'capacity': u'1099511627776',
> u'physdev': u'sdf', u'type': u'FCP', u'state': u'active', u'lun': u'1'},
> {u'capacity': u'1099511627776', u'physdev': u'sdg', u'type': u'FCP',
> u'state': u'active', u'lun': u'1'}, {u'capacity': u'1099511627776',
> u'physdev': u'sdh', u'type': u'FCP', u'state': u'active', u'lun': u'1'},
> {u'capacity': u'1099511627776', u'physdev': u'sdi', u'type': u'FCP',
> u'state': u'active', u'lun': u'1'}], u'devtype': u'FCP',
> u'physicalblocksize': u'512', u'pvUUID': u'', u'serial':
> u'S3PARdataVV_1619775', u'GUID': u'360002ac01d0060964d3f',
> u'productID': u'VV'}]}
>
> 2016-07-08 11:46:11 DEBUG otopi.context context._executeMethod:142 method
> exception
>
> Traceback (most recent call last):
>
>   File "/usr/lib/python2.7/site-packages/otopi/context.py", line 132, in
> _executeMethod
>
> method['method']()
>
>   File
> "/usr/share/ovirt-hosted-engine-setup/scripts/../plugins/gr-he-setup/storage/blockd.py",
> line 612, in _customization
>
> lunGUID = self._customize_lun(self.domainType, target)
>
>   File
> "/usr/share/ovirt-hosted-engine-setup/scripts/../plugins/gr-he-setup/storage/blockd.py",
> line 212, in _customize_lun
>
> available_luns = self._fc_get_lun_list()
>
>   File
> "/usr/share/ovirt-hosted-engine-setup/scripts/../plugins/gr-he-setup/storage/blockd.py",
> line 407, in _fc_get_lun_list
>
> for device in devices['devList']:
>
> KeyError: 'devList'
>
> ERROR otopi.context context._executeMethod:151 Failed to execute stage
> 'Environment customization': 'devList'
>
> ….
>
> FC volume is hosted on an HP 3PAR storage array.
>
>
>
> It is available, under multipathd, to the OS:
>
>
>
> [root@xxx-yyy-xxx ~]# multipath -l
>
> 360002ac01d0060964d3f dm-2 3PARdata,VV
>
> size=1.0T features='1 queue_if_no_path' hwhandler='0' wp=rw
>
> `-+- policy='round-robin 0' prio=0 status=active
>
>   |- 2:0:11:1 sdi 8:128 active undef running
>
>   |- 1:0:9:1  sdd 8:48  active undef running
>
>   |- 2:0:7:1  sdg 8:96  active undef running
>
>   |- 1:0:1:1  sdc 8:32  active undef running
>
>   |- 2:0:6:1  sdf 8:80  active undef running
>
>   |- 1:0:0:1  sdb 8:16  active undef running
>
>   |- 2:0:8:1  sdh 8:112 active undef running
>
>   `- 1:0:10:1 sde 8:64  active undef running
>
> [root@xxx-yyy-xxx ~]#
>
>
>
> Must I submit other relevant logs ? Which one ?
>
> Thanks in advance.

Hi Roberto,
we already opened a bug on it:

Re: [ovirt-users] [hosted-engine] engine VM didn't migrate to the fresh upgraded host

2016-07-07 Thread Simone Tiraboschi
On Thu, Jul 7, 2016 at 7:17 AM, Wee Sritippho  wrote:
> Hi, I followed this instruction:
>
> http://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine
>
> However, when I exited the global maintenance mode in step 7 and waited for
> about 15 minutes, the engine VM still doesn't migrate to the fresh upgraded
> host.

In you case it didn't migrated since both host-2 and host-3 were
already at 3400 points and so there wasn't any reason to migrate.

> BTW, after step 6, did I have to put the host out of its local maintenance
> mode? The instruction didn't state this so I guess it's a special case when
> upgrading and didn't do anything.

hosted-engine --set-maintenance --mode=none will also exit local
maintenance mode.

> [root@host01 me]# hosted-engine --vm-status
>
>
> --== Host 1 status ==--
>
> Status up-to-date  : True
> Hostname   : host01.ovirt.forest.go.th
> Host ID: 1
> Engine status  : {"reason": "vm not running on this
> host", "health": "bad", "vm": "down", "detail": "unknown"}
> Score  : 0
> stopped: False
> Local maintenance  : True
> crc32  : 33cc9d8c
> Host timestamp : 4993624
>
>
> --== Host 2 status ==--
>
> Status up-to-date  : True
> Hostname   : host02.ovirt.forest.go.th
> Host ID: 2
> Engine status  : {"reason": "vm not running on this
> host", "health": "bad", "vm": "down", "detail": "unknown"}
> Score  : 3400
> stopped: False
> Local maintenance  : False
> crc32  : 6dc9b311
> Host timestamp : 4244063
>
>
> --== Host 3 status ==--
>
> Status up-to-date  : True
> Hostname   : host03.ovirt.forest.go.th
> Host ID: 3
> Engine status  : {"health": "good", "vm": "up",
> "detail": "up"}
> Score  : 3400
> stopped: False
> Local maintenance  : False
> crc32  : 29513baf
> Host timestamp : 5537027
>
> Thank you
>
> --
> Wee
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] No hosted Engine VM in portal after migrating

2016-07-07 Thread Simone Tiraboschi
On Thu, Jul 7, 2016 at 1:34 PM, Matt .  wrote:
> Hi,
>
> I did a successfull migration to hosted engine and noticed there is no
> HostedEngine VM in the portal.

Do you just miss the hosted-engine VM or also the whole hosted-engine
storage domain?

> In earlier tests it showed up so I wonder if I can import it in some
> way so it exists in the portal.
>
> I used this migration way which earlier added the hosted engine and
> stayed in the portal also after restoring the old backup.
>
> http://www.ovirt.org/develop/developer-guide/engine/migrate-to-hosted-engine/
>
> Thanks,
>
> Matt
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] No hosted Engine VM in portal after migrating

2016-07-08 Thread Simone Tiraboschi
On Fri, Jul 8, 2016 at 9:25 AM, Matt . <yamakasi@gmail.com> wrote:
> Hi,
>
> I see on the engine gthe following message:
>
> 2016-07-08 09:24:24,986 INFO
> [org.ovirt.engine.core.bll.exportimport.ImportVmCommand]
> (org.ovirt.thread.pool-8-thread-5) [49c1c8c1] Lock Acquired to object
> 'EngineLock:{exclusiveLocks='[5576ec24-112e-4995-89f8-57e40c43cc5a=<VM,
> ACTION_TYPE_FAILED_VM_IS_BEING_IMPORTED$VmName HostedEngine>,
> HostedEngine=<VM_NAME, ACTION_TYPE_FAILED_NAME_ALREADY_USED>]',
> sharedLocks='[5576ec24-112e-4995-89f8-57e40c43cc5a=<REMOTE_VM,
> ACTION_TYPE_FAILED_VM_IS_BEING_IMPORTED$VmName HostedEngine>]'}'
>
> The issue is that there is no entry as VM in the database.

Adding Roy here.

> What should I check more ?
>
> Thanks,
>
> Matt
>
> 2016-07-07 23:26 GMT+02:00 Matt . <yamakasi@gmail.com>:
>> The storage domain is there, the host-engine is just not being added,
>> also not after some time waiting.
>>
>> 2016-07-07 14:15 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>> On Thu, Jul 7, 2016 at 1:34 PM, Matt . <yamakasi@gmail.com> wrote:
>>>> Hi,
>>>>
>>>> I did a successfull migration to hosted engine and noticed there is no
>>>> HostedEngine VM in the portal.
>>>
>>> Do you just miss the hosted-engine VM or also the whole hosted-engine
>>> storage domain?
>>>
>>>> In earlier tests it showed up so I wonder if I can import it in some
>>>> way so it exists in the portal.
>>>>
>>>> I used this migration way which earlier added the hosted engine and
>>>> stayed in the portal also after restoring the old backup.
>>>>
>>>> http://www.ovirt.org/develop/developer-guide/engine/migrate-to-hosted-engine/
>>>>
>>>> Thanks,
>>>>
>>>> Matt
>>>> ___
>>>> Users mailing list
>>>> Users@ovirt.org
>>>> http://lists.ovirt.org/mailman/listinfo/users
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Cannot add Hosted Engine Host because of missing answersfile

2016-07-04 Thread Simone Tiraboschi
On Mon, Jul 4, 2016 at 3:54 PM, Matt . <yamakasi@gmail.com> wrote:
> Indeed, that is what I thought, but does this mean I need to reinstall
> that whole VM ?

Yes, unfortunately the best option, since you installation never
completed, is to clean the hosted-engine storage domain and restart
from scratch.
If you use the engine appliance (yum install ovirt-engine-appliance)
the whole installation will take about 15/20 minutes (excluding
download time).


> And what about the hosted_storare and such ?
>
> 2016-07-04 12:36 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>> On Mon, Jul 4, 2016 at 10:48 AM, Matt . <yamakasi@gmail.com> wrote:
>>> Hi,
>>>
>>> Yes, the issue is that I don't have that info on the shared storage
>>> indeed. I was wondering if I just can create it manually.
>>
>> It's not that easy and not that safe.
>> It will be much easier and quicker to simply redeploy the first host.
>>
>>> I would like to add some to the documentation there if it's possible.
>>>
>>> Logs will be posted :)
>>>
>>> 2016-07-04 10:44 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>>> On Mon, Jul 4, 2016 at 10:25 AM, Matt . <yamakasi@gmail.com> wrote:
>>>>> Hi Sure I can, I will attach them today.
>>>>>
>>>>> The install failed at the end as the installer/deploy didn't like my
>>>>> hostfile on the vdsm host where I was doing the install. The install
>>>>> of the engine went fine.
>>>>
>>>> Do you mean it failed on the first host?
>>>> The configuration is written on the shared storage if and only if the
>>>> first host successfully deployed.
>>>> If you have an issue on the first host you cannot simply add other
>>>> hosts and you have to fix there first.
>>>>
>>>>> I don't think it's a bug, it could be tackled when such thing happens
>>>>> but still, it was a config issue what actually shouldn't be an issue
>>>>> to let the full install succeed.
>>>>>
>>>>> Do you upfront have an idea if I need to do a full new deploy ?
>>>>>
>>>>> I will grab the logs for you later on.
>>>>
>>>> Thanks
>>>>
>>>>> Do you want them in a bugreport ?
>>>>
>>>> Let's see the logs before
>>>>
>>>>> Thanks,
>>>>>
>>>>> Matt
>>>>>
>>>>> 2016-07-04 10:03 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>>>>> On Mon, Jul 4, 2016 at 6:31 AM, Matt . <yamakasi@gmail.com> wrote:
>>>>>>> Hi!
>>>>>>>
>>>>>>> Apperently there has been something gone wrong with the hosted engine
>>>>>>> install which I can perfectly start but cannot add any Host to the HA
>>>>>>> setup.
>>>>>>>
>>>>>>> I'm missing an answers.conf on my host where I did the installation.
>>>>>>>
>>>>>>> Has anyone an idea if I can fix this without a new full deploy ?
>>>>>>>
>>>>>>> I'm on 3.6.7
>>>>>>
>>>>>> Can you please report the version of ovirt-hosted-engine-ha and
>>>>>> ovirt-hosted-engine-setup from your working hosts?
>>>>>> Can you please attach hosted-engine-setup logs from your first host?
>>>>>>
>>>>>>> Some errorlog:
>>>>>>>
>>>>>>>
>>>>>>>   --== STORAGE CONFIGURATION ==--
>>>>>>>
>>>>>>>   During customization use CTRL-D to abort.
>>>>>>>   Please specify the storage you would like to use (glusterfs,
>>>>>>> iscsi, fc, nfs3, nfs4)[nfs3]:
>>>>>>>   Please specify the full shared storage connection path to
>>>>>>> use (example: host:/path): filer-01.my-domain.tld:/vol/vol0/oVIRT_HE
>>>>>>>   The specified storage location already contains a data
>>>>>>> domain. Is this an additional host setup (Yes, No)[Yes]?
>>>>>>> [ INFO  ] Installing on additional host
>>>>>>>   Please specify the Host ID [Must be integer, default: 2]:
>>>>>>>
>>>>>>>   --== SYSTEM CONFIGURATION ==--
>>>>>>>
>>>>>>

Re: [ovirt-users] 3.6 : Hosted_Storage unattached

2016-08-09 Thread Simone Tiraboschi
On Tue, Aug 9, 2016 at 2:14 PM, Alexis HAUSER
 wrote:
> Actually, I solved my problem by solving a bug I was affected by (SELinux 
> preventing to add storage domain), removing the unattached hosted_storage and 
> restarting ovirt-engine.
> It added it automatically with the VM as usual :)

Do you know if there is already an open ticket for that? if not,
opening one on https://bugzilla.redhat.com/ will help to get it fixed.
Thanks.

> - Mail original -
> De: "Alexis HAUSER" 
> À: "users" 
> Envoyé: Mardi 9 Août 2016 11:50:48
> Objet: [ovirt-users] 3.6 : Hosted_Storage unattached
>
> Hi, I installed a new node with a new hosted engine, version 3.6, added a 
> data domain, but I can't see the hosted_storage.
>
> I tried to use the "import storage" on it but it keeps having unattached 
> status, and in the logs I can see :
>
> "2016-08-09 05:39:32,821 WARN  
> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
> (ajp-/127.0.0.1:8702-1) [] Correlation ID: 1d9c8f67, Job ID: 
> 6e75d61e-3140-4f1c-a301-e683ddc28b1f, Call Stack: null, Custom Event ID: -1, 
> Message: This Data center compatibility version does not support importing a 
> data domain with its entities (VMs and Templates). The imported domain will 
> be imported without them."
>
> Any ideas ?
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4.0 Hosted Engine

2016-08-08 Thread Simone Tiraboschi
On Sun, Aug 7, 2016 at 3:42 PM, Arsène Gschwind
 wrote:
> Hi,
>
> I have an oVirt setup with 2 server using hosted-engine, both server
> resgistered properly the hosted-engine using :
> # hosted-engine --deploy
>
> but for some reason the second isn't recognized as a host for hosted-engine
> and I'm not able to migrate the hosted-engine.
> The error I get when trying to migrate:
>
> Cannot migrate VM. There is no host that satisfies current scheduling
> constraints. See below for details:
> The host xx did not satisfy internal filter HA because it is not a
> Hosted Engine host..
>
> I've tried to redeploy the hosted-engine but this will fail since the host
> already exists in the management DB.
> I've tried to redeploy the host using the GUI when editing the Host and set
> DEPLOY at Hosted Engine but in that case the Event just says the
> configuration was updated but nothing happens.
>
> Is there a way to check if the host is registered as hosted-engine host?

Can you please execute hosted-engine --vm-status on the problematic host?

Do you see any hosted-engine score for it in the engine?

> How could I register it correctly?
>
> Let me know if you need any logs.
> Thanks for any hint.
>
> Regards,
> Arsène
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4.0 Hosted Engine

2016-08-08 Thread Simone Tiraboschi
On Mon, Aug 8, 2016 at 10:54 AM, Arsène Gschwind
 wrote:
> Status up-to-date  : False
> Hostname   : 
> Host ID: 1

The issue is that on both the host you just see the status of host 1,
which is also not up-to-date at host 2 eyes, and you miss host 2.

Can you please check the status of ovirt-ha-agent and ovirt-ha-broker
services and share your /var/log/ovirt-hosted-engine-ha/agent.log from
the problematic host?
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] upgrade hosted-engine os ( not hosts )

2016-06-30 Thread Simone Tiraboschi
On Thu, Jun 30, 2016 at 8:34 AM, Paul Groeneweg | Pazion  wrote:
> Hi Yedidyah,
>
> Thank you for the comprehensive answers.
>
> I think I go for a complete reinstall ( read also OS upgrade tool is not
> adviced on 6.6 or higher as there might be newer packages as on 7 ). No
> doubting to re-use current VM or setup from scratch ( fresh host with new
> hosted-engine and existing storage domein ).

If you are planning to redeploy hosted-engien from scratch and
restoring on the new engine DB a backup of the previous one, please
carefully consider this:
https://bugzilla.redhat.com/show_bug.cgi?id=1240466#c21

So, if you choose that path you'll have also to manually remove any
reference to the previous hosted-engine from the restored DB.
I'd strongly suggest you to wait for the upgrade tool to be fixed
since manually doing this upgrade can be really error prone.

> You explain the steps ( 1 to 6 ), but then don't talk about storage domain
> import.
> Does it mean, when I reinstall the hosted-engine in the current he VM and
> restore an engine-backup ( step 5 ) I am able to start vm from Host and it
> is still connected to the master storage ( so no need for storage import) ?
>
> Best Regards,
> Paul Groeneweg
>
>
> Op do 30 jun. 2016 om 08:00 schreef Yedidyah Bar David :
>>
>> On Wed, Jun 29, 2016 at 10:07 PM, Paul Groeneweg | Pazion
>>  wrote:
>> >
>> > I am looking for a way to get my hosted-engine running on el7 so I can
>> > upgrade to oVirt 4.0. Currently my hosts already run el7, but my
>> > hosted-engine is still el6.
>> >
>> > I read
>> >
>> > https://www.ovirt.org/documentation/how-to/hosted-engine-host-OS-upgrade/
>> > but this is only about the hosts.
>> >
>> > I read https://www.ovirt.org/documentation/how-to/hosted-engine/, but it
>> > only mentions upgrade of the hosted-engine software, not the OS.
>> >
>> > I understood I can do a fresh hosted-engine install, and then import my
>> > storage domain to the new hosted engine, but:
>> >
>> > - Do I need to restore my hosted engine database? ( like described here:
>> >
>> > http://www.ovirt.org/develop/developer-guide/engine/migrate-to-hosted-engine/
>> > )
>>
>> You might not have to, if you only care about the imported VMs from your
>> storage. This will not keep other configuration, such as
>> users/roles/permissions
>> etc.
>>
>> > - Can I directly install hosted-engine 4.0 and then import the storage
>> > domain? Or should I install same hosted-engine version?
>>
>> AFAIK 4.0 engine can import 3.6 storage domains without problem.
>>
>> > - Do I first need another master storage domain or can I directly import
>> > my
>> > old master storage domain?
>>
>> No idea. Even if you do, you can create a small empty one and later remove
>> it.
>>
>> > - When importing the storage domain what is the risk it fails ( I have
>> > backups, but it would cost a day to restore all )
>>
>> No idea, but IIRC we got many successful reports and at most few failures
>> for this.
>>
>> > - How long would import take? few minutes or hours? ( I want to keep
>> > down
>> > time as low as possible ).
>>
>> Again no idea. Perhaps do some test?
>>
>> >
>> > Another option would be upgrade the OS ( with redhat-upgrade-tool ) or
>> > is
>> > this a path for disaster?
>>
>> Didn't work for us well, so we decided to not support it. If you decide to
>> try,
>> make sure you test carefully beforehand. From ovirt's POV:
>> 1. You'll need to handle postgresql upgrade.
>> 2. Right after OS upgrade, you'll still have (I think) el6 packages
>> of the engine. It will hopefully be in a good-enough state for upgrade
>> to 4.0, but we didn't test this.
>> 3. Specifically, if upgrade fails, rollback will most likely not work,
>> so you'll have to manually handle this - take a full vm backup and make
>> sure you can restore it.
>>
>> >
>> > I hope someone can tell me how I can smoothly upgrade my hosted-engine
>> > up to
>> > el7 and run oVirt 4.
>>
>> We are working on a tool/wizard to help with this process. It used to
>> work,
>> but at some point it was decided that one of the actions it does is risky
>> and was blocked, thus the tool is broken currently.
>>
>> You can invoke the tool by running: 'hosted-engine --upgrade-appliance'.
>> As noted above, this is currently broken.
>>
>> There are several open bugs about it, e.g.:
>>
>> https://bugzilla.redhat.com/show_bug.cgi?id=1319457
>> https://bugzilla.redhat.com/show_bug.cgi?id=1343425
>> https://bugzilla.redhat.com/show_bug.cgi?id=1343593 (closed, this is
>> what broke the tool)
>>
>> Basically, you can manually do what the tool is supposed to do:
>> 1. Make sure state is clean and stable (no running/pending storage
>> actions,
>> no VMs in the middle of migration etc), all clusters are compat level 3.6,
>> etc.
>> 2. Move to global maintenance
>> 3. backup the engine using engine-backup and keep the backup elsewhere
>> 4. Reinstall engine vm with el7 and 4.0 engine (the tool will use 

Re: [ovirt-users] Cannot add Hosted Engine Host because of missing answersfile

2016-07-04 Thread Simone Tiraboschi
On Mon, Jul 4, 2016 at 10:48 AM, Matt . <yamakasi@gmail.com> wrote:
> Hi,
>
> Yes, the issue is that I don't have that info on the shared storage
> indeed. I was wondering if I just can create it manually.

It's not that easy and not that safe.
It will be much easier and quicker to simply redeploy the first host.

> I would like to add some to the documentation there if it's possible.
>
> Logs will be posted :)
>
> 2016-07-04 10:44 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>> On Mon, Jul 4, 2016 at 10:25 AM, Matt . <yamakasi@gmail.com> wrote:
>>> Hi Sure I can, I will attach them today.
>>>
>>> The install failed at the end as the installer/deploy didn't like my
>>> hostfile on the vdsm host where I was doing the install. The install
>>> of the engine went fine.
>>
>> Do you mean it failed on the first host?
>> The configuration is written on the shared storage if and only if the
>> first host successfully deployed.
>> If you have an issue on the first host you cannot simply add other
>> hosts and you have to fix there first.
>>
>>> I don't think it's a bug, it could be tackled when such thing happens
>>> but still, it was a config issue what actually shouldn't be an issue
>>> to let the full install succeed.
>>>
>>> Do you upfront have an idea if I need to do a full new deploy ?
>>>
>>> I will grab the logs for you later on.
>>
>> Thanks
>>
>>> Do you want them in a bugreport ?
>>
>> Let's see the logs before
>>
>>> Thanks,
>>>
>>> Matt
>>>
>>> 2016-07-04 10:03 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>>>> On Mon, Jul 4, 2016 at 6:31 AM, Matt . <yamakasi@gmail.com> wrote:
>>>>> Hi!
>>>>>
>>>>> Apperently there has been something gone wrong with the hosted engine
>>>>> install which I can perfectly start but cannot add any Host to the HA
>>>>> setup.
>>>>>
>>>>> I'm missing an answers.conf on my host where I did the installation.
>>>>>
>>>>> Has anyone an idea if I can fix this without a new full deploy ?
>>>>>
>>>>> I'm on 3.6.7
>>>>
>>>> Can you please report the version of ovirt-hosted-engine-ha and
>>>> ovirt-hosted-engine-setup from your working hosts?
>>>> Can you please attach hosted-engine-setup logs from your first host?
>>>>
>>>>> Some errorlog:
>>>>>
>>>>>
>>>>>   --== STORAGE CONFIGURATION ==--
>>>>>
>>>>>   During customization use CTRL-D to abort.
>>>>>   Please specify the storage you would like to use (glusterfs,
>>>>> iscsi, fc, nfs3, nfs4)[nfs3]:
>>>>>   Please specify the full shared storage connection path to
>>>>> use (example: host:/path): filer-01.my-domain.tld:/vol/vol0/oVIRT_HE
>>>>>   The specified storage location already contains a data
>>>>> domain. Is this an additional host setup (Yes, No)[Yes]?
>>>>> [ INFO  ] Installing on additional host
>>>>>   Please specify the Host ID [Must be integer, default: 2]:
>>>>>
>>>>>   --== SYSTEM CONFIGURATION ==--
>>>>>
>>>>> [WARNING] A configuration file must be supplied to deploy Hosted
>>>>> Engine on an additional host.
>>>>> [ ERROR ] 'version' is not stored in the HE configuration image
>>>>> [ ERROR ] Unable to get the answer file from the shared storage
>>>>> [ ERROR ] Failed to execute stage 'Environment customization': Unable
>>>>> to get the answer file from the shared storage
>>>>> [ INFO  ] Stage: Clean up
>>>>> [ INFO  ] Generating answer file
>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160704062340.conf'
>>>>> [ INFO  ] Stage: Pre-termination
>>>>> [ INFO  ] Stage: Termination
>>>>> [ ERROR ] Hosted Engine deployment failed: this system is not
>>>>> reliable, please check the issue, fix and redeploy
>>>>>   Log file is located at
>>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160704062311-eqejnq.log
>>>>>
>>>>> I hope we can fix this.
>>>>> ___
>>>>> Users mailing list
>>>>> Users@ovirt.org
>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Cannot add Hosted Engine Host because of missing answersfile

2016-07-04 Thread Simone Tiraboschi
On Mon, Jul 4, 2016 at 6:31 AM, Matt .  wrote:
> Hi!
>
> Apperently there has been something gone wrong with the hosted engine
> install which I can perfectly start but cannot add any Host to the HA
> setup.
>
> I'm missing an answers.conf on my host where I did the installation.
>
> Has anyone an idea if I can fix this without a new full deploy ?
>
> I'm on 3.6.7

Can you please report the version of ovirt-hosted-engine-ha and
ovirt-hosted-engine-setup from your working hosts?
Can you please attach hosted-engine-setup logs from your first host?

> Some errorlog:
>
>
>   --== STORAGE CONFIGURATION ==--
>
>   During customization use CTRL-D to abort.
>   Please specify the storage you would like to use (glusterfs,
> iscsi, fc, nfs3, nfs4)[nfs3]:
>   Please specify the full shared storage connection path to
> use (example: host:/path): filer-01.my-domain.tld:/vol/vol0/oVIRT_HE
>   The specified storage location already contains a data
> domain. Is this an additional host setup (Yes, No)[Yes]?
> [ INFO  ] Installing on additional host
>   Please specify the Host ID [Must be integer, default: 2]:
>
>   --== SYSTEM CONFIGURATION ==--
>
> [WARNING] A configuration file must be supplied to deploy Hosted
> Engine on an additional host.
> [ ERROR ] 'version' is not stored in the HE configuration image
> [ ERROR ] Unable to get the answer file from the shared storage
> [ ERROR ] Failed to execute stage 'Environment customization': Unable
> to get the answer file from the shared storage
> [ INFO  ] Stage: Clean up
> [ INFO  ] Generating answer file
> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160704062340.conf'
> [ INFO  ] Stage: Pre-termination
> [ INFO  ] Stage: Termination
> [ ERROR ] Hosted Engine deployment failed: this system is not
> reliable, please check the issue, fix and redeploy
>   Log file is located at
> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160704062311-eqejnq.log
>
> I hope we can fix this.
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Unable to set up oVirt 4.0 HE using glusterfs storage

2016-07-04 Thread Simone Tiraboschi
On Sun, Jul 3, 2016 at 5:57 AM, Kevin Hung  wrote:
> Looks like there still needs to be some work done on oVirt 4.0 Node and
> ovirt-hosted-engine-setup before it's ready for general consumption. I have
> spent days trying to get this to work, and only got it running (on one host)
> after encountering 8 serious issues (7 below and the initial glusterfs one).
> I have not been able to successfully deploy a second host (see issue 7
> below). I will be moving back to deploying hosts using CentOS (with either
> oVirt 4.0 or oVirt 3.6) as I need a working oVirt deployment up and running.
>
> In case anyone is interested in reproducing the issues, I used the Node ISO
> here [1] and the latest (7/2/2016) engine appliance OVA here [2]. Those seem
> to be the "official" files as far as I can tell (which is difficult as the
> documentation is not clear).
>
> List of issues:
> 1. The error I mentioned seems to be an problem with the code. I bypassed it
> by deleting /usr/libexec/vdsm/hooks/before_network_setup/50_fcoe.
> 2. ovirt-hosted-engine-setup is unable to connect to the vdsm service if the
> FQDN of the node is not resolvable (i.e. if a DNS server is not entered in
> the initial setup). This should be checked in either the initial oVirt Node
> setup process or the beginning of ovirt-hosted-engine-setup.
> 3. The management bridge does not get created properly when the server is
> set up with a manually configured DNS server and running NetworkManager (the
> default on Node). It seems like a bug has been filed for this back in 2014.
> [3]
> 4. Using cloud-init with default values to customize the engine appliance
> can fail on the line "Creating/refreshing DWH database schema" if it takes
> longer than 600 seconds to return output. This may apply to any other step
> that takes a long time to complete. The VM no longer appears to be exist
> after the setup exits that so I am unable to debug.

600 seconds seams more than a reasonable time to create an empty DB,
if it requires more than 10 minutes for a simple/short operation there
is probably something strange with the storage.

> 5. Without using cloud-init, the setup creates an engine VM that I cannot
> log into (it does not seem to use the engine admin password or a blank
> password).

Yes, the engine VM host-name and its root password are configured via
cloud-init and there is not default password.
If you want to avoid using cloud-init you have to reset the root
password of the engine VM as for any el7 machine.

> 6. Destroying the VM (option 4) leaves the files intact on the shared
> storage so I cannot restart setup without deleting those first. This may be
> intentional, but the use of kvm terminology (destroy for power off) is not
> common, not to mention that "virsh -r list --all" does not list the VM
> anymore.

On failures, there is not just the engine VM disk but a whole storage
domain for hosted-engine which also contains ancillary disks.
Re-deploying over a dirty storage is not supported so please clean up
the whole storage domain on failures.

> 7. Unable to deploy second host through web UI (error "Failed to configure
> management network on host node2 due to setup networks failure.") or using

This is not hosted-engine specific:
https://bugzilla.redhat.com/show_bug.cgi?id=1350763

> ovirt-hosted-engine-setup (it looks like it can't connect to or doesn't
> start the broker service).
> 8. Random errors to stderr: "vcpu0 unhandled rdmsr" (this seems to be an

Are you running in a nested env?

> upstream bug) and "multipath: error getting device" (this has been an issue
> for years with oVirt and seems to be due to multipathing being on by default
> even for systems where that does not apply).
>
> [1]
> http://resources.ovirt.org/pub/ovirt-4.0/iso/ovirt-node-ng-installer/ovirt-node-ng-installer-ovirt-4.0-2016062412.iso
> [2]
> http://jenkins.ovirt.org/view/All/job/ovirt-appliance_ovirt-4.0_build-artifacts-el7-x86_64/
> [3] https://bugzilla.redhat.com/show_bug.cgi?id=1160423
>
>
> On 7/1/2016 8:37 PM, Kevin Hung wrote:
>>
>> It looks like I'm now getting an error when the deployment tries to
>> configure the management bridge.
>>
>> Setup log:
>>
>> 2016-07-01 20:29:47 INFO otopi.plugins.gr_he_common.network.bridge
>> bridge._misc:
>> 372 Configuring the management bridge
>> 2016-07-01 20:29:48 DEBUG otopi.plugins.gr_he_common.network.bridge
>> bridge._misc
>> :384 networks: {'ovirtmgmt': {'nic': 'eno1', 'ipaddr': u'192.168.1.211',
>> 'netmask': u'255.255.255.0', 'bootproto': u'none', 'gateway':
>> u'192.168.1.1', 'defaultRoute': True}}
>> 2016-07-01 20:29:48 DEBUG otopi.plugins.gr_he_common.network.bridge
>> bridge._misc
>> :385 bonds: {}
>> 2016-07-01 20:29:48 DEBUG otopi.plugins.gr_he_common.network.bridge
>> bridge._misc
>> :386 options: {'connectivityCheck': False}
>> 2016-07-01 20:29:48 DEBUG otopi.context context._executeMethod:142 method
>> exception
>> Traceback (most recent call last):
>>   File 

Re: [ovirt-users] Cannot add Hosted Engine Host because of missing answersfile

2016-07-04 Thread Simone Tiraboschi
On Mon, Jul 4, 2016 at 10:25 AM, Matt . <yamakasi@gmail.com> wrote:
> Hi Sure I can, I will attach them today.
>
> The install failed at the end as the installer/deploy didn't like my
> hostfile on the vdsm host where I was doing the install. The install
> of the engine went fine.

Do you mean it failed on the first host?
The configuration is written on the shared storage if and only if the
first host successfully deployed.
If you have an issue on the first host you cannot simply add other
hosts and you have to fix there first.

> I don't think it's a bug, it could be tackled when such thing happens
> but still, it was a config issue what actually shouldn't be an issue
> to let the full install succeed.
>
> Do you upfront have an idea if I need to do a full new deploy ?
>
> I will grab the logs for you later on.

Thanks

> Do you want them in a bugreport ?

Let's see the logs before

> Thanks,
>
> Matt
>
> 2016-07-04 10:03 GMT+02:00 Simone Tiraboschi <stira...@redhat.com>:
>> On Mon, Jul 4, 2016 at 6:31 AM, Matt . <yamakasi@gmail.com> wrote:
>>> Hi!
>>>
>>> Apperently there has been something gone wrong with the hosted engine
>>> install which I can perfectly start but cannot add any Host to the HA
>>> setup.
>>>
>>> I'm missing an answers.conf on my host where I did the installation.
>>>
>>> Has anyone an idea if I can fix this without a new full deploy ?
>>>
>>> I'm on 3.6.7
>>
>> Can you please report the version of ovirt-hosted-engine-ha and
>> ovirt-hosted-engine-setup from your working hosts?
>> Can you please attach hosted-engine-setup logs from your first host?
>>
>>> Some errorlog:
>>>
>>>
>>>   --== STORAGE CONFIGURATION ==--
>>>
>>>   During customization use CTRL-D to abort.
>>>   Please specify the storage you would like to use (glusterfs,
>>> iscsi, fc, nfs3, nfs4)[nfs3]:
>>>   Please specify the full shared storage connection path to
>>> use (example: host:/path): filer-01.my-domain.tld:/vol/vol0/oVIRT_HE
>>>   The specified storage location already contains a data
>>> domain. Is this an additional host setup (Yes, No)[Yes]?
>>> [ INFO  ] Installing on additional host
>>>   Please specify the Host ID [Must be integer, default: 2]:
>>>
>>>   --== SYSTEM CONFIGURATION ==--
>>>
>>> [WARNING] A configuration file must be supplied to deploy Hosted
>>> Engine on an additional host.
>>> [ ERROR ] 'version' is not stored in the HE configuration image
>>> [ ERROR ] Unable to get the answer file from the shared storage
>>> [ ERROR ] Failed to execute stage 'Environment customization': Unable
>>> to get the answer file from the shared storage
>>> [ INFO  ] Stage: Clean up
>>> [ INFO  ] Generating answer file
>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160704062340.conf'
>>> [ INFO  ] Stage: Pre-termination
>>> [ INFO  ] Stage: Termination
>>> [ ERROR ] Hosted Engine deployment failed: this system is not
>>> reliable, please check the issue, fix and redeploy
>>>   Log file is located at
>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160704062311-eqejnq.log
>>>
>>> I hope we can fix this.
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [Call for feedback] did you install/update to 4.1.0?

2017-02-03 Thread Simone Tiraboschi
On Fri, Feb 3, 2017 at 7:20 PM, Simone Tiraboschi <stira...@redhat.com>
wrote:

>
>
> On Fri, Feb 3, 2017 at 5:22 PM, Ralf Schenk <r...@databay.de> wrote:
>
>> Hello,
>>
>> of course:
>>
>> [root@microcloud27 mnt]# sanlock client status
>> daemon 8a93c9ea-e242-408c-a63d-a9356bb22df5.microcloud
>> p -1 helper
>> p -1 listener
>> p -1 status
>>
>> sanlock.log attached. (Beginning 2017-01-27 where everything was fine)
>>
> Thanks, the issue is here:
>
> 2017-02-02 19:01:22+0100 4848 [1048]: s36 lockspace 
> 7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96:3:/rhev/data-center/mnt/glusterSD/glusterfs.rxmgmt.databay.de:_engine/7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96/dom_md/ids:0
> 2017-02-02 19:03:42+0100 4988 [12983]: s36 delta_acquire host_id 3 busy1 3 15 
> 13129 7ad427b1-fbb6-4cee-b9ee-01f596fddfbb.microcloud
> 2017-02-02 19:03:43+0100 4989 [1048]: s36 add_lockspace fail result -262
>
> Could you please check if you have other hosts contending for the same ID
> (id=3 in this case).
>

Another option is to manually force a sanlock renewal on that host and
check what happens, something like:
sanlock client renewal -s 7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96:3:/rhev/data-
center/mnt/glusterSD/glusterfs.rxmgmt.databay.de:_engine/7c8deaa8-be02-4aaf-
b9b4-ddc8da99ad96/dom_md/ids:0


>
>
>> Bye
>>
>> Am 03.02.2017 um 16:12 schrieb Simone Tiraboschi:
>>
>> The hosted-engine storage domain is mounted for sure,
>> but the issue is here:
>> Exception: Failed to start monitoring domain
>> (sd_uuid=7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96, host_id=3): timeout
>> during domain acquisition
>>
>> The point is that in VDSM logs I see just something like:
>> 2017-02-02 21:05:22,283 INFO  (jsonrpc/1) [dispatcher] Run and protect:
>> repoStats(options=None) (logUtils:49)
>> 2017-02-02 21:05:22,285 INFO  (jsonrpc/1) [dispatcher] Run and protect:
>> repoStats, Return response: {u'a7fbaaad-7043-4391-9523-3bedcdc4fb0d':
>> {'code': 0, 'actual': True, 'version': 0, 'acquired': True, 'delay':
>> '0.000748727', 'lastCheck': '0.1', 'valid': True},
>> u'2b2a44fc-f2bd-47cd-b7af-00be59e30a35': {'code': 0, 'actual': True,
>> 'version': 0, 'acquired': True, 'delay': '0.00082529', 'lastCheck': '0.1',
>> 'valid': True}, u'5d99af76-33b5-47d8-99da-1f32413c7bb0': {'code': 0,
>> 'actual': True, 'version': 4, 'acquired': True, 'delay': '0.000349356',
>> 'lastCheck': '5.3', 'valid': True}, u'7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96':
>> {'code': 0, 'actual': True, 'version': 4, 'acquired': False, 'delay':
>> '0.000377052', 'lastCheck': '0.6', 'valid': True}} (logUtils:52)
>>
>> Where the other storage domains have 'acquired': True whil it's
>> always 'acquired': False for the hosted-engine storage domain.
>>
>> Could you please share your /var/log/sanlock.log from the same host and
>> the output of
>>  sanlock client status
>> ?
>>
>>
>>
>>
>> On Fri, Feb 3, 2017 at 3:52 PM, Ralf Schenk <r...@databay.de> wrote:
>>
>>> Hello,
>>>
>>> I also put host in Maintenance and restarted vdsm while ovirt-ha-agent
>>> is running. I can mount the gluster Volume "engine" manually in the host.
>>>
>>> I get this repeatedly in /var/log/vdsm.log:
>>>
>>> 2017-02-03 15:29:28,891 INFO  (MainThread) [vds] Exiting (vdsm:167)
>>> 2017-02-03 15:29:30,974 INFO  (MainThread) [vds] (PID: 11456) I am the
>>> actual vdsm 4.19.4-1.el7.centos microcloud27 (3.10.0-514.6.1.el7.x86_64)
>>> (vdsm:145)
>>> 2017-02-03 15:29:30,974 INFO  (MainThread) [vds] VDSM will run with cpu
>>> affinity: frozenset([1]) (vdsm:251)
>>> 2017-02-03 15:29:31,013 INFO  (MainThread) [storage.check] Starting
>>> check service (check:91)
>>> 2017-02-03 15:29:31,017 INFO  (MainThread) [storage.Dispatcher] Starting
>>> StorageDispatcher... (dispatcher:47)
>>> 2017-02-03 15:29:31,017 INFO  (check/loop) [storage.asyncevent] Starting
>>>  (asyncevent:122)
>>> 2017-02-03 15:29:31,156 INFO  (MainThread) [dispatcher] Run and protect:
>>> registerDomainStateChangeCallback(callbackFunc=>> object at 0x2881fc8>) (logUtils:49)
>>> 2017-02-03 15:29:31,156 INFO  (MainThread) [dispatcher] Run and protect:
>>> registerDomainStateChangeCallback, Return response: None (logUtils:52)
>>> 2017-02-03 15:29:31,160 INFO  (MainThread) [MOM] Preparing MOM interface
>>> (momIF:49)
>>> 2017-02-03 15:29:31,161 INFO  (MainThread) [MOM] Using named unix socket
>>> /var/run/vdsm/mom-vdsm.sock (momIF:58)
>>> 2017-02-03 15:29:31,162 IN

Re: [ovirt-users] [Call for feedback] did you install/update to 4.1.0?

2017-02-03 Thread Simone Tiraboschi
On Fri, Feb 3, 2017 at 5:22 PM, Ralf Schenk <r...@databay.de> wrote:

> Hello,
>
> of course:
>
> [root@microcloud27 mnt]# sanlock client status
> daemon 8a93c9ea-e242-408c-a63d-a9356bb22df5.microcloud
> p -1 helper
> p -1 listener
> p -1 status
>
> sanlock.log attached. (Beginning 2017-01-27 where everything was fine)
>
Thanks, the issue is here:

2017-02-02 19:01:22+0100 4848 [1048]: s36 lockspace
7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96:3:/rhev/data-center/mnt/glusterSD/glusterfs.rxmgmt.databay.de:_engine/7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96/dom_md/ids:0
2017-02-02 19:03:42+0100 4988 [12983]: s36 delta_acquire host_id 3
busy1 3 15 13129 7ad427b1-fbb6-4cee-b9ee-01f596fddfbb.microcloud
2017-02-02 19:03:43+0100 4989 [1048]: s36 add_lockspace fail result -262

Could you please check if you have other hosts contending for the same ID
(id=3 in this case).


> Bye
>
> Am 03.02.2017 um 16:12 schrieb Simone Tiraboschi:
>
> The hosted-engine storage domain is mounted for sure,
> but the issue is here:
> Exception: Failed to start monitoring domain 
> (sd_uuid=7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96,
> host_id=3): timeout during domain acquisition
>
> The point is that in VDSM logs I see just something like:
> 2017-02-02 21:05:22,283 INFO  (jsonrpc/1) [dispatcher] Run and protect:
> repoStats(options=None) (logUtils:49)
> 2017-02-02 21:05:22,285 INFO  (jsonrpc/1) [dispatcher] Run and protect:
> repoStats, Return response: {u'a7fbaaad-7043-4391-9523-3bedcdc4fb0d':
> {'code': 0, 'actual': True, 'version': 0, 'acquired': True, 'delay':
> '0.000748727', 'lastCheck': '0.1', 'valid': True},
> u'2b2a44fc-f2bd-47cd-b7af-00be59e30a35': {'code': 0, 'actual': True,
> 'version': 0, 'acquired': True, 'delay': '0.00082529', 'lastCheck': '0.1',
> 'valid': True}, u'5d99af76-33b5-47d8-99da-1f32413c7bb0': {'code': 0,
> 'actual': True, 'version': 4, 'acquired': True, 'delay': '0.000349356',
> 'lastCheck': '5.3', 'valid': True}, u'7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96':
> {'code': 0, 'actual': True, 'version': 4, 'acquired': False, 'delay':
> '0.000377052', 'lastCheck': '0.6', 'valid': True}} (logUtils:52)
>
> Where the other storage domains have 'acquired': True whil it's
> always 'acquired': False for the hosted-engine storage domain.
>
> Could you please share your /var/log/sanlock.log from the same host and
> the output of
>  sanlock client status
> ?
>
>
>
>
> On Fri, Feb 3, 2017 at 3:52 PM, Ralf Schenk <r...@databay.de> wrote:
>
>> Hello,
>>
>> I also put host in Maintenance and restarted vdsm while ovirt-ha-agent is
>> running. I can mount the gluster Volume "engine" manually in the host.
>>
>> I get this repeatedly in /var/log/vdsm.log:
>>
>> 2017-02-03 15:29:28,891 INFO  (MainThread) [vds] Exiting (vdsm:167)
>> 2017-02-03 15:29:30,974 INFO  (MainThread) [vds] (PID: 11456) I am the
>> actual vdsm 4.19.4-1.el7.centos microcloud27 (3.10.0-514.6.1.el7.x86_64)
>> (vdsm:145)
>> 2017-02-03 15:29:30,974 INFO  (MainThread) [vds] VDSM will run with cpu
>> affinity: frozenset([1]) (vdsm:251)
>> 2017-02-03 15:29:31,013 INFO  (MainThread) [storage.check] Starting check
>> service (check:91)
>> 2017-02-03 15:29:31,017 INFO  (MainThread) [storage.Dispatcher] Starting
>> StorageDispatcher... (dispatcher:47)
>> 2017-02-03 15:29:31,017 INFO  (check/loop) [storage.asyncevent] Starting
>>  (asyncevent:122)
>> 2017-02-03 15:29:31,156 INFO  (MainThread) [dispatcher] Run and protect:
>> registerDomainStateChangeCallback(callbackFunc=> at 0x2881fc8>) (logUtils:49)
>> 2017-02-03 15:29:31,156 INFO  (MainThread) [dispatcher] Run and protect:
>> registerDomainStateChangeCallback, Return response: None (logUtils:52)
>> 2017-02-03 15:29:31,160 INFO  (MainThread) [MOM] Preparing MOM interface
>> (momIF:49)
>> 2017-02-03 15:29:31,161 INFO  (MainThread) [MOM] Using named unix socket
>> /var/run/vdsm/mom-vdsm.sock (momIF:58)
>> 2017-02-03 15:29:31,162 INFO  (MainThread) [root] Unregistering all
>> secrets (secret:91)
>> 2017-02-03 15:29:31,164 INFO  (MainThread) [vds] Setting channels'
>> timeout to 30 seconds. (vmchannels:223)
>> 2017-02-03 15:29:31,165 INFO  (MainThread) [vds.MultiProtocolAcceptor]
>> Listening at :::54321 (protocoldetector:185)
>> 2017-02-03 15:29:31,354 INFO  (vmrecovery) [vds] recovery: completed in
>> 0s (clientIF:495)
>> 2017-02-03 15:29:31,371 INFO  (BindingXMLRPC) [vds] XMLRPC server running
>> (bindingxmlrpc:63)
>> 2017-02-03 15:29:31,471 INFO  (periodic/1) [dispatcher] Run and protect:
>> repoStats(options=None) (logUtils:49)
>> 2017-02-03 15:29:31,472 INFO  (periodic/1) [dispatcher] Run and protect:
>> repoStats, Return

Re: [ovirt-users] [Call for feedback] did you install/update to 4.1.0?

2017-02-06 Thread Simone Tiraboschi
On Mon, Feb 6, 2017 at 12:42 PM, Ralf Schenk <r...@databay.de> wrote:

> Hello,
>
> I set the host to maintenance mode and tried to undeploy engine via GUI.
> The action in GUI doesn't show an error but afterwards it still shows only
> "Undeploy" on hosted-engine tab od the host.
>
> Even removing the host from the cluster doesn't work because the GUI says
> "The hosts maekred with * still have hosted engine deployed on them. Hosted
> engine should be undeployed before they are removed"
>
Yes, sorry: it's now a two step process, you have first to undeploy
hosted-engine from the host and only then you could remove the host.



> Bye
> Am 06.02.2017 um 11:44 schrieb Simone Tiraboschi:
>
>
>
> On Sat, Feb 4, 2017 at 11:52 AM, Ralf Schenk <r...@databay.de> wrote:
>
>> Hello,
>>
>> I have set up 3 hosts for engine, 2 of them are working correct. There is
>> no other host even having broker/agent installed. Is it possible that the
>> error occurs because the hosts are multihomed (Management IP, IP for
>> storage) and can communicate with different IP's ?
>>
> Having multiple logical networks for storage, management and so on is a
> good practice and it's advised so I tend to exclude any issue there.
> The point is why your microcloud27.sub.mydomain.de fails acquiring a lock
> as host 3.
> Probably the simplest fix is just setting it in maintenance mode from the
> engine, removing it and deploying it from the engine as an hosted engine
> host again.
>
>
>
>
> --
>
>
> *Ralf Schenk*
> fon +49 (0) 24 05 / 40 83 70 <+49%202405%20408370>
> fax +49 (0) 24 05 / 40 83 759 <+49%202405%204083759>
> mail *r...@databay.de* <r...@databay.de>
>
> *Databay AG*
> Jens-Otto-Krag-Straße 11
> D-52146 Würselen
> *www.databay.de* <http://www.databay.de>
>
> Sitz/Amtsgericht Aachen • HRB:8437 • USt-IdNr.: DE 210844202
> Vorstand: Ralf Schenk, Dipl.-Ing. Jens Conze, Aresch Yavari, Dipl.-Kfm.
> Philipp Hermanns
> Aufsichtsratsvorsitzender: Wilhelm Dohmen
> --
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [Call for feedback] did you install/update to 4.1.0?

2017-02-06 Thread Simone Tiraboschi
On Sat, Feb 4, 2017 at 11:52 AM, Ralf Schenk <r...@databay.de> wrote:

> Hello,
>
> I have set up 3 hosts for engine, 2 of them are working correct. There is
> no other host even having broker/agent installed. Is it possible that the
> error occurs because the hosts are multihomed (Management IP, IP for
> storage) and can communicate with different IP's ?
>
Having multiple logical networks for storage, management and so on is a
good practice and it's advised so I tend to exclude any issue there.
The point is why your microcloud27.sub.mydomain.de fails acquiring a lock
as host 3.
Probably the simplest fix is just setting it in maintenance mode from the
engine, removing it and deploying it from the engine as an hosted engine
host again.



> hosted-engine --vm-status on both working hosts seems correct: (3 is out
> of order...)
>
> [root@microcloud21 ~]# hosted-engine --vm-status
>
>
> --== Host 1 status ==--
>
> conf_on_shared_storage : True
> Status up-to-date  : True
> Hostname   : microcloud21.sub.mydomain.de
> Host ID: 1
> Engine status  : {"health": "good", "vm": "up",
> "detail": "up"}
> Score  : 3400
> stopped: False
> Local maintenance  : False
> crc32  : 5941227d
> local_conf_timestamp   : 152316
> Host timestamp : 152302
> Extra metadata (valid at timestamp):
> metadata_parse_version=1
> metadata_feature_version=1
> timestamp=152302 (Sat Feb  4 11:49:29 2017)
> host-id=1
> score=3400
> vm_conf_refresh_time=152316 (Sat Feb  4 11:49:43 2017)
> conf_on_shared_storage=True
> maintenance=False
> state=EngineUp
> stopped=False
>
>
> --== Host 2 status ==--
>
> conf_on_shared_storage : True
> Status up-to-date  : True
> Hostname   : microcloud24.sub.mydomain.de
> Host ID: 2
> Engine status  : {"reason": "vm not running on this
> host", "health": "bad", " vm": "down",
> "detail": "unknown"}
> Score  : 3400
> stopped: False
> Local maintenance  : False
> crc32  : 77e25433
> local_conf_timestamp   : 157637
> Host timestamp : 157623
> Extra metadata (valid at timestamp):
> metadata_parse_version=1
> metadata_feature_version=1
> timestamp=157623 (Sat Feb  4 11:49:34 2017)
> host-id=2
> score=3400
> vm_conf_refresh_time=157637 (Sat Feb  4 11:49:48 2017)
> conf_on_shared_storage=True
> maintenance=False
> state=EngineDown
> stopped=False
>
>
> --== Host 3 status ==--
>
> conf_on_shared_storage : True
> Status up-to-date  : False
> Hostname   : microcloud27.sub.mydomain.de
> Host ID: 3
> Engine status  : unknown stale-data
> Score  : 0
> stopped: True
> Local maintenance  : False
> crc32  : 74798986
> local_conf_timestamp   : 77946
> Host timestamp : 77932
> Extra metadata (valid at timestamp):
> metadata_parse_version=1
> metadata_feature_version=1
> timestamp=77932 (Fri Feb  3 15:19:25 2017)
> host-id=3
> score=0
> vm_conf_refresh_time=77946 (Fri Feb  3 15:19:39 2017)
> conf_on_shared_storage=True
> maintenance=False
> state=AgentStopped
> stopped=True
>
> Am 03.02.2017 um 19:20 schrieb Simone Tiraboschi:
>
>
>
> On Fri, Feb 3, 2017 at 5:22 PM, Ralf Schenk <r...@databay.de> wrote:
>
>> Hello,
>>
>> of course:
>>
>> [root@microcloud27 mnt]# sanlock client status
>> daemon 8a93c9ea-e242-408c-a63d-a9356bb22df5.microcloud
>> p -1 helper
>> p -1 listener
>> p -1 status
>>
>> sanlock.log attached. (Beginning 2017-01-27 where everything was fine)
>>
> Thanks, the issue is here:
>
> 2017-02-02 19:01:22+0100 4848 [1048]: s36 lockspace 
> 7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96:3:/rhev/data-center/mnt/glusterSD/glusterfs.sub.mydomain.de:_eng

Re: [ovirt-users] Restoring Hosted-Engine from a stale backup

2017-02-06 Thread Simone Tiraboschi
On Mon, Feb 6, 2017 at 1:52 PM, Doug Ingham <dou...@gmail.com> wrote:

> Hi All, Simone,
>
> On 24 January 2017 at 10:11, Simone Tiraboschi <stira...@redhat.com>
> wrote:
>
>>
>>
>> On Tue, Jan 24, 2017 at 1:49 PM, Doug Ingham <dou...@gmail.com> wrote:
>>
>>> Hey guys,
>>>  Just giving this a bump in the hope that someone might be able to
>>> advise...
>>>
>>> Hi all,
>>>>  One of our engines has had a DB failure* & it seems there was an
>>>> unnoticed problem in its backup routine, meaning the last backup I've got
>>>> is a couple of weeks old.
>>>> Luckily, VDSM has kept the underlying VMs running without any
>>>> interruptions, so my objective is to get the HE back online & get the hosts
>>>> & VMs back under its control with minimal downtime.
>>>>
>>>> So, my questions are the following...
>>>>
>>>>1. What problems can I expect to have with VMs added/modified since
>>>>the last backup?
>>>>
>>>> Modified VMs will be reverted to the previous configuration; additional
>> VMs should be seen as external VMs, then you could import.
>>
>
> Given VDSM kept the VMs up whilst the HE's been down, how will the running
> VMs that were present before & after the backup be affected?
>
> Many of the VMs that were present during the last backup are now on
> different hosts, including the HE VM. Will that cause any issues?
>

For normal VMs I don't expect any issue: the engine will simply update the
correspondent record once it will find them on the managed hosts.
A serious issue could instead happen with HA VMs:
if the engine finds earlier an HA VM as running on a different host it will
simply update its record, the issue is if it finds earlier the VM a not on
the original host since it will try to restart it causing a split brain and
probably a VM corruption.
I opened a bug to track it:
https://bugzilla.redhat.com/show_bug.cgi?id=1419649



>
>
>>
>>
>>>
>>>>1. As it's only the DB that's been affected, can I skip redeploying
>>>>the Engine & jump straight to restoring the DB & rerunning engine-setup?
>>>>
>>>>
>> Yes, if the engine VM is fine, you could just import the previous backup
>> and run engine-setup again.
>> Please set the global maintenance mode for hosted-engine since
>> engine-backup and engine-setup are going to bring down the engine.
>>
>
> As per above, do I still only need to import the previous backup even if
> the all of the VMs (including the HE VM) are now on different hosts to when
> the backup was made?
>

Please take care of the HA VMs.


>
>
> And as for the future, is it going to be necessary to always keep an
> unused host in the cluster to allow for emergency restores? I'm a bit
> concerned that if we ever utilised all of our hosts for running VMs, then
> we'd be completely stuck if the HE ever imploded again.
>

Honestly I don't see any special issue there.


>
> Cheers,
> --
> Doug
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted engine errors after 4.1 upgrade.

2017-02-08 Thread Simone Tiraboschi
On Wed, Feb 8, 2017 at 8:59 AM, Yedidyah Bar David  wrote:

> On Wed, Feb 8, 2017 at 2:31 AM, Todd Punderson 
> wrote:
> > Seeing issues with my hosted engine, it seems it's unable to extract
> vm.conf
> > from storage. My ovirt-hosted-engine-ha/agent.log is full of this
> repeating
> > over and over. This is happening on all 3 of my hosts. My storage is
> > glusterfs on the hosts themselves.
> >
> > Hopefully this is enough info to get started.
>

Another step is editing /etc/ovirt-hosted-engine-ha/agent-log.conf changing
from

[logger_root]
level=INFO
handlers=syslog,logfile
propagate=0

to

[logger_root]
level=DEBUG
handlers=syslog,logfile
propagate=0

and restart ovirt-ha-agent to get more detailed info about the issue.


> >
> > Thanks!
> >
> > MainThread::INFO::2017-02-07
> > 19:27:33,063::hosted_engine::612::ovirt_hosted_engine_ha.
> agent.hosted_engine.HostedEngine::(_initialize_vdsm)
> > Initializing VDSM
> > MainThread::INFO::2017-02-07
> > 19:27:35,455::hosted_engine::639::ovirt_hosted_engine_ha.
> agent.hosted_engine.HostedEngine::(_initialize_storage_images)
> > Connecting the storage
> > MainThread::INFO::2017-02-07
> > 19:27:35,456::storage_server::219::ovirt_hosted_engine_ha.
> lib.storage_server.StorageServer::(connect_storage_server)
> > Connecting storage server
> > MainThread::INFO::2017-02-07
> > 19:27:40,169::storage_server::226::ovirt_hosted_engine_ha.
> lib.storage_server.StorageServer::(connect_storage_server)
> > Connecting storage server
> > MainThread::INFO::2017-02-07
> > 19:27:40,202::storage_server::233::ovirt_hosted_engine_ha.
> lib.storage_server.StorageServer::(connect_storage_server)
> > Refreshing the storage domain
> > MainThread::INFO::2017-02-07
> > 19:27:40,418::hosted_engine::666::ovirt_hosted_engine_ha.
> agent.hosted_engine.HostedEngine::(_initialize_storage_images)
> > Preparing images
> > MainThread::INFO::2017-02-07
> > 19:27:40,419::image::126::ovirt_hosted_engine_ha.lib.
> image.Image::(prepare_images)
> > Preparing images
> > MainThread::INFO::2017-02-07
> > 19:27:43,370::hosted_engine::669::ovirt_hosted_engine_ha.
> agent.hosted_engine.HostedEngine::(_initialize_storage_images)
> > Reloading vm.conf from the shared storage domain
> > MainThread::INFO::2017-02-07
> > 19:27:43,371::config::206::ovirt_hosted_engine_ha.agent.
> hosted_engine.HostedEngine.config::(refresh_local_conf_file)
> > Trying to get a fresher copy of vm configuration from the OVF_STORE
> > MainThread::INFO::2017-02-07
> > 19:27:45,968::ovf_store::103::ovirt_hosted_engine_ha.lib.
> ovf.ovf_store.OVFStore::(scan)
> > Found OVF_STORE: imgUUID:3e14c1b5-5ade-4827-aad4-66c59824acd2,
> > volUUID:3cbeeb3b-f755-4d42-a654-8dab34213792
> > MainThread::INFO::2017-02-07
> > 19:27:46,257::ovf_store::103::ovirt_hosted_engine_ha.lib.
> ovf.ovf_store.OVFStore::(scan)
> > Found OVF_STORE: imgUUID:9b49968b-5a62-4ab2-a2c5-b94bc0b2d3ce,
> > volUUID:8f4d69c5-73a7-4e8c-a58f-909b55efec7d
> > MainThread::INFO::2017-02-07
> > 19:27:46,355::ovf_store::112::ovirt_hosted_engine_ha.lib.
> ovf.ovf_store.OVFStore::(getEngineVMOVF)
> > Extracting Engine VM OVF from the OVF_STORE
> > MainThread::INFO::2017-02-07
> > 19:27:46,366::ovf_store::119::ovirt_hosted_engine_ha.lib.
> ovf.ovf_store.OVFStore::(getEngineVMOVF)
> > OVF_STORE volume path:
> > /rhev/data-center/mnt/glusterSD/ovirt01-gluster.doonga.org:
> _engine/536cd721-4396-4029-b1ea-8ce84738137e/images/9b49968b-5a62-4ab2-
> a2c5-b94bc0b2d3ce/8f4d69c5-73a7-4e8c-a58f-909b55efec7d
> > MainThread::ERROR::2017-02-07
> > 19:27:46,389::ovf_store::124::ovirt_hosted_engine_ha.lib.
> ovf.ovf_store.OVFStore::(getEngineVMOVF)
> > Unable to extract HEVM OVF
> > MainThread::ERROR::2017-02-07
> > 19:27:46,390::config::235::ovirt_hosted_engine_ha.agent.
> hosted_engine.HostedEngine.config::(refresh_local_conf_file)
> > Unable to get vm.conf from OVF_STORE, falling back to initial vm.conf
>
> Can you please attach the output of:
>
> sudo -u vdsm dd
> if=/rhev/data-center/mnt/glusterSD/ovirt01-gluster.doonga.org:
> _engine/536cd721-4396-4029-b1ea-8ce84738137e/images/9b49968b-5a62-4ab2-
> a2c5-b94bc0b2d3ce/8f4d69c5-73a7-4e8c-a58f-909b55efec7d
> | tar -tvf -
>
> Thanks.
>
> Did everything work well in 4.0? How did you upgrade?
>
> Best,
> --
> Didi
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [Call for feedback] did you install/update to 4.1.0?

2017-02-03 Thread Simone Tiraboschi
I see there an ERROR on stopMonitoringDomain but I cannot see the
correspondent  startMonitoringDomain; could you please look for it?

On Fri, Feb 3, 2017 at 1:16 PM, Ralf Schenk <r...@databay.de> wrote:

> Hello,
>
> attached is my vdsm.log from the host with hosted-engine-ha around the
> time-frame of agent timeout that is not working anymore for engine (it
> works in Ovirt and is active). It simply isn't working for engine-ha
> anymore after Update.
>
> At 2017-02-02 19:25:34,248 you'll find an error corresponoding to agent
> timeout error.
>
> Bye
>
>
>
> Am 03.02.2017 um 11:28 schrieb Simone Tiraboschi:
>
> 3. Three of my hosts have the hosted engine deployed for ha. First all
>>> three where marked by a crown (running was gold and others where silver).
>>> After upgrading the 3 Host deployed hosted engine ha is not active anymore.
>>>
>>> I can't get this host back with working ovirt-ha-agent/broker. I already
>>> rebooted, manually restarted the services but It isn't able to get cluster
>>> state according to
>>> "hosted-engine --vm-status". The other hosts state the host status as
>>> "unknown stale-data"
>>>
>>> I already shut down all agents on all hosts and issued a "hosted-engine
>>> --reinitialize-lockspace" but that didn't help.
>>>
>>> Agents stops working after a timeout-error according to log:
>>>
>>> MainThread::INFO::2017-02-02 19:24:52,040::hosted_engine::8
>>> 41::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_domain_monitor_status)
>>> VDSM domain monitor status: PENDING
>>> MainThread::INFO::2017-02-02 19:24:59,185::hosted_engine::8
>>> 41::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_domain_monitor_status)
>>> VDSM domain monitor status: PENDING
>>> MainThread::INFO::2017-02-02 19:25:06,333::hosted_engine::8
>>> 41::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_domain_monitor_status)
>>> VDSM domain monitor status: PENDING
>>> MainThread::INFO::2017-02-02 19:25:13,554::hosted_engine::8
>>> 41::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_domain_monitor_status)
>>> VDSM domain monitor status: PENDING
>>> MainThread::INFO::2017-02-02 19:25:20,710::hosted_engine::8
>>> 41::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_domain_monitor_status)
>>> VDSM domain monitor status: PENDING
>>> MainThread::INFO::2017-02-02 19:25:27,865::hosted_engine::8
>>> 41::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_domain_monitor_status)
>>> VDSM domain monitor status: PENDING
>>> MainThread::ERROR::2017-02-02 19:25:27,866::hosted_engine::8
>>> 15::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_domain_monitor)
>>> Failed to start monitoring domain 
>>> (sd_uuid=7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96,
>>> host_id=3): timeout during domain acquisition
>>> MainThread::WARNING::2017-02-02 19:25:27,866::hosted_engine::4
>>> 69::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
>>> Error while monitoring engine: Failed to start monitoring domain
>>> (sd_uuid=7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96, host_id=3): timeout
>>> during domain acquisition
>>> MainThread::WARNING::2017-02-02 19:25:27,866::hosted_engine::4
>>> 72::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
>>> Unexpected error
>>> Traceback (most recent call last):
>>>   File 
>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
>>> line 443, in start_monitoring
>>> self._initialize_domain_monitor()
>>>   File 
>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
>>> line 816, in _initialize_domain_monitor
>>> raise Exception(msg)
>>> Exception: Failed to start monitoring domain
>>> (sd_uuid=7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96, host_id=3): timeout
>>> during domain acquisition
>>> MainThread::ERROR::2017-02-02 19:25:27,866::hosted_engine::4
>>> 85::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
>>> Shutting down the agent because of 3 failures in a row!
>>> MainThread::INFO::2017-02-02 19:25:32,087::hosted_engine::8
>>> 41::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_domain_monitor_status)
>>> VDSM domain monitor status: PENDING
>>> MainThread::INFO::2017-02-02 19:25:34,250::hosted_eng

Re: [ovirt-users] [Call for feedback] did you install/update to 4.1.0?

2017-02-03 Thread Simone Tiraboschi
On Fri, Feb 3, 2017 at 11:17 AM, Sandro Bonazzola 
wrote:

>
>
> On Fri, Feb 3, 2017 at 10:54 AM, Ralf Schenk  wrote:
>
>> Hello,
>>
>> I upgraded my cluster of 8 hosts with gluster storage and
>> hosted-engine-ha. They were already Centos 7.3 and using Ovirt 4.0.6 and
>> gluster 3.7.x packages from storage-sig testing.
>>
>> I'm missing the storage listed under storage tab but this is already
>> filed by a bug. Increasing Cluster and Storage Compability level and also
>> "reset emulated machine" after having upgraded one host after another
>> without the need to shutdown vm's works well. (VM's get sign that there
>> will be changes after reboot).
>>
>> Important: you also have to issue a yum update on the host for upgrading
>> additional components like i.e. gluster to 3.8.x. I was frightened of this
>> step but It worked well except a configuration issue I was responsible for
>> in gluster.vol (I had "transport socket, rdma")
>>
>> Bugs/Quirks so far:
>>
>> 1. After restarting a single VM that used RNG-Device I got an error (it
>> was german) but like "RNG Device not supported by cluster". I hat to
>> disable RNG Device save the settings. Again settings and enable RNG Device.
>> Then machine boots up.
>> I think there is a migration step missing from /dev/random to
>> /dev/urandom for exisiting VM's.
>>
>
> Tomas, Francesco, Michal, can you please follow up on this?
>
>
>
>> 2. I'm missing any gluster specific management features as my gluster is
>> not managable in any way from the GUI. I expected to see my gluster now in
>> dashboard and be able to add volumes etc. What do I need to do to "import"
>> my existing gluster (Only one volume so far) to be managable ?
>>
>
> Sahina, can you please follow up on this?
>
>
>> 3. Three of my hosts have the hosted engine deployed for ha. First all
>> three where marked by a crown (running was gold and others where silver).
>> After upgrading the 3 Host deployed hosted engine ha is not active anymore.
>>
>> I can't get this host back with working ovirt-ha-agent/broker. I already
>> rebooted, manually restarted the services but It isn't able to get cluster
>> state according to
>> "hosted-engine --vm-status". The other hosts state the host status as
>> "unknown stale-data"
>>
>> I already shut down all agents on all hosts and issued a "hosted-engine
>> --reinitialize-lockspace" but that didn't help.
>>
>> Agents stops working after a timeout-error according to log:
>>
>> MainThread::INFO::2017-02-02 19:24:52,040::hosted_engine::8
>> 41::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_domain_monitor_status)
>> VDSM domain monitor status: PENDING
>> MainThread::INFO::2017-02-02 19:24:59,185::hosted_engine::8
>> 41::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_domain_monitor_status)
>> VDSM domain monitor status: PENDING
>> MainThread::INFO::2017-02-02 19:25:06,333::hosted_engine::8
>> 41::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_domain_monitor_status)
>> VDSM domain monitor status: PENDING
>> MainThread::INFO::2017-02-02 19:25:13,554::hosted_engine::8
>> 41::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_domain_monitor_status)
>> VDSM domain monitor status: PENDING
>> MainThread::INFO::2017-02-02 19:25:20,710::hosted_engine::8
>> 41::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_domain_monitor_status)
>> VDSM domain monitor status: PENDING
>> MainThread::INFO::2017-02-02 19:25:27,865::hosted_engine::8
>> 41::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_domain_monitor_status)
>> VDSM domain monitor status: PENDING
>> MainThread::ERROR::2017-02-02 19:25:27,866::hosted_engine::8
>> 15::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_domain_monitor)
>> Failed to start monitoring domain 
>> (sd_uuid=7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96,
>> host_id=3): timeout during domain acquisition
>> MainThread::WARNING::2017-02-02 19:25:27,866::hosted_engine::4
>> 69::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
>> Error while monitoring engine: Failed to start monitoring domain
>> (sd_uuid=7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96, host_id=3): timeout
>> during domain acquisition
>> MainThread::WARNING::2017-02-02 19:25:27,866::hosted_engine::4
>> 72::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
>> Unexpected error
>> Traceback (most recent call last):
>>   File 
>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
>> line 443, in start_monitoring
>> self._initialize_domain_monitor()
>>   File 
>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
>> line 816, in _initialize_domain_monitor
>> raise Exception(msg)
>> Exception: Failed to start monitoring domain
>> (sd_uuid=7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96, host_id=3): timeout
>> during domain acquisition
>> MainThread::ERROR::2017-02-02 

Re: [ovirt-users] [Call for feedback] did you install/update to 4.1.0?

2017-02-03 Thread Simone Tiraboschi
t;
> 2017-02-03 15:29:35,920 INFO  (Reactor thread) [ProtocolDetector.AcceptorImpl]
> Accepted connection from ::1:49506 (protocoldetector:72)
> 2017-02-03 15:29:35,929 INFO  (Reactor thread) [ProtocolDetector.Detector]
> Detected protocol stomp from ::1:49506 (protocoldetector:127)
> 2017-02-03 15:29:35,930 INFO  (Reactor thread) [Broker.StompAdapter]
> Processing CONNECT request (stompreactor:102)
> 2017-02-03 15:29:35,930 INFO  (JsonRpc (StompReactor))
> [Broker.StompAdapter] Subscribe command received (stompreactor:129)
> 2017-02-03 15:29:36,067 INFO  (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC call
> Host.ping succeeded in 0.00 seconds (__init__:515)
> 2017-02-03 15:29:36,071 INFO  (jsonrpc/1) [throttled] Current
> getAllVmStats: {} (throttledlog:105)
> 2017-02-03 15:29:36,071 INFO  (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC call
> Host.getAllVmStats succeeded in 0.00 seconds (__init__:515)
> 2017-02-03 15:29:46,435 INFO  (periodic/0) [dispatcher] Run and protect:
> repoStats(options=None) (logUtils:49)
> 2017-02-03 15:29:46,435 INFO  (periodic/0) [dispatcher] Run and protect:
> repoStats, Return response: {} (logUtils:52)
> 2017-02-03 15:29:46,439 ERROR (periodic/0) [root] failed to retrieve
> Hosted Engine HA info (api:252)
> Traceback (most recent call last):
>   File "/usr/lib/python2.7/site-packages/vdsm/host/api.py", line 231, in
> _getHaInfo
> stats = instance.get_all_stats()
>   File 
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py",
> line 103, in get_all_stats
> self._configure_broker_conn(broker)
>   File 
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py",
> line 180, in _configure_broker_conn
> dom_type=dom_type)
>   File 
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
> line 177, in set_storage_domain
> .format(sd_type, options, e))
> RequestError: Failed to set storage domain FilesystemBackend, options
> {'dom_type': 'glusterfs', 'sd_uuid': '7c8deaa8-be02-4aaf-b9b4-ddc8da99ad96'}:
> Request failed:  ted_engine_ha.lib.storage_backends.BackendFailureException'>
> 2017-02-03 15:29:51,095 INFO  (jsonrpc/2) [jsonrpc.JsonRpcServer] RPC call
> Host.getAllVmStats succeeded in 0.00 seconds (__init__:515)
> 2017-02-03 15:29:51,219 INFO  (jsonrpc/3) [jsonrpc.JsonRpcServer] RPC call
> Host.setKsmTune succeeded in 0.00 seconds (__init__:515)
> 2017-02-03 15:30:01,444 INFO  (periodic/1) [dispatcher] Run and protect:
> repoStats(options=None) (logUtils:49)
> 2017-02-03 15:30:01,444 INFO  (periodic/1) [dispatcher] Run and protect:
> repoStats, Return response: {} (logUtils:52)
> 2017-02-03 15:30:01,448 ERROR (periodic/1) [root] failed to retrieve
> Hosted Engine HA info (api:252)
>
>
>
> Am 03.02.2017 um 13:39 schrieb Simone Tiraboschi:
>
> I see there an ERROR on stopMonitoringDomain but I cannot see the
> correspondent  startMonitoringDomain; could you please look for it?
>
> On Fri, Feb 3, 2017 at 1:16 PM, Ralf Schenk <r...@databay.de> wrote:
>
>> Hello,
>>
>> attached is my vdsm.log from the host with hosted-engine-ha around the
>> time-frame of agent timeout that is not working anymore for engine (it
>> works in Ovirt and is active). It simply isn't working for engine-ha
>> anymore after Update.
>>
>> At 2017-02-02 19:25:34,248 you'll find an error corresponoding to agent
>> timeout error.
>>
>> Bye
>>
>>
>>
>> Am 03.02.2017 um 11:28 schrieb Simone Tiraboschi:
>>
>> 3. Three of my hosts have the hosted engine deployed for ha. First all
>>>> three where marked by a crown (running was gold and others where silver).
>>>> After upgrading the 3 Host deployed hosted engine ha is not active anymore.
>>>>
>>>> I can't get this host back with working ovirt-ha-agent/broker. I
>>>> already rebooted, manually restarted the services but It isn't able to get
>>>> cluster state according to
>>>> "hosted-engine --vm-status". The other hosts state the host status as
>>>> "unknown stale-data"
>>>>
>>>> I already shut down all agents on all hosts and issued a "hosted-engine
>>>> --reinitialize-lockspace" but that didn't help.
>>>>
>>>> Agents stops working after a timeout-error according to log:
>>>>
>>>> MainThread::INFO::2017-02-02 19:24:52,040::hosted_engine::8
>>>> 41::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_domain_monitor_status)
>>>> VDSM domain monitor status: PENDING
>>>> MainThread::INFO::2017-02-02 19:24:59

Re: [ovirt-users] Use of Direct Attached Storage as Storage Type

2017-01-24 Thread Simone Tiraboschi
On Mon, Jan 23, 2017 at 9:55 PM, Fernando Frediani <
fernando.fredi...@upx.com.br> wrote:

> I am trying to use a DAC (Direct Attached Storage) in a Hardware which has
> it as the only option (DELL VRTX) but I see no matching option at Storage
> Type. It has there: NFS, POSIX compliant FS, GlusterFS, iSCSI and Fiber
> Channel. The two closest are iSCSI and Fiberchannel as they use CLVM in the
> backend I understand, but of course they are not the same thing as DAC.
>
> Has anyone had a similar scenario and how solved it ?
>

Did you tried simply choosing FC for that? do you see your volumes in that
case?


>
> Thanks
>
> Fernando
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Update NFS storage of the HE server

2017-01-24 Thread Simone Tiraboschi
On Mon, Jan 23, 2017 at 6:51 PM, Claude Duropcher <
claude.duroc...@cptaq.gouv.qc.ca> wrote:

> Simone, thanks for replying,
>
> According to the documentation, global mode seems to be a superset of
> local mode. Does setting to local mode maintenance on all hosts in the
> cluster really bring any benefits?
>

No, not really: they are different and global is not superset of local.
global maintenance mode is needed if you need to do maintenance on the
engine VM, no host will restart the engine VM. local maintenance mode is if
you need to do maintenance or upgrade a single host.



>
>-
>
>global - All high-availability agents in the cluster are disabled from
>monitoring the state of the engine virtual machine. The global maintenance
>mode must be applied for any setup or upgrade operations that require the
>engine to be stopped, such as upgrading to a later version of oVirt.
>-
>
>local - The high-availability agent on the host issuing the command is
>disabled from monitoring the state of the engine virtual machine. The host
>is exempt from hosting the engine virtual machine while in local
>maintenance mode; if hosting the engine virtual machine when placed into
>this mode, the engine will be migrated to another host, provided there is a
>suitable contender. The local maintenance mode is recommended when applying
>system changes or updates to the host.
>
> Le 2017-01-23 à 12:13, Simone Tiraboschi a écrit :
>
>
>
> On Mon, Jan 23, 2017 at 4:20 PM, Claude Duropcher <
> claude.duroc...@cptaq.gouv.qc.ca> wrote:
>
>> Hi,
>>
>> We need to update/reboot the NFS server on wich the hosted engine is
>> installed. This server also hosts a couple of extra storage domains (ISO,
>> Export and a iScsi domain). I'm wondering if this procedure is complete for
>> an oVirt 4.0 installation :
>>
>> 1-set the environment to global maintenance
>>
>> 2-set the extra domains to maintenance mode
>>
>> 3-shutdown the hosted engine
>>
>> 4-update/reboot the storage server
>>
>> 5-reboot the hosted engine
>>
>> 6-reactivate extra domains
>>
>> 7-disable global maintenance mode
>>
>> I'm wondering what's happening after step 4 and how ovirt will react when
>> the NFS storage disapear and come back.
>>
>>
> ovirt-ha-agent will try to reconnect the shared storage if hosted-engine
> for all the time; depending on different factors, sanlock will fail
> updating the hosted-engine lease ans this can bring the watchdog to reboot
> the host.
> Setting also local maintenance mode on the involved hosts is a good idea.
>
>
>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Restoring Hosted-Engine from a stale backup

2017-01-24 Thread Simone Tiraboschi
On Tue, Jan 24, 2017 at 1:49 PM, Doug Ingham  wrote:

> Hey guys,
>  Just giving this a bump in the hope that someone might be able to
> advise...
>
> Hi all,
>>  One of our engines has had a DB failure* & it seems there was an
>> unnoticed problem in its backup routine, meaning the last backup I've got
>> is a couple of weeks old.
>> Luckily, VDSM has kept the underlying VMs running without any
>> interruptions, so my objective is to get the HE back online & get the hosts
>> & VMs back under its control with minimal downtime.
>>
>> So, my questions are the following...
>>
>>1. What problems can I expect to have with VMs added/modified since
>>the last backup?
>>
>> Modified VMs will be reverted to the previous configuration; additional
VMs should be seen as external VMs, then you could import.


>
>>1. As it's only the DB that's been affected, can I skip redeploying
>>the Engine & jump straight to restoring the DB & rerunning engine-setup?
>>
>>
Yes, if the engine VM is fine, you could just import the previous backup
and run engine-setup again.
Please set the global maintenance mode for hosted-engine since
engine-backup and engine-setup are going to bring down the engine.

>
>>1. The original docs I read didn't mention that it's best to leave a
>>host in maintenance mode before running the engine backup, so my plan is 
>> to
>>install a new temporary host on a separate server, re-add the old hosts &
>>then once everything's back up, remove the temporary host. Are there any
>>faults in this plan?
>>2. When it comes to deleting the old HE VM, the docs point to a
>>paywalled guide on redhat.com...?
>>
>>  To add a bit more info to 4), I'm referring to the following...
>
> Note: If the Engine database is restored successfully, but the Engine
>> virtual machine appears to be Down and cannot be migrated to another
>> self-hosted engine host, you can enable a new Engine virtual machine and
>> remove the dead Engine virtual machine from the environment by following
>> the steps provided in https://access.redhat.com/solutions/1517683.
>>
> Source: http://www.ovirt.org/documentation/self-hosted/
> chap-Backing_up_and_Restoring_an_EL-Based_Self-Hosted_Environment/
>
>
If you are re-importing the backup in place on the initial engine VM you
don't have to.
The point is just if you are migrating to a new engine VM and so you have
to remove the entry of the previous one to let the auto-import process
trigger again.


> CentOS 7
>> oVirt 4.0.4
>> Gluster 3.8
>>
>> * Apparently a write somehow cleared fsync, despite not actually having
>> been written to disk?! No idea how that happened...
>>
>> Many thanks,
>> --
>> Doug
>>
>
> Cheers,
> --
> Doug
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Official Hyperconverged Gluster oVirt upgrade procedure?

2017-01-26 Thread Simone Tiraboschi
On Thu, Jan 26, 2017 at 10:16 AM, Ralf Schenk  wrote:

> Hello,
>
> i would appreciate any hint, too. I'm on 4.0.6 on Centos 7.3 since
> yesterday but I'm frightened what I need to do to upgrade and be able to
> manage gluster from GUI then.
>

1. Setting global maintenance mode,
2. upgrading the engine on the engine VM as for a regular engine,
3. exit the global maintenance mode
4. upgrade the host (once at time!!!) from the engine

should be enough


> Bye
>
> Am 25.01.2017 um 21:32 schrieb Hanson:
>
> Hi Guys,
>
> Just wondering if we have an updated manual or whats the current procedure
> for upgrading the nodes in a hyperconverged ovirt gluster pool?
>
> Ie Nodes run 4.0 oVirt, as well as GlusterFS, and hosted-engine running in
> a gluster storage domain.
>
> Put node in maintenance mode and disable glusterfs from ovirt gui, run yum
> update?
>
> Thanks!
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
> --
>
>
> *Ralf Schenk*
> fon +49 (0) 24 05 / 40 83 70 <+49%202405%20408370>
> fax +49 (0) 24 05 / 40 83 759 <+49%202405%204083759>
> mail *r...@databay.de* 
>
> *Databay AG*
> Jens-Otto-Krag-Straße 11
> D-52146 Würselen
> *www.databay.de* 
>
> Sitz/Amtsgericht Aachen • HRB:8437 • USt-IdNr.: DE 210844202
> Vorstand: Ralf Schenk, Dipl.-Ing. Jens Conze, Aresch Yavari, Dipl.-Kfm.
> Philipp Hermanns
> Aufsichtsratsvorsitzender: Wilhelm Dohmen
> --
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] How to resume hosted-engine --vm-start-paused

2017-01-25 Thread Simone Tiraboschi
On Wed, Jan 25, 2017 at 5:28 AM, Matt .  wrote:

> I'm trying to resume a pause started HostedEngine but I'm failing and
> I see lots of topics about some option --vm-resume being needed.
>
> "virsh resume HostedEngine" asks me for a username which is not set,
> so I'm wondering how to get this working.
>
>
you can try with
virsh -c qemu:///system?authfile=/etc/ovirt-hosted-engine/virsh_auth.conf
resume HostedEngine

but I agree that --vm-resume will be needed as well.


> Thanks!
>
> Matt
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Connect to Hosted-Engine using ----add-console-password --password=

2017-01-25 Thread Simone Tiraboschi
On Wed, Jan 25, 2017 at 4:51 AM, Matt .  wrote:

> Oops, I was connecting using Spice and needed VNC. Maybe it's nice to
> put this in the output or the help for the console command.
>

The console commands now directly points to the serial console and it's
intended to bu run on the host where the engine VM is running for
troubleshooting activities: it's text only but you don't need any graphical
subsystem on the host you are going to use.

Then you could still connect with remote viewer from an external machine
but you have to identify the right port and the protocol;
if you are using spice you need also to download the right certificates.
Could you please open an RFE to print out the list of command to connect
from a remote client with remote viewer in the output of hosted-engine
--console ?


>
> 2017-01-25 4:48 GMT+01:00 Matt . :
> > I'm figuring out how to connect to the hosted engine using virt-viewer
> > on a client, is there any more information about this ? You set the
> > console password but then ? on what port do we connect, and I assume
> > we connect the normal way for normal VM's Consoles ?
> >
> > Any info would be nice.
> >
> > Thanks.
> >
> > Matt
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Migrate hosted engine to a new storage domain

2017-01-27 Thread Simone Tiraboschi
On Thu, Jan 26, 2017 at 7:14 PM, Logan Kuhn 
wrote:

> I read through that and it makes sense.  It also said that moving the
> hosted engine's storage was a 4.1 feature, but I didn't see it on the
> release notes for RC 2.  Is it still planned for 4.1?  Do you have the BZ
> handy so that I can track it?
>
>
Yes, it's in.
It's tracked here:
https://bugzilla.redhat.com/show_bug.cgi?id=1240466
https://bugzilla.redhat.com/show_bug.cgi?id=1235200

We have still something to improve on corner cases if the hosts you want to
filter out in the restore process were not in maintenance mode when you
took the backup but the functionality is there.


> Logan
>
>
> On Thu, Jan 19, 2017 at 10:06 AM, Logan Kuhn 
> wrote:
>
>> Exactly what I needed.  Thank you
>>
>> On Thu, Jan 19, 2017 at 1:10 AM, Yedidyah Bar David 
>> wrote:
>>
>>> On Thu, Jan 19, 2017 at 4:01 AM, Logan Kuhn 
>>> wrote:
>>> > Hi
>>> >
>>> > We are planning on moving to a different storage solution and I'm
>>> curious,
>>> > is there a way to migrate the hosted engine's storage domain to the new
>>> > solution?  It's NFS currently and can be NFS on the new storage as
>>> well.
>>> >
>>> > From what I've read it looks like it should be possible to
>>> >
>>> > Take a full backup of the engine VM
>>> > Deploy another hosted engine VM with hosted-engine --deploy
>>> > Install/configure CentOS 7.3
>>> > Deploy new engine with engine-setup
>>> > Then restore the backup into the new VM.
>>> >
>>> > What I'm not sure of is if that backup will contain enough of it's
>>> data to
>>> > restore to a completely different storage domain?
>>> >
>>> > Also, the engine database is on a remote server, the data warehouse
>>> service
>>> > and all other aspects of the hosted engine reside on the VM.
>>>
>>> Please check the list archive:
>>>
>>> http://lists.ovirt.org/pipermail/users/2017-January/078739.html
>>>
>>> Best,
>>>
>>> >
>>> > Thanks,
>>> > Logan
>>> >
>>> > ___
>>> > Users mailing list
>>> > Users@ovirt.org
>>> > http://lists.ovirt.org/mailman/listinfo/users
>>> >
>>>
>>>
>>>
>>> --
>>> Didi
>>>
>>
>>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [ovirt-devel] Translate hint

2017-01-27 Thread Simone Tiraboschi
On Thu, Jan 26, 2017 at 11:50 AM, Gianluca Cecchi  wrote:

> On Thu, Jan 26, 2017 at 11:40 AM, Yuko Katabami 
> wrote:
>
>>
>>
>> You should notify the engineering team when you finished working.
>> Pulling translations from Zanata into the repo is a manual task, so
>> unless you ask the team, it will not be imported to the product.
>> Particularly with non-latest versions.
>>
>> Kind regards,
>>
>> Yuko
>>
>>>
>>>
> Thank you very much Yuko for the explanation! I'm going to test what you
> suggested.
> BTW: what should I do to "notify the engineering team"?
>
>
A message on de...@ovirt.org is enough


> Cheers,
> Gianluca
>
> ___
> Devel mailing list
> de...@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/devel
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Memory for Engine Machine

2017-02-20 Thread Simone Tiraboschi
On Mon, Feb 20, 2017 at 9:39 PM, Roy Golan  wrote:

>
>
> On Feb 20, 2017 9:35 PM, "Doug Ingham"  wrote:
>
> 16GB is just the recommended amount of memory. The more items your Engine
> has to manage, the more memory it will consume, so whilst it might not be
> using that amount of memory at the moment, it will do as you expand your
> cluster.
>
>
> It will never really need this amount of memory. There is no reason to
> configure the heap for 16gb.
>
>
>
> On 20 February 2017 at 16:22, FERNANDO FREDIANI  > wrote:
>
>> Hello folks
>>
>> I have a Engine dedicated machine running with 4GB of memory. It has been
>> working fine without any apparent issues.
>>
>> If I check the system memory usage it rarely goes over 1.5GB.
>>
>> But when I upgrade oVirt Engine it complains with the following message:
>> "[WARNING] Less than 16384MB of memory is available".
>>
>
> Please open a bug with all the details, we should address that.
>

Currently we are tuning the java heap size to 1/4 of the system memory.
On hosted-engine-setup side we propose to assign to the engine VM 16GB (so
java heap = 4GB) as recommended and 4GB as a minimum suggested value.



>
>> Why is all that required if the real usage doesn't show that need ? Or am
>> I missing anything ?
>>
>> Fernando Frediani
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>
>
>
> --
> Doug
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Attaching ISO to hosted engine for OS upgrade

2017-02-22 Thread Simone Tiraboschi
On Wed, Feb 22, 2017 at 8:04 PM, Chris Adams <c...@cmadams.net> wrote:

> Once upon a time, Simone Tiraboschi <stira...@redhat.com> said:
> > Then ee have a specific helper utility for 3.6/el6 -> 4.0/el7:
> > https://www.ovirt.org/develop/release-management/features/
> hosted-engine-migration-to-4-0/
>
> Ahh, that looks better.  I was looking at this:
>
> https://www.ovirt.org/documentation/migration-engine-36-to-40/
>
> which just kind of glosses over how to upgrade the OS. :)
>
> I do usually use my custom CentOS install (rather than the appliance);
> is there a way to do that?
>

No, not with that flow and not in 4.1 at all.
ovirt-engine-appliance is basically just a centos VM plus ovirt-engine.
Once deployed you can tune it as you need.


>
> Also, is it normally recommended to upgrade one major release at a time?
>

For the engine it's not just recommended, it's mandatory!


> In other words, aside from the engine CentOS6->7 step, would upgrading
> from 3.5 to 4.1 need to go through 3.6 and 4.0 along the way?
> --
> Chris Adams <c...@cmadams.net>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Guest Agent Running unconfined on Centos 7

2017-02-22 Thread Simone Tiraboschi
On Wed, Feb 22, 2017 at 10:05 PM, Michal Skrivanek 
wrote:

> > On 22 Feb 2017, at 16:46, Jiri Belka  wrote:
> >
> > - Original Message -
> >> From: "Alan Griffiths" 
> >> To: "Ovirt Users" 
> >> Sent: Friday, February 10, 2017 4:25:28 PM
> >> Subject: [ovirt-users] Guest Agent Running unconfined on Centos 7
> >>
> >> Hi,
> >>
> >> I'm running ovirt-guest-agent from Centos 7 EPEL and I notice that it's
> >> running unconfined rather than within its own domain.
> >>
> >> I see there is a rhev_agentd_exec_t
>
> That sound suspicious on its own. Are you sure you haven't mixed rhev
> and ovirt agents in the same guest at some point? Restoring selinux
> context doesn't help?
>
>
Here the same:
[root@c72he20170222h1 ~]# yum list installed | grep rhev
fence-agents-rhevm.x86_64 4.0.11-47.el7_3.2
 @updates
[root@c72he20170222h1 ~]# yum list installed | grep ovirt-guest-agent
ovirt-guest-agent-common.noarch   1.0.12-4.el7
@epel
[root@c72he20170222h1 ~]# ps auxZ  | grep guest-agent
system_u:system_r:unconfined_service_t:s0 ovirtag+ 732 0.2  0.6 441796
36036 ? Ssl  16:59   0:46 /usr/bin/python
/usr/share/ovirt-guest-agent/ovirt-guest-agent.py
unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023 root 6938 0.0  0.0
112648 964 pts/0 S+ 22:31   0:00 grep --color=auto guest-agent
[root@c72he20170222h1 ~]# semanage fcontext -l | grep rhev_agentd
/var/log/rhev-agent(/.*)?  all files
 system_u:object_r:rhev_agentd_log_t:s0
/var/log/ovirt-guest-agent(/.*)?   all files
 system_u:object_r:rhev_agentd_log_t:s0
/usr/lib/systemd/system/ovirt-guest-agent.*regular file
system_u:object_r:rhev_agentd_unit_file_t:s0
/var/run/rhev-agentd\.pid  regular file
system_u:object_r:rhev_agentd_var_run_t:s0
/usr/share/ovirt-guest-agent   regular file
system_u:object_r:rhev_agentd_exec_t:s0
/var/run/ovirt-guest-agent\.pidregular file
system_u:object_r:rhev_agentd_var_run_t:s0
/usr/share/rhev-agent/rhev-agentd\.py  regular file
system_u:object_r:rhev_agentd_exec_t:s0
/usr/share/rhev-agent/LockActiveSession\.pyregular file
system_u:object_r:rhev_agentd_exec_t:s0
/usr/share/ovirt-guest-agent/LockActiveSession\.py regular file
system_u:object_r:rhev_agentd_exec_t:s0




> >> type, which I attempted to assign to
> >> ovirt-guest-agent.py but it still starts up as unconfined. Is there a
> >> supported process for getting ovirt-guest into its own domain? Or a
> reason
> >> why it's not possible?
> >>
> >> Thanks,
> >>
> >> Alan
> >
> > Hm, it seems many ovirt services run unconfined. For ovirt GA, it seems
> > there's missing glue between systemd -> python -> GA script.
> >
> > Vinzenz, any idea?
> >
> > j.
> > ___
> > Users mailing list
> > Users@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> >
> >
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Attaching ISO to hosted engine for OS upgrade

2017-02-22 Thread Simone Tiraboschi
On Wed, Feb 22, 2017 at 5:33 PM, Chris Adams  wrote:

> I'm working on upgrading an oVirt 3.5 setup.  The physical hosts are
> running CentOS 7, but the hosted engine is CentOS 6.  The upgrade notes
> are "back up the engine, upgrade/reinstall the OS, then restore", but I
> can't see how to actually install CentOS 7 on the engine.
>
> Am I supposed to re-run "hosted-engine --deploy"?  Wouldn't that try to
> re-register the physical hosts, or can I interrupt it to restore the
> backup?
>
> Or, is there a way to just attach an install ISO to the engine VM and
> boot from that?
>

On 3.6 we release the engine also for el6 so for 3.5/el6 to 3.6/el6 you
have just to:
- set global maintenance mode,
- add the 3.6 repo to the engine VM,
- run yum update ovirt-*-setup* on the engine VM
- run engine-setup
- upgrade the host to 3.6 setting maintenance mode one host at a time
- raise the cluster compatibility level

Then ee have a specific helper utility for 3.6/el6 -> 4.0/el7:
https://www.ovirt.org/develop/release-management/features/hosted-engine-migration-to-4-0/



> --
> Chris Adams 
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] high number of IOPS from ovirt hosts

2017-02-12 Thread Simone Tiraboschi
On Fri, Feb 10, 2017 at 12:46 PM, Andrea Fagiani <
andrea.fagi...@immobiliare.it> wrote:

> Hi,
>
> We've got a 8-host cluster running oVirt 4.1; each host is connected via
> iscsi to a EQL 4100 storage (2x).
> Most of the VMs use barely any IOPS, however on our storage controller
> we've noticed a constant steam of reads, about 2000 IOPS; any idea what
> might cause the ovirt hosts to hammer on the storage so much?
>
> Looking at iotop, it seems that systemd-udevd is being activated every few
> seconds (and its IO% is usually fairly high) so I assume this might be
> somehow related.
>

mmmh this could be an hint: how many LUNs are you exposing on the iSCSI
portal used for the hosted-engine LUN?


>
> Thanks,
> Andrea
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] questions about migrating to self-hosted engine from bare-metal

2017-02-14 Thread Simone Tiraboschi
On Mon, Feb 13, 2017 at 7:41 PM, cmc  wrote:

> Hi Yaniv,
>
> Log attached. There's a bit in there where I'd tried setting the host
> record to point to the new host as an experiment, but I changed it
> back, and it still fails with the same error every time.
>
>
It fails due to this one:
2017-02-13 13:07:45,812 ERROR (vm/642a0b9a) [virt.vm]
(vmId='642a0b9a-49fc-4ccc-8976-f6685953d0e8') The vm start process failed
(vm:616)
Traceback (most recent call last):
  File "/usr/share/vdsm/virt/vm.py", line 552, in _startUnderlyingVm
self._run()
  File "/usr/share/vdsm/virt/vm.py", line 1994, in _run
self._connection.createXML(domxml, flags),
  File "/usr/lib/python2.7/site-packages/vdsm/libvirtconnection.py", line
123, in wrapper
ret = f(*args, **kwargs)
  File "/usr/lib/python2.7/site-packages/vdsm/utils.py", line 941, in
wrapper
return func(inst, *args, **kwargs)
  File "/usr/lib64/python2.7/site-packages/libvirt.py", line 3782, in
createXML
if ret is None:raise libvirtError('virDomainCreateXML() failed',
conn=self)
libvirtError: invalid argument: could not find capabilities for arch=x86_64
domaintype=kvm
2017-02-13 13:07:45,852 INFO  (vm/642a0b9a) [virt.vm]
(vmId='642a0b9a-49fc-4ccc-8976-f6685953d0e8') Changed state to Down:
invalid argument: could not find capabilities for arch=x86_64
domaintype=kvm  (code=1) (vm:1199)

Is your CPU properly configured for virtualization? Are you running on a
nested env?



> Thanks,
>
> Cam
>
>
> On Mon, Feb 13, 2017 at 6:22 PM, Yaniv Kaul  wrote:
> >
> > On Mon, Feb 13, 2017 at 6:26 PM, cmc  wrote:
> >>
> >>
> >> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-
> setup-20170213141937-0wgc31.log
> >
> >
> > Anything in the above log?
> > Y.
> >
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] questions about migrating to self-hosted engine from bare-metal

2017-02-14 Thread Simone Tiraboschi
On Tue, Feb 14, 2017 at 2:36 PM, cmc  wrote:

> Hi Simone,
>
> >>
> >
> > It fails due to this one:
> > 2017-02-13 13:07:45,812 ERROR (vm/642a0b9a) [virt.vm]
> > (vmId='642a0b9a-49fc-4ccc-8976-f6685953d0e8') The vm start process
> failed
> > (vm:616)
> > Traceback (most recent call last):
> >   File "/usr/share/vdsm/virt/vm.py", line 552, in _startUnderlyingVm
> > self._run()
> >   File "/usr/share/vdsm/virt/vm.py", line 1994, in _run
> > self._connection.createXML(domxml, flags),
> >   File "/usr/lib/python2.7/site-packages/vdsm/libvirtconnection.py",
> line
> > 123, in wrapper
> > ret = f(*args, **kwargs)
> >   File "/usr/lib/python2.7/site-packages/vdsm/utils.py", line 941, in
> > wrapper
> > return func(inst, *args, **kwargs)
> >   File "/usr/lib64/python2.7/site-packages/libvirt.py", line 3782, in
> > createXML
> > if ret is None:raise libvirtError('virDomainCreateXML() failed',
> > conn=self)
> > libvirtError: invalid argument: could not find capabilities for
> arch=x86_64
> > domaintype=kvm
> > 2017-02-13 13:07:45,852 INFO  (vm/642a0b9a) [virt.vm]
> > (vmId='642a0b9a-49fc-4ccc-8976-f6685953d0e8') Changed state to Down:
> invalid
> > argument: could not find capabilities for arch=x86_64 domaintype=kvm
> > (code=1) (vm:1199)
> >
> > Is your CPU properly configured for virtualization? Are you running on a
> > nested env?
> >
>
> Yes, I assume it should be configured for virtualisation, since this
> VM I'm creating to replace the bare-metal engine is running on one of
> the two hosts that make up the cluster, and host 30 other running VMs
> currently. Should it not run on the same cluster perhaps?
>
> My steps are:
>
> 1. Create a VM on the cluster. Install the ovirt release rpm, install
> the packages ovirt-hosted-engine-setup and ovirt-engine-appliance
> 2. run hosted-engine --deploy
>
> ...and that is as far as I have gotten so far.
>

OK, a bit of confusion here:
ovirt-hosted-engine-setup is going to create a VM for you based on
ovirt-engine-appliance (with the engine).
If you manually create a VM (L1) on your physical host (L0) to run
hosted-engine-setup there, hosted-engine-setup will create a VM (L2) for
the engine running inside the L1 VM and this requires nested virtualization
support which I think it's not enabled by default on oVirt hosts.
But the point is that you simply have to run ovirt-hosted-engine-setup on
your physical host and not on a VM otherwise all the HA mechanism of
hosted-engine will make no sense if nothing is bringing up your virtual
hosts for you.

hosted-engine-setup requires an host with no others running VMs so, if it's
already managed my an engine, move it to maintenance and remove it from
your cluster.



>
> Thanks,
>
> Cam
>
> >
> >>
> >> Thanks,
> >>
> >> Cam
> >>
> >>
> >> On Mon, Feb 13, 2017 at 6:22 PM, Yaniv Kaul  wrote:
> >> >
> >> > On Mon, Feb 13, 2017 at 6:26 PM, cmc  wrote:
> >> >>
> >> >>
> >> >>
> >> >> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-
> setup-20170213141937-0wgc31.log
> >> >
> >> >
> >> > Anything in the above log?
> >> > Y.
> >> >
> >>
> >> ___
> >> Users mailing list
> >> Users@ovirt.org
> >> http://lists.ovirt.org/mailman/listinfo/users
> >>
> >
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] OVN routing and firewalling in oVirt

2017-02-16 Thread Simone Tiraboschi
On Thu, Feb 16, 2017 at 2:20 PM, Gianluca Cecchi 
wrote:

> Hello,
> how do we manage routing between different OVN networks in oVirt?
> And between OVN networks and physical ones?
>

Take a look at this blog post:
http://blog.spinhirne.com/2016/09/the-ovn-gateway-router.html

The good news is that a distributed NAT is going to be introduced with OVN
2.7:
https://patchwork.ozlabs.org/patch/726766/


> Based on architecture read here:
> http://openvswitch.org/support/dist-docs/ovn-architecture.7.html
>
> I see terms for logical routers and gateway routers respectively but how
> to apply to oVirt configuration?
> Do I have to choose between setting up a specialized VM or a physical one:
> is it applicable/advisable to put on oVirt host itself the gateway
> functionality?
>
> Is there any security policy (like security groups in Openstack) to
> implement?
>
> Thanks,
> Gianluca
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] OVN routing and firewalling in oVirt

2017-02-16 Thread Simone Tiraboschi
On Thu, Feb 16, 2017 at 4:49 PM, Gianluca Cecchi <gianluca.cec...@gmail.com>
wrote:

> On Thu, Feb 16, 2017 at 2:26 PM, Simone Tiraboschi <stira...@redhat.com>
> wrote:
>
>>
>>
>> On Thu, Feb 16, 2017 at 2:20 PM, Gianluca Cecchi <
>> gianluca.cec...@gmail.com> wrote:
>>
>>> Hello,
>>> how do we manage routing between different OVN networks in oVirt?
>>> And between OVN networks and physical ones?
>>>
>>
>> Take a look at this blog post:
>> http://blog.spinhirne.com/2016/09/the-ovn-gateway-router.html
>>
>
> Great!
> Actually using the previous blog post of the series:
> http://blog.spinhirne.com/2016/09/an-introduction-to-ovn-routing.html
>

It was something I wished to show this Monday in the workshop but we were
really out of time!


>
>
> I was able to complete routing between two different oVirt subnets:
>
> In oVirt I have previously created:
>
> ovn_net1 network with subnet subn1 (defined as 172.16.10.0/24 with gw
> 172.16.10.1)
> so that ip usable range is from 172.16.10.1 to 172.16.10.254
>
> ovn_net2 network with subnet subn2 (defined as 192.168.10.0/24 with gw
> 192.168.10.1)
> so that ip usable range is from 192.168.10.1 to 192.168.10.254
>
> I have to VMs defined on the two subnets:
> vm1 172.16.10.2
> vm2 192.168.10.101
>
> on central server (that is my engine)
> # define the new logical switches
> # no, already created from inside oVirt: they are ovn_net1 and ovn_net2
>
> # add the router
> ovn-nbctl lr-add net1net2
>
> # create router port for the connection to net1
> ovn-nbctl lrp-add net1net2 net1 02:ac:10:ff:01:29 172.16.10.1/24
>
> # create the net1 switch port for connection to net1net2
> ovn-nbctl lsp-add ovn_net1 net1-net1net2
> ovn-nbctl lsp-set-type net1-net1net2 router
> ovn-nbctl lsp-set-addresses net1-net1net2 02:ac:10:ff:01:29
> ovn-nbctl lsp-set-options net1-net1net2 router-port=net1
>
> # create router port for the connection to net2
> ovn-nbctl lrp-add net1net2 net2 02:ac:10:ff:01:93 192.168.10.1/24
>
> # create the net2 switch port for connection to net1net2
> ovn-nbctl lsp-add ovn_net2 net2-net1net2
> ovn-nbctl lsp-set-type net2-net1net2 router
> ovn-nbctl lsp-set-addresses net2-net1net2 02:ac:10:ff:01:93
> ovn-nbctl lsp-set-options net2-net1net2 router-port=net2
>
> # show config
> ovn-nbctl show
>
> [root@ractorshe ~]# ovn-nbctl show
> switch 38cca50c-e8b2-43fe-b585-2ee815191939 (ovn_net1)
> port 5562d95d-060f-4c64-b535-0e460ae6aa5a
> addresses: ["00:1a:4a:16:01:52 dynamic"]
> port 87fea70a-583b-4484-b72b-030e2f175aa6
> addresses: ["00:1a:4a:16:01:53 dynamic"]
> port net1-net1net2
> addresses: ["02:ac:10:ff:01:29"]
> port 99f619fc-29d2-4d40-8c28-4ce9291eb97a
> addresses: ["00:1a:4a:16:01:51 dynamic"]
> switch 6a0e7a92-8edc-44dd-970a-2b1f5c07647d (ovn_net2)
> port net2-net1net2
> addresses: ["02:ac:10:ff:01:93"]
> port 9b7a79a3-aa38-43b1-abd4-58370171755e
> addresses: ["00:1a:4a:16:01:54 dynamic"]
> router 59d79312-a434-4150-be46-285a9f37df8d (net1net2)
> port net2
> mac: "02:ac:10:ff:01:93"
> networks: ["192.168.10.1/24"]
> port net1
> mac: "02:ac:10:ff:01:29"
> networks: ["172.16.10.1/24"]
> [root@ractorshe ~]#
>
> And now vm1 is able to ping both the gateways ip on subn1 and subn2 and to
> ssh into vm2
> It remains a sort of spof the fact of the central ovn server, where the
> logical router lives... but for initial testing it is ok
>

Are you sure? did you tried bringing it down?

AFAIU, OVN is already providing distributed routing since 2.6: if the node
where you have the oVirt OVN provider and the OVN controller with
northbound and southbound DB is down you cannot edit logical networks but
the existing flows should still be there.



>
> Thanks again,
> Gianluca
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Update NFS storage of the HE server

2017-01-23 Thread Simone Tiraboschi
On Mon, Jan 23, 2017 at 4:20 PM, Claude Duropcher <
claude.duroc...@cptaq.gouv.qc.ca> wrote:

> Hi,
>
> We need to update/reboot the NFS server on wich the hosted engine is
> installed. This server also hosts a couple of extra storage domains (ISO,
> Export and a iScsi domain). I'm wondering if this procedure is complete for
> an oVirt 4.0 installation :
>
> 1-set the environment to global maintenance
>
> 2-set the extra domains to maintenance mode
>
> 3-shutdown the hosted engine
>
> 4-update/reboot the storage server
>
> 5-reboot the hosted engine
>
> 6-reactivate extra domains
>
> 7-disable global maintenance mode
>
> I'm wondering what's happening after step 4 and how ovirt will react when
> the NFS storage disapear and come back.
>
>
ovirt-ha-agent will try to reconnect the shared storage if hosted-engine
for all the time; depending on different factors, sanlock will fail
updating the hosted-engine lease ans this can bring the watchdog to reboot
the host.
Setting also local maintenance mode on the involved hosts is a good idea.



> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] HostedEngine with HA

2016-08-19 Thread Simone Tiraboschi
On Thu, Aug 18, 2016 at 6:38 PM, Carlos Rodrigues <c...@eurotux.com> wrote:

> On Thu, 2016-08-18 at 17:45 +0200, Simone Tiraboschi wrote:
>
> On Thu, Aug 18, 2016 at 5:43 PM, Carlos Rodrigues <c...@eurotux.com>
> wrote:
>
>
> I increase hosted_engine disk space to 160G. How do i force to create
> OVF_STORE.
>
>
> I think that restarting the engine on the engine VM will trigger it
> although I'm not sure that it was a size issue.
>
>
> I found to OVF_STORE on another storage domain with "Domain Type" "Data
> (Master)"
>
>
Each storage domain has its own OVF_STORE volumes; you should get them also
on the hosted-engine storage domain.
Not really sure about how to trigger it again; adding Roy here.



> ​
>
>
>
> Regards,
> Carlos Rodrigues
>
> On Thu, 2016-08-18 at 12:14 +0100, Carlos Rodrigues wrote:
>
>
> On Thu, 2016-08-18 at 12:34 +0200, Simone Tiraboschi wrote:
>
>
>
> On Thu, Aug 18, 2016 at 12:11 PM, Carlos Rodrigues <c...@eurotux.co
> m>
> wrote:
>
>
>
>
> On Thu, 2016-08-18 at 11:53 +0200, Simone Tiraboschi wrote:
>
>
>
>
>
>
> On Thu, Aug 18, 2016 at 11:50 AM, Carlos Rodrigues <cmar@eurotu
> x.
> com>
> wrote:
>
>
>
>
> On Thu, 2016-08-18 at 11:42 +0200, Simone Tiraboschi wrote:
>
>
>
>
> On Thu, Aug 18, 2016 at 11:25 AM, Carlos Rodrigues <cmar@eu
> ro
> tux.
> com> wrote:
>
>
>
>
>
> On Thu, 2016-08-18 at 11:04 +0200, Simone Tiraboschi
> wrote:
>
>
>
>
>
> On Thu, Aug 18, 2016 at 10:36 AM, Carlos Rodrigues
> <cmar@
> euro
> tux.com>
> wrote:
>
>
>
>
>
>
> On Thu, 2016-08-18 at 10:27 +0200, Simone Tiraboschi
> wrote:
>
>
>
>
>
>
> On Thu, Aug 18, 2016 at 10:22 AM, Carlos Rodrigues
> <cmar@
> eurotux.
> com>
> wrote:
>
>
>
>
>
>
>
> On Thu, 2016-08-18 at 08:54 +0200, Simone
> Tiraboschi
> wrote:
>
>
>
>
>
>
>
> On Tue, Aug 16, 2016 at 12:53 PM, Carlos
> Rodrigues  mar@euro
> tux.
> com>
> wrote:
>
>
>
>
>
>
>
>
> On Sun, 2016-08-14 at 14:22 +0300, Roy Golan
> wrote:
>
>
>
>
>
>
>
>
>
>
> On 12 August 2016 at 20:23, Carlos
> Rodrigues
>  r@eurotu
> x.co
> m>
> wrote:
>
>
>
>
>
>
>
>
> Hello,
>
> I have one cluster with two hosts with
> power
> management
> correctly
> configured and one virtual machine with
> HostedEngine
> over
> shared
> storage with FiberChannel.
>
> When i shutdown the network of host with
> HostedEngine
> VM,  it
> should be
> possible the HostedEngine VM migrate
> automatically to
> another
> host?
>
> migrate on which network?
>
>
>
>
>
>
>
>
> What is the expected behaviour on this HA
> scenario?
>
>
> After a few minutes your vm will be
> shutdown
> by
> the High
> Availability
> agent, as it can't see network, and started
> on
> another
> host.
>
>
>
> I'm testing this scenario and after shutdown
> network, it
> should
> be
> expected that agent shutdown ha and started
> on
> another
> host,
> but
> after
> couple minutes nothing happens and on host
> with
> network we
> getting
> the
> following messages:
>
> Aug 16 11:44:08 ied-
> blade11.install.eurotux.local
> ovirt-ha-
> agent[2779]:
> ovirt-ha-agent
> ovirt_hosted_engine_ha.agent.hosted_engine.Ho
> st
> edEn
> gine.con
> fig
> ERROR
> Unable to get vm.conf from OVF_STORE, falling
> back
> to
> initial
> vm.conf
>
> I think the HA agent its trying to get vm
> configuration but
> some
> how it
> can't get vm.conf to start VM.
>
>
> No, this is a different issues.
> In 3.6 we added a feature to let the engine
> manage
> also the
> engine VM
> itself; ovirt-ha-agent will pickup the latest
> engine
> VM
> configuration
> from the OVF_STORE which is managed by the
> engine.
> If something goes wrong, ovirt-ha-agent could
> fallback to the
> initial
> (bootstrap time) vm.conf. This will normally
> happen
> till you
> add
> your
> first regular storage domain and the engine
> imports
> the
> engine
> VM.
>
>
> But i already have my first storage domain and
> storage
> engine
> domain
> and already imported engine VM.
>
> I'm using 4.0 version.
>
>
> This seams an issue, can you please share your
> /var/log/ovirt-hosted-engine-ha/agent.log ?
>
>
> I sent it in attachment.
>
>
> Nothing

Re: [ovirt-users] HostedEngine with HA

2016-08-19 Thread Simone Tiraboschi
On Fri, Aug 19, 2016 at 12:07 PM, Carlos Rodrigues <c...@eurotux.com> wrote:
> On Fri, 2016-08-19 at 10:47 +0100, Carlos Rodrigues wrote:
>> On Fri, 2016-08-19 at 11:36 +0200, Simone Tiraboschi wrote:
>> >
>> >
>> >
>> > On Fri, Aug 19, 2016 at 11:29 AM, Carlos Rodrigues <c...@eurotux.co
>> > m>
>> > wrote:
>> > >
>> > > After night, the OVF_STORE it was created:
>> > >
>> >
>> > It's quite strange that it got so long but now it looks fine.
>> >
>> > If the ISO_DOMAIN that I see in your screenshot is served by the
>> > engine VM itself, I suggest to remove it and export from an
>> > external
>> > server.
>> > Serving the ISO storage domain from the engine VM itself is not a
>> > good idea since when the engine VM is down you can experiment long
>> > delays before getting the engine VM restarted due to the
>> > unavailable
>> > storage domain.
>>
>> Ok, thank you for advice.
>>
>> Now, apparently is all ok. I'll do more tests with HA and any issue
>> i'll tell you.
>>
>> Thank you for your support.
>>
>> Regards,
>> Carlos Rodrigues
>>
>
> I shutdown the network of host with engine VM and i expected that other
> host fence the host and start engine VM but i don't see any fence
> action and the "free" host keep trying to start VM but get and error of
> sanlock
>
> Aug 19 11:03:03 ied-blade11.install.eurotux.local kernel: qemu-kvm:
> sending ioctl 5326 to a partition!
> Aug 19 11:03:03 ied-blade11.install.eurotux.local kernel: qemu-kvm:
> sending ioctl 80200204 to a partition!
> Aug 19 11:03:03 ied-blade11.install.eurotux.local kvm[7867]: 1 guest
> now active
> Aug 19 11:03:03 ied-blade11.install.eurotux.local sanlock[884]: 2016-
> 08-19 11:03:03+0100 1023 [903]: r3 paxos_acquire owner 1 delta 1 9
> 245502 alive
> Aug 19 11:03:03 ied-blade11.install.eurotux.local sanlock[884]: 2016-
> 08-19 11:03:03+0100 1023 [903]: r3 acquire_token held error -243
> Aug 19 11:03:03 ied-blade11.install.eurotux.local sanlock[884]: 2016-
> 08-19 11:03:03+0100 1023 [903]: r3 cmd_acquire 2,9,7862 acquire_token
> -243 lease owned by other host
> Aug 19 11:03:03 ied-blade11.install.eurotux.local libvirtd[1369]:
> resource busy: Failed to acquire lock: error -243
> Aug 19 11:03:03 ied-blade11.install.eurotux.local kernel: ovirtmgmt:
> port 2(vnet0) entered disabled state
> Aug 19 11:03:03 ied-blade11.install.eurotux.local kernel: device vnet0
> left promiscuous mode
> Aug 19 11:03:03 ied-blade11.install.eurotux.local kernel: ovirtmgmt:
> port 2(vnet0) entered disabled state
> Aug 19 11:03:03 ied-blade11.install.eurotux.local kvm[7885]: 0 guests
> now active
> Aug 19 11:03:03 ied-blade11.install.eurotux.local systemd-
> machined[7863]: Machine qemu-4-HostedEngine terminated.

Maybe you hit this one:
https://bugzilla.redhat.com/show_bug.cgi?id=1322849


Can you please check it as described in comment 28 and eventually
apply the workaround in comment 18?



>> > > Regards,
>> > > Carlos Rodrigues
>> > >
>> > > On Fri, 2016-08-19 at 08:29 +0200, Simone Tiraboschi wrote:
>> > > >
>> > > >
>> > > >
>> > > > On Thu, Aug 18, 2016 at 6:38 PM, Carlos Rodrigues <cmar@eurotux
>> > > > .c
>> > > > om> wrote:
>> > > > >
>> > > > > On Thu, 2016-08-18 at 17:45 +0200, Simone Tiraboschi wrote:
>> > > > > >
>> > > > > > On Thu, Aug 18, 2016 at 5:43 PM, Carlos Rodrigues <cmar@eur
>> > > > > > ot
>> > > > > > ux.com> wrote:
>> > > > > > >
>> > > > > > >
>> > > > > > > I increase hosted_engine disk space to 160G. How do i
>> > > > > > > force
>> > > > > > > to create
>> > > > > > > OVF_STORE.
>> > > > > >
>> > > > > > I think that restarting the engine on the engine VM will
>> > > > > > trigger it
>> > > > > > although I'm not sure that it was a size issue.
>> > > > > >
>> > > > >
>> > > > > I found to OVF_STORE on another storage domain with "Domain
>> > > > > Type" "Data (Master)"
>> > > > >
>> > > > >
>> > > >
>> > > >

Re: [ovirt-users] Hosted Engine 3.6 Questions

2016-08-22 Thread Simone Tiraboschi
On Thu, Aug 18, 2016 at 6:38 PM, C Williams  wrote:
> Hello,
>
> We have an installation with 2 filers on CentOS 7.2 with the hosted engine
> appliance. We use iscsi for our backend storage.
>
> We have some issues
>
> We had to dedicate an entire dedicated storage domain for the sole use of
> the Hosted Engine appliance. Is this required ?

Yes, it is.

> Or can the Hosted Engine
> appliance co-exist on another storage domain with VMs ?  I have not been
> able to migrate the Hosted Engine VM to other storage. Because of this
> limitation and to be efficient with our storage use, we asked our storage
> admin to make a small iscsi LUN 20GB for the storage domain for the
> appliance. However, we are constantly getting errors regarding low disk
> space for this domain even though it has only the 10GB appliance.

A storage domain on block devices requires 5 or 6 GB of ancillary data
due to extents size and, if I'm not wrong, there is a low disk space
threshold at about 4 GB; using slightly larger LUN will help.

> Could this size problem be related to my problems with importing VMWare VMs
> into our larger (Multi TB) storage using the Import tool in 3.6.
>
> Thank You
>
>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Restoring Hosted Engine to failed

2017-02-26 Thread Simone Tiraboschi
On Sun, Feb 26, 2017 at 8:02 AM, Yedidyah Bar David  wrote:

> On Fri, Feb 24, 2017 at 6:34 PM, Logan Kuhn 
> wrote:
> > Hi
> >
> > I'm in the process of trying to restore our previous hosted engine onto
> new,
> > permanent, storage.  I've followed this guide:
> > https://www.ovirt.org/develop/developer-guide/engine/
> migrate-to-hosted-engine/


The point is that the DB you restored contains references to the
hosted-engine storage domain and to the engine VM you used on your initial
hosted-engine env.
You have to clean up these references to let the engine correctly import
the new hosted-engine env.
With 4.1 we introduced two new options to engine-backup in order to make
correctly working for a backup took on a previous hosted engine env, they
are:
 --he-remove-storage-vm Removes the hosted-engine storage
domain, all its entities and the hosted-engine VM during restore.
 --he-remove-hosts  Removes all the hosted-engine hosts
during restore.

Hosted-engine host should be redeployed as well since they point to the
previous env as well.

The correct path to migrate from a previous hosted-engine env to a new one
is:
- set global maintenance mode on the previous env
- set maintenance mode for the hosted engine hosts on the previous engine
- take the backup of the previous engine
Follow
https://www.ovirt.org/develop/developer-guide/engine/migrate-to-hosted-engine/
adding  '--he-remove-storage-vm  --he-remove-hosts ' to the DB restore
command.



> > It went pretty smoothly up until the hosted engine deploy tried to add
> the
> > host to the environment. The error was Unable to add HEVM to the manager.
> > However, I can login to the environment just fine, but it does not have
> the
> > HEVM and the host it's currently running on is down.
> >
> > I've attached the hosted engine setup log.  Let me know if there are any
> > other logs that will be helpful
>
> Please check/share also: From the engine:
> /var/log/ovirt-engine/engine.log
> /var/log/ovirt-engine/host-deploy/*
>
> From the host:
> /var/log/vdsm/vdsm.log
> /tmp/ovirt* (if found. If all is ok they are copied to engine and removed)
>
> Best,
>
> >
> > Logan
> >
> > ___
> > Users mailing list
> > Users@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> >
>
>
>
> --
> Didi
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] 3.5->3.6 did not import hosted engine storage domain

2017-02-26 Thread Simone Tiraboschi
On Fri, Feb 24, 2017 at 8:24 PM, Chris Adams  wrote:

> So, on to my next upgrade issue (sorry for all the questions and thanks
> for everybody's help)...  I upgraded my test cluster from 3.5 to 3.6
> (latest version of each, all on CentOS 7 except the engine on CentOS 6).
> Now I'm working on the next step, upgrading to 4.0 and migrating the HE
> to the appliance.
>
> When I went from 3.5 to 3.6, I ended up with an fhanswers.conf in the
> shared storage that only contained "None"; I fixed that based on some
> mailing list messages (but just mentioning it in case it could be
> related).
>
> My problem is that the hosted engine storage domain did not get imported
> into the engine DB, so I can't proceed with "hosted-engine
> --upgrade-appliance".  I didn't see any errors, so I'm not sure how that
> happened.  I'm also not sure how to fix that.
>
> Suggestions?
>

Can you please attach your engine.log ?


> --
> Chris Adams 
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] 3.5->3.6 did not import hosted engine storage domain

2017-03-01 Thread Simone Tiraboschi
On Wed, Mar 1, 2017 at 5:04 PM, Chris Adams <c...@cmadams.net> wrote:

> Once upon a time, Simone Tiraboschi <stira...@redhat.com> said:
> > On Wed, Mar 1, 2017 at 3:19 PM, Chris Adams <c...@cmadams.net> wrote:
> > > However, now I'm back into the catch-22 of 3.6.7+hosted engine: the
> > > cluster compatibility level can't be raised while there's a running VM,
> > > and that includes the hosted engine.
> >
> > Please see this one:
> > https://bugzilla.redhat.com/show_bug.cgi?id=1364557
> >
> > Simply define 'InClusterUpgrade' scheduling policy on the HE VM cluster
>
> I first tried setting the policy, but got "Error while executing action:
> The set cluster compatibility version does not allow mixed major host OS
> versions. Can not start the cluster upgrade."; I guess this is because
> my hosts are CentOS 7 and the engine is CentOS 6?
>

This is not an issue, are you sure that all the hosts are el7 based?


>
> I tried changing the engine config to skip that check from comment 10
> step 3, but got:
> - Can not start cluster upgrade mode, see below for details:
> - VM HostedEngine with id 4a035efd-a041-4e46-84db-01cf79400913 is
>   configured to be not migratable.
>
> I did the SQL update from comment 1, and then I could set the policy.
>
> However, I still can't change the cluster compatibility version.
> --
> Chris Adams <c...@cmadams.net>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Mirgration issues

2017-03-02 Thread Simone Tiraboschi
Please take care that directly moving away or restarting VMs under oVirt
hood could cause conflicts with potentially dangerous results:
https://bugzilla.redhat.com/show_bug.cgi?id=1419649#c11

On Wed, Mar 1, 2017 at 9:40 AM, Arman Khalatyan  wrote:

> to use virsh you can set password:
>
>  saslpasswd2 -a libvirt foo (foo is some username)
>
> then use virsh with that username as usual
>
>
> On Thu, Feb 23, 2017 at 8:02 PM, Sven Achtelik 
> wrote:
>
>> Hi All,
>>
>>
>>
>> so sorry for not seeing this. I used vdsCLient just because I didn’t know
>> how to get virsh working. (It asked for user/password) Now that I managed
>> to get this working things look different
>>
>>
>>
>> [root@ovirt-node02 ~]# vdsClient -s localhost list table
>>
>> e051b38c-fd63-40f0-8d64-26c12ff7b880  41294  HostedEngine
>> Up   172.16.1.9
>>
>> [root@ovirt-node02 ~]# virsh -r list --all
>>
>>  IdName   State
>>
>> 
>>
>> 1 HostedEngine   running
>>
>> - data_p   shut off
>>
>>
>>
>> How could this vm persist over reboots and reinstalls? And how would I
>> remove this ?
>>
>>
>>
>>
>>
>> *Von:* Michal Skrivanek [mailto:michal.skriva...@redhat.com]
>> *Gesendet:* Donnerstag, 23. Februar 2017 17:35
>> *An:* Sven Achtelik ; Martin Sivak <
>> msi...@redhat.com>
>> *Cc:* Arman Khalatyan ; users 
>>
>> *Betreff:* Re: [ovirt-users] Mirgration issues
>>
>>
>>
>>
>>
>> On 23 Feb 2017, at 15:04, Sven Achtelik  wrote:
>>
>>
>>
>> Did that twice and it didn’t change anything.
>>
>>
>>
>> *Von:* Arman Khalatyan [mailto:arm2...@gmail.com ]
>> *Gesendet:* Donnerstag, 23. Februar 2017 15:02
>> *An:* Sven Achtelik 
>> *Cc:* Yanir Quinn ; users 
>> *Betreff:* Re: [ovirt-users] Mirgration issues
>>
>>
>>
>> engine gui.
>>
>>
>>
>> On Thu, Feb 23, 2017 at 1:46 PM, Sven Achtelik 
>> wrote:
>>
>> Do you mean just reinstalling from the Engine gui or reinstalling it
>> completely including the OS?
>>
>>
>>
>> *Von:* Arman Khalatyan [mailto:arm2...@gmail.com]
>> *Gesendet:* Donnerstag, 23. Februar 2017 13:45
>>
>>
>> *An:* Sven Achtelik 
>> *Cc:* Yanir Quinn ; users 
>> *Betreff:* Re: [ovirt-users] Mirgration issues
>>
>>
>>
>> Just random thought: try to reinstall the "bad" host:)
>>
>>
>>
>> On Thu, Feb 23, 2017 at 1:32 PM, Sven Achtelik 
>> wrote:
>>
>> Yes, all hosts are identical and going through the values there is no
>> difference between them.
>>
>>
>>
>> [root@ovirt-node02 ~]# systemctl status vdsmd -l
>>
>> ● vdsmd.service - Virtual Desktop Server Manager
>>
>>Loaded: loaded (/usr/lib/systemd/system/vdsmd.service; enabled;
>> vendor preset: enabled)
>>
>>Active: active (running) since Tue 2017-02-21 08:00:54 CST; 1 day 22h
>> ago
>>
>>   Process: 3571 ExecStartPre=/usr/libexec/vdsm/vdsmd_init_common.sh
>> --pre-start (code=exited, status=0/SUCCESS)
>>
>> Main PID: 3742 (vdsm)
>>
>>CGroup: /system.slice/vdsmd.service
>>
>>├─3742 /usr/bin/python2 /usr/share/vdsm/vdsm
>>
>>├─4021 /usr/libexec/ioprocess --read-pipe-fd 57
>> --write-pipe-fd 56 --max-threads 10 --max-queued-requests 10
>>
>>├─ /usr/libexec/ioprocess --read-pipe-fd 41
>> --write-pipe-fd 40 --max-threads 10 --max-queued-requests 10
>>
>>├─5120 /usr/libexec/ioprocess --read-pipe-fd 71
>> --write-pipe-fd 70 --max-threads 10 --max-queued-requests 10
>>
>>├─5232 /usr/libexec/ioprocess --read-pipe-fd 79
>> --write-pipe-fd 78 --max-threads 10 --max-queued-requests 10
>>
>>├─5533 /usr/libexec/ioprocess --read-pipe-fd 87
>> --write-pipe-fd 86 --max-threads 10 --max-queued-requests 10
>>
>>├─5576 /usr/libexec/ioprocess --read-pipe-fd 109
>> --write-pipe-fd 108 --max-threads 10 --max-queued-requests 10
>>
>>└─5589 /usr/libexec/ioprocess --read-pipe-fd 116
>> --write-pipe-fd 114 --max-threads 10 --max-queued-requests 10
>>
>>
>>
>> Feb 23 06:22:50 ovirt-node02.mgmt.lan.company.lan vdsm[3742]: vdsm
>> vds.dispatcher ERROR SSL error receiving from 
>> > connected ('::1', 36484, 0, 0) at 0x2548518>: unexpected eof
>>
>> Feb 23 06:22:50 ovirt-node02.mgmt.lan.company.lan vdsm[3742]: vdsm
>> vds.dispatcher ERROR SSL error receiving from 
>> > connected ('::1', 36486, 0, 0) at 0x2b7bd88>: unexpected eof
>>
>> Feb 23 06:22:53 ovirt-node02.mgmt.lan.company.lan vdsm[3742]: vdsm
>> vds.dispatcher ERROR SSL error receiving from 
>> > connected ('::1', 36490, 0, 0) at 0x25cb2d8>: unexpected eof
>>
>> Feb 23 06:22:54 ovirt-node02.mgmt.lan.company.lan vdsm[3742]: vdsm
>> vds.dispatcher ERROR SSL error receiving from 
>> > connected ('::1', 36494, 0, 

Re: [ovirt-users] Virsh

2017-03-02 Thread Simone Tiraboschi
On Thu, Mar 2, 2017 at 3:10 PM, Gianluca Cecchi 
wrote:

> On Thu, Mar 2, 2017 at 12:49 PM, Koen Vanoppen 
> wrote:
>
>> [root@mercury1 ~]# saslpasswd2 -a libvirt koen
>> Password:
>> Again (for verification):
>> [root@mercury1 ~]# virsh list --all
>> Please enter your authentication name: koen
>> Please enter your password:
>> error: failed to connect to the hypervisor
>> error: no valid connection
>> error: authentication failed: authentication failed
>>
>>
> I can only say that I just tested on my environment, with plain CentOS 7.3
> in oVirt 4.1 and it works.
>
> In theory, your connection string should use unix domain sockets if I'm
> not wrong and should be the same as "-c qemu:///system"
> In fact, using that connection URI I get the same prompts as without
> anything (only thing I just get the login/pwd prompt before running any
> command).
>
> Possibly there is something SELinux related? Is it enabled?
>
> Strange enough I'm verifying in my 4.1 system that I can actually run this
> command below without any password.
> (obviously all the caveat of running it out of oVirt are applicable...)
>
> [root@ovmsrv05 ~]# virsh -c qemu://ovmsrv05.mydomain/system
> Welcome to virsh, the virtualization interactive terminal.
>
> Type:  'help' for help with commands
>'quit' to quit
>
> virsh # list
>  IdName   State
> 
>  2 raclab1running
>  10c7testovn1 running
>
> virsh #
>
> This happens using the hostname used for the host when added to oVirt infra
> Instead if I use localhost I get
>
> [root@ovmsrv05 ~]# virsh -c qemu://localhost/system
> 2017-03-02 13:58:16.190+: 25221: info : libvirt version: 2.0.0,
> package: 10.el7_3.4 (CentOS BuildSystem ,
> 2017-01-17-23:37:48, c1bm.rdu2.centos.org)
> 2017-03-02 13:58:16.190+: 25221: info : hostname: ovmsrv05.mydomain
> 2017-03-02 13:58:16.190+: 25221: warning :
> virNetTLSContextCheckCertificate:1125 : Certificate check failed
> Certificate [session] owner does not match the hostname localhost
> error: failed to connect to the hypervisor
> error: authentication failed: Failed to verify peer's certificate
> [root@ovmsrv05 ~]#
>
> Does this command work for you too in 4.0?
> Is it in general a bug or a feature? Or anything cached (I don't think so
> because I can execute the same on another host where I didn't run anything
> before and where I didn't use the saslpasswd2 command to add a local virsh
> user)?
>

It's a feature: we configure it for TLS/x509 authentication for the engine
over TCP and SASL authentication for the local access overt the unix domain
socket.



>
> Gianluca
>
>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] 3.6 : Hosted engine High Availability

2016-08-25 Thread Simone Tiraboschi
On Thu, Aug 25, 2016 at 10:17 AM, Alexis HAUSER
 wrote:
> Hi,
>
>
> I'm trying to "deploy" the hosted engine (3.6) on a second node for HA. I 
> used hosted-engine --delpoy, answering the script questions it's a new host 
> setup.
>
> However it is requesting me for a FQDN, an engine name and a password for 
> admin@internal. Is this a normal behavior ?

No, it's not: you have to point to the same storage server you used
for the first host, hosted-engine-setup will detect the existing
installation and it will ask if you are going to add an additional
host.
At that point it will consume the answer file saved on the shared storage.

> It seems very strange to me that the configuration isn't taken from the 
> hosted engine storage, do you think I forgot a step ?
>
>
> Thanks in advance
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] 3.6 : Hosted engine High Availability

2016-08-25 Thread Simone Tiraboschi
On Thu, Aug 25, 2016 at 11:57 AM, Alexis HAUSER
 wrote:
>
>>This instead is an issue of your env:
>>your additional host is still named 'localhost.localdomain': of course
>>the engine VM will be able to resolve it but it will not reach your
>>host; in that case the engine will just try to add the engine VM
>>itself as an host creating a mess.
>>Since we saw this issue many times in the past from user logs, we are
>>now explicitly rejecting 'localhost.localdomain' as an host address.
>
> This is that part that confused me a bit : I need to set up a new FQDN for 
> the engine, but a different one right ? So each engine on each node needs a 
> different FQDN ?

No, you have just to use a sensate globally resolvable FQDN for your
additional host: 'localhost.localdomain' is not.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] 3.6 : Hosted engine High Availability

2016-08-25 Thread Simone Tiraboschi
On Thu, Aug 25, 2016 at 11:19 AM, Alexis HAUSER
 wrote:
>>No, it's not: you have to point to the same storage server you used
>>for the first host, hosted-engine-setup will detect the existing
>>installation and it will ask if you are going to add an additional
>>host.
>>At that point it will consume the answer file saved on the shared storage.
>
>
> This is exactly what I've done but it stills asks me those question about 
> FQDN, hostname and password, and as we can see it seems to find the answer 
> file :
>
>
>
>
>
>
>
> hosted-engine --deploy
> [ INFO  ] Stage: Initializing
> [ INFO  ] Generating a temporary VNC password.
> [ INFO  ] Stage: Environment setup
>   Continuing will configure this host for serving as hypervisor and 
> create a VM where you have to install the engine afterwards.
>   Are you sure you want to continue? (Yes, No)[Yes]:
>   Configuration files: []
>   Log file: 
> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160825082809-4evynk.log
>   Version: otopi-1.4.2 (otopi-1.4.2-1.el7ev)
>   It has been detected that this program is executed through an SSH 
> connection without using screen.
>   Continuing with the installation may lead to broken installation if 
> the network connection fails.
>   It is highly recommended to abort the installation and run it 
> inside a screen session using command "screen".
>   Do you want to continue anyway? (Yes, No)[No]: yes
> [ INFO  ] Hardware supports virtualization
> [ INFO  ] Bridge ovirtmgmt already created
> [ INFO  ] Stage: Environment packages setup
> [ INFO  ] Stage: Programs detection
> [ INFO  ] Stage: Environment setup
> [ INFO  ] Stage: Environment customization
>
>   --== STORAGE CONFIGURATION ==--
>
>   During customization use CTRL-D to abort.
>   Please specify the storage you would like to use (glusterfs, iscsi, 
> fc, nfs3, nfs4)[nfs3]: iscsi
>   Please specify the iSCSI portal IP address: x.x.x.x
>   Please specify the iSCSI portal port [3260]:
>   Please specify the iSCSI portal user:
>   Please specify the target name 
> (iqn.1984-05.com.dell:powervault., 
> iqn.1984-05.com.dell:powervault., 
> iqn.1984-05.com.dell:powervault., 
> iqn.1984-05.com.dell:powervault.) 
> [iqn.1984-05.com.dell:powervault.]:
>   The following luns have been found on the requested target:
> [1] 36002219000897d5e25bd5754b30f   836GiB  DELL  
>   MD3000i
> status: used, paths: 1 active
>
> [2] 36002219000897d5e2ea457a406f7   100GiB  DELL  
>   MD3000i
> status: used, paths: 1 active
>
> [3] 36002219000897d5e2e5357980305   150GiB  DELL  
>   MD3000i
> status: free, paths: 1 active
>
> [4] 36002219000897d5e2fa357ad109e   500GiB  DELL  
>   MD3000i
> status: free, paths: 1 active
>
> [5] 36002219000897d5e2fa857ad11c2   80GiB   DELL  
>   MD3000i
> status: free, paths: 1 active
>
> [6] 36002219000897d5e2f1657a920a2   587GiB  DELL  
>   MD3000i
> status: used, paths: 1 active
>
>   Please select the destination LUN (1, 2, 3, 4, 5, 6) [1]: 2
>   The specified storage location already contains a data domain. Is 
> this an additional host setup (Yes, No)[Yes]?
> [ INFO  ] Installing on additional host
>   Please specify the Host ID [Must be integer, default: 2]:
>
>   --== SYSTEM CONFIGURATION ==--
>
> [WARNING] A configuration file must be supplied to deploy Hosted Engine on an 
> additional host.
> [ INFO  ] Answer file successfully loaded

It seams it correctly loaded it.

>   --== NETWORK CONFIGURATION ==--
>
> [ INFO  ] Additional host deployment, firewall manager is 'iptables'
>   The following CPU types are supported by this host:
>  - model_Haswell-noTSX: Intel Haswell-noTSX Family
>  - model_SandyBridge: Intel SandyBridge Family
>  - model_Westmere: Intel Westmere Family
>  - model_Nehalem: Intel Nehalem Family
>  - model_Penryn: Intel Penryn Family
>  - model_Conroe: Intel Conroe Family
>
>   --== HOSTED ENGINE CONFIGURATION ==--
>
>   Enter the name which will be used to identify this host inside the 
> Administrator Portal [hosted_engine_2]:

This is fine since the name of your new host is not in the answer file.

>   Enter 'admin@internal' user password that will be used for 
> accessing the Administrator Portal:
>   Confirm 

Re: [ovirt-users] 3.6 : Hosted engine High Availability

2016-08-25 Thread Simone Tiraboschi
On Thu, Aug 25, 2016 at 3:12 PM, Alexis HAUSER
 wrote:
>
>
>> This is that part that confused me a bit : I need to set up a new FQDN for 
>> the engine, but a different one right ? So each engine on each node needs a 
>> different FQDN ?
>
>> No, you have just to use a sensate globally resolvable FQDN for your 
>> additional host: 'localhost.localdomain' is not.
>
> Well actually when I use the FQDN of my current hosted engine, I doesn't seem 
> to resolve...
>
>
> Host address:  [localhost.localdomain]: engine.mydomain.com
> [ ERROR ] Host name is not valid: vm-rhemgr.rennes.enst-bretagne.fr did not 
> resolve into an IP address
>   Please provide the address of this host.
>   Note: The engine VM and all the other hosts should be able to 
> correctly resolve it.

Can you please share your hosted-engine-setup logs?
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] 3.6 : Hosted engine High Availability

2016-08-25 Thread Simone Tiraboschi
On Thu, Aug 25, 2016 at 4:26 PM, Alexis HAUSER
 wrote:
>
>> Can you please share your hosted-engine-setup logs?
>
> Yes of course, here they are :)

OK, the issue is here:
2016-08-25 12:49:04 DEBUG otopi.plugins.otopi.dialog.human
human.queryString:156 query OVESETUP_NETWORK_FQDN_HOST_HOSTNAME
2016-08-25 12:49:04 DEBUG otopi.plugins.otopi.dialog.human
dialog.__logString:219 DIALOG:SEND Please provide the
address of this host.
2016-08-25 12:49:04 DEBUG otopi.plugins.otopi.dialog.human
dialog.__logString:219 DIALOG:SEND Note: The engine VM
and all the other hosts should be able to correctly resolve it.
2016-08-25 12:49:04 DEBUG otopi.plugins.otopi.dialog.human
dialog.__logString:219 DIALOG:SEND Host address:
[localhost.localdomain]:
2016-08-25 12:49:37 DEBUG otopi.plugins.otopi.dialog.human
dialog.__logString:219 DIALOG:RECEIVEvm-rhemgr.mydomain.com
2016-08-25 12:49:37 DEBUG
otopi.plugins.ovirt_hosted_engine_setup.network.bridge
hostname.test_hostname:411 test_hostname exception
Traceback (most recent call last):
  File "/usr/lib/python2.7/site-packages/ovirt_setup_lib/hostname.py",
line 407, in test_hostname
not_local_text,
  File "/usr/lib/python2.7/site-packages/ovirt_setup_lib/hostname.py",
line 252, in _validateFQDNresolvability
fqdn=fqdn,
RuntimeError: vm-rhemgr.mydomain.com did not resolve into an IP address
2016-08-25 12:49:37 ERROR
otopi.plugins.ovirt_hosted_engine_setup.network.bridge
dialog.queryEnvKey:115 Host name is not valid: vm-rhemgr.mydomain.com
did not resolve into an IP address



'Please provide the address of THIS host.' means that you have to
enter/validate the address of the host you are going to add (the host
where you are running hosted-engine --deploy command).

Let's try to recap:
the fqdn of your engine VM is 'vm-rhemgr.mydomain.com',
the fqdn of your host is currently 'localhost.localdomain' but it's
not acceptable (try to run 'ssh localhost.localdomain' on the engine
VM and see where are you getting...)

So you have just to configure a valid fqdn on your additional host
(something like 'my2ndhost.mydomain.com') and confirm it when asked by
that question.

Normally we suggest to rely on a properly configured DNS; you can just
work entering values under '/etc/hosts' but it's up to you to properly
maintain it:
- the engine VM should be able to resolve the address of all the hosts
to contact them: this is not true in your env, with
'localhost.localdomain' your engine VM will not reach your host...
- each host should be able to resolve the address of all the other
hosts and also the address of the engine VM: this is not true in your
env as I read 'RuntimeError: vm-rhemgr.mydomain.com did not resolve
into an IP address'
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] HELP Upgrade hypervisors from CentOS 6.8 to CentOS 7

2016-09-02 Thread Simone Tiraboschi
On Fri, Sep 2, 2016 at 5:21 PM, VONDRA Alain  wrote:

> Hi,
> I'd like to upgrade my oVirt environment from 3.5 to 3.6 and maybe 4,
> actually the manager oVirt is un 3.5 version, installed on a CentOS 7.2 and
> the two hypervisors are installed on CentOS 6.8.
> I need anyway to upgrade the hosts to be able to move to 3.6.
> I've tried to upgrade the first host, but I had of course issues telling
> that it can't be possible to mix different OS in the same cluster, I've
> also tried to create another cluster to put this host without success.
> What is the best way to upgrade cleanly and safely ?
> Thank you in advance for your advices
>
>
Follow this:
https://www.ovirt.org/documentation/how-to/hosted-engine-host-OS-upgrade/

Simply ignore the steps that refers to the engine VM if you are with the
engine on a physical system.


> Alain VONDRA
>
> Chargé d'exploitation des Systèmes d'Information
> Direction Administrative et Financière
> +33 1 44 39 77 76
>
> UNICEF France
> 3 rue Duguay Trouin  75006
> PARIS
> www.unicef.fr
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


<    1   2   3   4   5   6   7   8   9   10   >