Re: [ovirt-users] New User 3.6 - ovirt-image-repository

2016-04-12 Thread Jack Greene
Thanks for the reply.

 

You’re right the v1 images aren’t really what I need but the cirros disk was 
helpful.  This is my 1st engine setup and the ovirt-image-repository has the 
URL of http://glance.ovirt.org:9292 in external providers.  That works after I 
used the tab at the top but only v1 images are available.  I did create my own 
Centos 6.7 minimal template. 

 

 

 

From: Oved Ourfali [mailto:oourf...@redhat.com] 
Sent: Tuesday, April 12, 2016 9:59 PM
To: Jack Greene 
Cc: users 
Subject: Re: [ovirt-users] New User 3.6 - ovirt-image-repository

 

You should access /api/v2 rather than just /v2.

I was surprised to see /v1 actually works, but /api/v2 should give you what you 
need.

What issue do you see in oVirt with the image repository?

I must admit that the images there are not that up-to-date, so not sure they 
would be sufficient for you anyway.

 

 

On Tue, Apr 12, 2016 at 5:46 PM, Jack Greene  > wrote:

I'm really excited to try out oVirt and have installed it 3 times using the 
hosted engine documentation.  I continue to get stalled at deploying a new vm 
from the ovirt-image-repository.  I suspect there is a problem with the glance 
server.  I can access the base URL and the v1 version URL but not the v2 
version URL.

Is there a way to tell my new engine to use the v1 repo or are those templates 
incompatible with the new engine?

oVirt Engine Version: 3.6.4.1-1.el6

http://glance.ovirt.org:9292 - ok
http://glance.ovirt.org:9292/v1/ - ok
http://glance.ovirt.org:9292/v2/ - 404

Thanks,  Jack

___
Users mailing list
Users@ovirt.org  
http://lists.ovirt.org/mailman/listinfo/users



 

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] New User 3.6 - ovirt-image-repository

2016-04-12 Thread Oved Ourfali
You should access /api/v2 rather than just /v2.
I was surprised to see /v1 actually works, but /api/v2 should give you what
you need.
What issue do you see in oVirt with the image repository?
I must admit that the images there are not that up-to-date, so not sure
they would be sufficient for you anyway.


On Tue, Apr 12, 2016 at 5:46 PM, Jack Greene  wrote:

> I'm really excited to try out oVirt and have installed it 3 times using
> the hosted engine documentation.  I continue to get stalled at deploying a
> new vm from the ovirt-image-repository.  I suspect there is a problem with
> the glance server.  I can access the base URL and the v1 version URL but
> not the v2 version URL.
>
> Is there a way to tell my new engine to use the v1 repo or are those
> templates incompatible with the new engine?
>
> oVirt Engine Version: 3.6.4.1-1.el6
>
> http://glance.ovirt.org:9292 - ok
> http://glance.ovirt.org:9292/v1/ - ok
> http://glance.ovirt.org:9292/v2/ - 404
>
> Thanks,  Jack
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] How to Change oVirt Engine Appliance to use Spice?

2016-04-12 Thread Richard Harmonson
How do I change the oVirt Engine Appliance from VNC to the Spice Console?

Thank you, Richard
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] New User 3.6 - ovirt-image-repository < Solved?

2016-04-12 Thread Jack Greene
Ok figured out I was clicking in the wrong place looking for the repo

http://i.imgur.com/bJUD0Co.png

Seems the icon in the navigation pane does nothing except allow you to see
your external providers.  The repo is actually across the top in the
resource tabs under storage.  Feeling dumb (been looking at this for 2 days)
but posting the image in case it helps another new user.

>From the Red Hat Community -
http://community.redhat.com/blog/2014/10/up-and-running-with-ovirt-3-5/
>From the storage tab, you should see an "ovirt-image-repository" entry next
to a little OpenStack logo. Clicking on this domain will bring up a menu of
images available in this repository. Click on the "CirrOS" image (which is
very small and perfect for testing) in the list and then click "Import,"
before hitting the OK button in the pop-up dialog to continue.


-Original Message-
From: users-boun...@ovirt.org [mailto:users-boun...@ovirt.org] On Behalf Of
Jack Greene
Sent: Tuesday, April 12, 2016 7:46 AM
To: users@ovirt.org
Subject: [ovirt-users] New User 3.6 - ovirt-image-repository

I'm really excited to try out oVirt and have installed it 3 times using the
hosted engine documentation.  I continue to get stalled at deploying a new
vm from the ovirt-image-repository.  I suspect there is a problem with the
glance server.  I can access the base URL and the v1 version URL but not the
v2 version URL.

Is there a way to tell my new engine to use the v1 repo or are those
templates incompatible with the new engine?

oVirt Engine Version: 3.6.4.1-1.el6

http://glance.ovirt.org:9292 - ok
http://glance.ovirt.org:9292/v1/ - ok
http://glance.ovirt.org:9292/v2/ - 404

Thanks,  Jack

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Looking for Python-SDK documentation

2016-04-12 Thread Brett I. Holcomb
   1. On Tue, 2016-04-12 at 23:03 +0200, Frank Thommen wrote:
> On 12.04.16 13:57, Alexander Wels wrote:
> > 
> > On Monday, April 11, 2016 05:03:28 PM Brett I. Holcomb wrote:
> > > 
> > > On Mon, 2016-04-11 at 20:41 +0200, Frank Thommen wrote:
> > > > 
> > > > On 11.04.16 20:17, Brett I. Holcomb wrote:
> > > > [...]
> > > I'm in the same boat as Frank.  I've done programming in various
> > > languages since Fortran without the numbers  but only when
> > > needed
> > > for my job as an Engineer so I'm not a professional but just
> > > trying to
> > > get a job done.  It would be nice to have a full reference so we
> > > know
> > > what to provide.  When trying to connect with the api I finally
> > > figured
> > > out to use ca_file (like ca-file on the command line).  Raz's
> > > reference
> > > is more complete but still leaves a lot out.  The newer
> > > equivalent of
> > > Raz's reference seems to be http://www.ovirt.org/develop/release-
> > > manage
> > > ment/features/infra/python-sdk/.
> > The Python/Java/Ruby SDKs are simply language specific wrappers
> > around the REST
> > api. So if you want a full list of all the options available check
> > out the
> > REST api documentation. You will have to translate a particular
> > REST api
> > field/feature to the SDK, but all the SDKs are generated from the
> > REST api
> > interface definition so the naming and everything should be the
> > same.
> That makes sense, thanks.  On the other hand I could also not find
> any 
> REST api reference.  Just some pages with examples a la "find it out 
> yourself from there"...
> 
> frank
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users

Same here.  A lot of examples, some more detailed but no "API Reference
Manual for oVirt REST API" like we had in the old days .

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Looking for Python-SDK documentation

2016-04-12 Thread Frank Thommen

On 12.04.16 13:57, Alexander Wels wrote:

On Monday, April 11, 2016 05:03:28 PM Brett I. Holcomb wrote:

On Mon, 2016-04-11 at 20:41 +0200, Frank Thommen wrote:

On 11.04.16 20:17, Brett I. Holcomb wrote:
[...]

I'm in the same boat as Frank.  I've done programming in various
languages since Fortran without the numbers  but only when needed
for my job as an Engineer so I'm not a professional but just trying to
get a job done.  It would be nice to have a full reference so we know
what to provide.  When trying to connect with the api I finally figured
out to use ca_file (like ca-file on the command line).  Raz's reference
is more complete but still leaves a lot out.  The newer equivalent of
Raz's reference seems to be http://www.ovirt.org/develop/release-manage
ment/features/infra/python-sdk/.


The Python/Java/Ruby SDKs are simply language specific wrappers around the REST
api. So if you want a full list of all the options available check out the
REST api documentation. You will have to translate a particular REST api
field/feature to the SDK, but all the SDKs are generated from the REST api
interface definition so the naming and everything should be the same.


That makes sense, thanks.  On the other hand I could also not find any 
REST api reference.  Just some pages with examples a la "find it out 
yourself from there"...


frank
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Looking for Python-SDK documentation

2016-04-12 Thread Brett I. Holcomb
On Tue, 2016-04-12 at 07:57 -0400, Alexander Wels wrote:
> On Monday, April 11, 2016 05:03:28 PM Brett I. Holcomb wrote:
> > 
> > On Mon, 2016-04-11 at 20:41 +0200, Frank Thommen wrote:
> > > 
> > > On 11.04.16 20:17, Brett I. Holcomb wrote:
> > > > 
> > > > On Mon, 2016-04-11 at 14:38 -0300, Amador Pahim wrote:
> > > > > 
> > > > > On 04/11/2016 02:07 PM, Brett I. Holcomb wrote:
> > > > > > 
> > > > > > On Mon, 2016-04-11 at 17:27 +0200, Frank Thommen wrote:
> > > > > > > 
> > > > > > > Dear all,
> > > > > > > 
> > > > > > > I'm currently lost at finding any documentation about the
> > > > > > > Python SDK
> > > > > > > (http://www.ovirt.org/develop/release-management/features
> > > > > > > /inf
> > > > > > > ra/python-sdk/)
> > > > > > > like provided classes, functions etc..  There are some
> > > > > > > examples on the
> > > > > > > mentioned page, but I cannot find a complete
> > > > > > > documentation.  Our oVirt
> > > > > > > server is running CentOS 7 with ovirt-engine-sdk-python
> > > > > > > installed.
> > > > > > > However there doesn't seem to exist an ovirt-engine-sdk-
> > > > > > > python-docs
> > > > > > > package and I couldn't find any appropriate link on the
> > > > > > > oVirt
> > > > > > > documentation pages (http://www.ovirt.org/documentation/)
> > > > > > > .
> > > > > > > 
> > > > > > > Any ideas, where the documentation is available?
> > > > > > > 
> > > > > > > Cheers
> > > > > > > Frank
> > > > > > > ___
> > > > > > > Users mailing list
> > > > > > > Users@ovirt.org 
> > > > > > > http://lists.ovirt.org/mailman/listinfo/users
> > > > > > > 
> > > > > > Other than what you found I found this but that's
> > > > > > all.  Doesn't
> > > > > > seem
> > > > > > to be much other than examples an the one that shows what
> > > > > > to
> > > > > > import.
> > > > > >  I, too would like to find what your are looking for so I
> > > > > > can
> > > > > > use it.
> > > > > There is this RHEV documentation that can be helpful:
> > > > > https://access.redhat.com/documentation/en-US/Red_Hat_Enterpr
> > > > > ise_
> > > > > Virtualization/3.3/html/Developer_Guide/chap-
> > > > > Python_Quick_Start_Example.html
> > > > > 
> > > > > > 
> > > > > > http://www.ovirt.org/develop/api/pythonapi/
> > > > > > 
> > > > > > 
> > > > > > ___
> > > > > > Users mailing list
> > > > > > Users@ovirt.org 
> > > > > > http://lists.ovirt.org/mailman/listinfo/users
> > > > > > 
> > > > Thanks.  That is helpful but as Frank said it would be nice to
> > > > have
> > > > an
> > > > api reference for each class and function that gives the
> > > > parameters,
> > > > return values, and other stuff usually found in an API
> > > > reference.  The
> > > > examples are helpful but don't give all the information abut a
> > > > function
> > > > or class.
> > > Thanks to all who answered.  Brett brings it to the point:  All
> > > sent 
> > > links so far are indeed helpful - thanks a lot - but not the
> > > reference I 
> > > expected. 
> > > https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_
> > > Virt
> > > ualization/3.6/html/Python_SDK_Guide/chap-
> > > Python_Reference_Documentation.html#Python_Reference_Documentatio
> > > n 
> > > mentions `pydoc`, but this documentation seems to be provided
> > > only
> > > for 
> > > some modules or to be incomplete.  Also for me not being a
> > > software 
> > > developper and newish to Python, the `pydoc` information is not
> > > very 
> > > useful.  Where can I e.g. find the documentation for vms.get()
> > > and 
> > > vms.add() (just to name teo concrete examples)?
> > > 
> > > Frank
> > > ___
> > > Users mailing list
> > > Users@ovirt.org
> > > http://lists.ovirt.org/mailman/listinfo/users
> > I'm in the same boat as Frank.  I've done programming in various
> > languages since Fortran without the numbers  but only when
> > needed
> > for my job as an Engineer so I'm not a professional but just trying
> > to
> > get a job done.  It would be nice to have a full reference so we
> > know
> > what to provide.  When trying to connect with the api I finally
> > figured
> > out to use ca_file (like ca-file on the command line).  Raz's
> > reference
> > is more complete but still leaves a lot out.  The newer equivalent
> > of
> > Raz's reference seems to be http://www.ovirt.org/develop/release-ma
> > nage
> > ment/features/infra/python-sdk/.
> The Python/Java/Ruby SDKs are simply language specific wrappers
> around the REST 
> api. So if you want a full list of all the options available check
> out the 
> REST api documentation. You will have to translate a particular REST
> api 
> field/feature to the SDK, but all the SDKs are generated from the
> REST api 
> interface definition so the naming and 

Re: [ovirt-users] Inconsistency beteween hosted engine id and sanlock id

2016-04-12 Thread Nir Soffer
On Tue, Apr 12, 2016 at 6:12 PM, Martin Sivak  wrote:
> Hi,
>
> thanks for the summary, this is what I was suspecting.
>
> Just a clarification about the hosted engine host-id and lockspace.
> Hosted engine has a separate lockspace from VDSM and uses
> hosted-engine's host-id there consistently to protect a metadata
> whiteboard. It has nothing to do with the VM and there is no conflict
> here.
>
>
> The issue seems to be that the VDSM lockspace is being used when
> connect storage domain is called and both hosted engine and
> ovirt-engine can call the connect command. Unfortunately hosted engine
> does not know the vds_spm_id when mounting the volume for the first
> time (even before ovirt-engine VM is started) and uses the host-id for
> that.
>
> Now, there is probably no issue when all hosts accessing that storage
> domain are hosted engine enabled right from the start as the storage
> domain is mounted to all hosts before the engine starts and the
> locking uses consistent id (hosted engine host-id).
>
> The problem surfaces on a host where the engine manages to call the
> "connect hosted engine storage domain" first. Because engine uses the
> vds_spm_id for the requested lease and a collision happens.
>
> I do not see any easy fix at this moment, maybe except telling engine
> to use hosted engine id when it tries to connect the hosted engine
> storage domain. That feels like a hack, but might work.
>
> There also seems to be a bug for this issue now:
> https://bugzilla.redhat.com/show_bug.cgi?id=1322849
>
> Simone/Nir can you please comment on the issue to confirm that our
> findings are correct?

The only host id that hosted engine can use to connect a host to
storage is the host id managed by engine. Using another id will fail to connect
the host to storage if another host is up and using this id.

If the operation is successful with the incorrect host id, another host will
fail to connect to storage because the first host is using its id.

The only way is to sync engine with the ids managed by hosted engine, or hosted
engine with the ids managed by engine. We cannot use both in the same time.

Once hosted engine is running, you cannot change the host id of the host its
running on, since releasing the lockspace for acquiring again with the correct
host id will cause sanlock to kill the engine vm.

For working around this issue, we need to export the host is from engine, store
them in hosted engine configuration, and use only these ids to connect to
storage after a restart.

Since host ids are static and change only when you add or remove a host to
engine, I think we can store the host id locally on the host, in the same way
we keep the host certificates or uuid (/etc/pki/vdsm, /etc/vdsm/vdsm.id).

Thoughts?

Nir

>
> Thanks
>
> Regards
>
> --
> Martin Sivak
> SLA / oVirt
>
> On Tue, Apr 12, 2016 at 4:31 PM, Baptiste Agasse
>  wrote:
>> Hi all,
>>
>> Last week we had problem on our ovirt infrastructure. The hosted engine 
>> didn't came up after the reboot of the host which hosted it. With the help 
>> of some people on #ovirt IRC channel (msivak, nsoffer and some others, thank 
>> to all of them) i managed to have my hosted engine up and running, but the 
>> underlying problem is still there. I think there is an inconsistency between 
>> sanlock ID of the hosts.
>>
>> Some background:
>>
>> We installed ovirt in 3.5 on CentOS 7 about 9 month ago. We have one DC with 
>> two clusters:
>>
>> cluster 1: 4 hosts (virt1, virt2, virt3, virt4) that were installed with 
>> 'hosted-engine --deploy' so there are capable to run the engine VM.
>> cluster 2: 2 hosts (virt6 and virt7) that were installed via the webui, so 
>> are 'normal' ovirt hosts.
>>
>> Since that we have successfully upgraded ovirt to 3.6 and set our cluster to 
>> 3.6 compatibility mode.
>>
>> Some weeks after something broke and the virt4 host rebooted. After some 
>> help on the IRC channel, i managed to get the engine vm up and running. 
>> After that i dug into the problem that seems to be around the sanlock part.
>>
>> After explanations, that i understand is:
>>
>> sanlock manage locks at DC level. there is an hosted_engine lock to manage 
>> who run the VM and there is a vdsm level lock on the hosted_engine disk (or 
>> any other VM disks) to know who can write to the disk.
>>
>> The problem in my case is that on some hosts that were installed in 3.5, the 
>> hosted_engine ID and the vds_spm_id are not the same, and some other host 
>> have it vds_spm_id identical to some other host hosted_engine_id. So in some 
>> case, some host can't acquire the lock on some disks and have different IDs 
>> in the sanlock space.
>>
>> Example, im my case:
>>
>> #
>> # For the hosted_engine hosts:
>> #
>> [root@virt1 ~]# grep host_id /etc/ovirt-hosted-engine/hosted-engine.conf
>> host_id=1
>>
>> [root@virt2 ~]# grep host_id /etc/ovirt-hosted-engine/hosted-engine.conf
>> host_id=2
>>
>> 

Re: [ovirt-users] Hosted engine on gluster problem

2016-04-12 Thread Nir Soffer
On Tue, Apr 12, 2016 at 3:05 PM, Luiz Claudio Prazeres Goncalves
 wrote:
> Hi Sandro, I've been using gluster with 3 external hosts for a while and
> things are working pretty well, however this single point of failure looks
> like a simple feature to implement,but critical to anyone who wants to use
> gluster on production  . This is not hyperconvergency which has other
> issues/implications. So , why not have this feature out on 3.6 branch? It
> looks like just let vdsm use the 'backupvol-server' option when mounting the
> engine domain and make the property tests.

Can you explain what is the problem, and what is the suggested solution?

Engine and vdsm already support the backupvol-server option - you can
define this option in the storage domain options when you create a gluster
storage domain. With this option vdsm should be able to connect to gluster
storage domain even if a brick is down.

If you don't have this option in engine , you probably cannot add it with hosted
engine setup, since for editing it you must put the storage domain in
maintenance
and if you do this the engine vm will be killed :-) This is is one of
the issues with
engine managing the storage domain it runs on.

I think the best way to avoid this issue, is to add a DNS entry
providing the addresses
of all the gluster bricks, and use this address for the gluster
storage domain. This way
the glusterfs mount helper can mount the domain even if one of the
gluster bricks
are down.

Again, we will need some magic from the hosted engine developers to modify the
address of the hosted engine gluster domain on existing system.

Nir

>
> Could you add this feature to the next release of 3.6 branch?
>
> Thanks
> Luiz
>
> Em ter, 12 de abr de 2016 05:03, Sandro Bonazzola 
> escreveu:
>>
>> On Mon, Apr 11, 2016 at 11:44 PM, Bond, Darryl 
>> wrote:
>>>
>>> My setup is hyperconverged. I have placed my test results in
>>> https://bugzilla.redhat.com/show_bug.cgi?id=1298693
>>>
>>
>> Ok, so you're aware about the limitation of the single point of failure.
>> If you drop the host referenced in hosted engine configuration for the
>> initial setup it won't be able to connect to shared storage even if the
>> other hosts in the cluster are up since the entry point is down.
>> Note that hyperconverged deployment is not supported in 3.6.
>>
>>
>>>
>>>
>>> Short description of setup:
>>>
>>> 3 hosts with 2 disks each set up with gluster replica 3 across the 6
>>> disks volume name hosted-engine.
>>>
>>> Hostname hosted-storage configured in /etc//hosts to point to the host1.
>>>
>>> Installed hosted engine on host1 with the hosted engine storage path =
>>> hosted-storage:/hosted-engine
>>>
>>> Install first engine on h1 successful. Hosts h2 and h3 added to the
>>> hosted engine. All works fine.
>>>
>>> Additional storage and non-hosted engine hosts added etc.
>>>
>>> Additional VMs added to hosted-engine storage (oVirt Reports VM and
>>> Cinder VM). Additional VM's are hosted by other storage - cinder and NFS.
>>>
>>> The system is in production.
>>>
>>>
>>> Engine can be migrated around with the web interface.
>>>
>>>
>>> - 3.6.4 upgrade released, follow the upgrade guide, engine is upgraded
>>> first , new Centos kernel requires host reboot.
>>>
>>> - Engine placed on h2 -  h3 into maintenance (local) upgrade and Reboot
>>> h3 - No issues - Local maintenance removed from h3.
>>>
>>> - Engine placed on h3 -  h2 into maintenance (local) upgrade and Reboot
>>> h2 - No issues - Local maintenance removed from h2.
>>>
>>> - Engine placed on h3 -h1 into mainteance (local) upgrade and reboot h1 -
>>> engine crashes and does not start elsewhere, VM(cinder)  on h3 on same
>>> gluster volume pauses.
>>>
>>> - Host 1 takes about 5 minutes to reboot (Enterprise box with all it's
>>> normal BIOS probing)
>>>
>>> - Engine starts after h1 comes back and stabilises
>>>
>>> - VM(cinder) unpauses itself,  VM(reports) continued fine the whole time.
>>> I can do no diagnosis on the 2 VMs as the engine is not available.
>>>
>>> - Local maintenance removed from h1
>>>
>>>
>>> I don't believe the issue is with gluster itself as the volume remains
>>> accessible on all hosts during this time albeit with a missing server
>>> (gluster volume status) as each gluster server is rebooted.
>>>
>>> Gluster was upgraded as part of the process, no issues were seen here.
>>>
>>>
>>> I have been able to duplicate the issue without the upgrade by following
>>> the same sort of timeline.
>>>
>>>
>>> 
>>> From: Sandro Bonazzola 
>>> Sent: Monday, 11 April 2016 7:11 PM
>>> To: Richard Neuboeck; Simone Tiraboschi; Roy Golan; Martin Sivak; Sahina
>>> Bose
>>> Cc: Bond, Darryl; users
>>> Subject: Re: [ovirt-users] Hosted engine on gluster problem
>>>
>>>
>>>
>>> On Mon, Apr 11, 2016 at 9:37 AM, Richard Neuboeck
>>> > wrote:
>>> Hi 

Re: [ovirt-users] Inconsistency beteween hosted engine id and sanlock id

2016-04-12 Thread Baptiste Agasse


- Le 12 Avr 16, à 17:29, Simone Tiraboschi stira...@redhat.com a écrit :

> On Tue, Apr 12, 2016 at 5:12 PM, Martin Sivak  wrote:
>> Hi,
>>
>> thanks for the summary, this is what I was suspecting.
>>
>> Just a clarification about the hosted engine host-id and lockspace.
>> Hosted engine has a separate lockspace from VDSM and uses
>> hosted-engine's host-id there consistently to protect a metadata
>> whiteboard. It has nothing to do with the VM and there is no conflict
>> here.
>>
>>
>> The issue seems to be that the VDSM lockspace is being used when
>> connect storage domain is called and both hosted engine and
>> ovirt-engine can call the connect command. Unfortunately hosted engine
>> does not know the vds_spm_id when mounting the volume for the first
>> time (even before ovirt-engine VM is started) and uses the host-id for
>> that.
>>
>> Now, there is probably no issue when all hosts accessing that storage
>> domain are hosted engine enabled right from the start as the storage
>> domain is mounted to all hosts before the engine starts and the
>> locking uses consistent id (hosted engine host-id).
>>
>> The problem surfaces on a host where the engine manages to call the
>> "connect hosted engine storage domain" first. Because engine uses the
>> vds_spm_id for the requested lease and a collision happens.
>>
>> I do not see any easy fix at this moment, maybe except telling engine
>> to use hosted engine id when it tries to connect the hosted engine
>> storage domain. That feels like a hack, but might work.
>>
>> There also seems to be a bug for this issue now:
>> https://bugzilla.redhat.com/show_bug.cgi?id=1322849

IMHO, it seems that this bug is related.

>>
>> Simone/Nir can you please comment on the issue to confirm that our
>> findings are correct?
> 
> I feel so but probably the solution you proposed is not enough since
> we also allow to mix hosted-engine enabled hosts and regular hosts
> (where you don't have any hosted-engine id) in the same cluster and,
> once the hosted-engine-storage domain got imported by engine, the
> engine will going to connect it on all of them.

More than in one cluster, i think that problem is data-center wide. As you said 
as the hosted-engine-storage domain got imported by engine, any host in the DC 
can connect to it (in my case virt4 and virt7 are hosts from different 
clusters, but in the same DC).

> 
>> Thanks
>>
>> Regards
>>
>> --
>> Martin Sivak
>> SLA / oVirt
>>
>> On Tue, Apr 12, 2016 at 4:31 PM, Baptiste Agasse
>>  wrote:
>>> Hi all,
>>>
>>> Last week we had problem on our ovirt infrastructure. The hosted engine 
>>> didn't
>>> came up after the reboot of the host which hosted it. With the help of some
>>> people on #ovirt IRC channel (msivak, nsoffer and some others, thank to all 
>>> of
>>> them) i managed to have my hosted engine up and running, but the underlying
>>> problem is still there. I think there is an inconsistency between sanlock 
>>> ID of
>>> the hosts.
>>>
>>> Some background:
>>>
>>> We installed ovirt in 3.5 on CentOS 7 about 9 month ago. We have one DC 
>>> with two
>>> clusters:
>>>
>>> cluster 1: 4 hosts (virt1, virt2, virt3, virt4) that were installed with
>>> 'hosted-engine --deploy' so there are capable to run the engine VM.
>>> cluster 2: 2 hosts (virt6 and virt7) that were installed via the webui, so 
>>> are
>>> 'normal' ovirt hosts.
>>>
>>> Since that we have successfully upgraded ovirt to 3.6 and set our cluster 
>>> to 3.6
>>> compatibility mode.
>>>
>>> Some weeks after something broke and the virt4 host rebooted. After some 
>>> help on
>>> the IRC channel, i managed to get the engine vm up and running. After that i
>>> dug into the problem that seems to be around the sanlock part.
>>>
>>> After explanations, that i understand is:
>>>
>>> sanlock manage locks at DC level. there is an hosted_engine lock to manage 
>>> who
>>> run the VM and there is a vdsm level lock on the hosted_engine disk (or any
>>> other VM disks) to know who can write to the disk.
>>>
>>> The problem in my case is that on some hosts that were installed in 3.5, the
>>> hosted_engine ID and the vds_spm_id are not the same, and some other host 
>>> have
>>> it vds_spm_id identical to some other host hosted_engine_id. So in some 
>>> case,
>>> some host can't acquire the lock on some disks and have different IDs in the
>>> sanlock space.
>>>
>>> Example, im my case:
>>>
>>> #
>>> # For the hosted_engine hosts:
>>> #
>>> [root@virt1 ~]# grep host_id /etc/ovirt-hosted-engine/hosted-engine.conf
>>> host_id=1
>>>
>>> [root@virt2 ~]# grep host_id /etc/ovirt-hosted-engine/hosted-engine.conf
>>> host_id=2
>>>
>>> [root@virt3 ~]# grep host_id /etc/ovirt-hosted-engine/hosted-engine.conf
>>> host_id=3
>>>
>>> [root@virt4 ~]# grep host_id /etc/ovirt-hosted-engine/hosted-engine.conf
>>> host_id=4
>>>
>>> #
>>> # For all hosts, including hosted engine:
>>> #
>>> [root@virt1 ~]# sanlock client status
>>> daemon 

[ovirt-users] New User 3.6 - ovirt-image-repository

2016-04-12 Thread Jack Greene
I'm really excited to try out oVirt and have installed it 3 times  
using the hosted engine documentation.  I continue to get stalled at  
deploying a new vm from the ovirt-image-repository.  I suspect there  
is a problem with the glance server.  I can access the base URL and  
the v1 version URL but not the v2 version URL.


Is there a way to tell my new engine to use the v1 repo or are those  
templates incompatible with the new engine?


oVirt Engine Version: 3.6.4.1-1.el6

http://glance.ovirt.org:9292 - ok
http://glance.ovirt.org:9292/v1/ - ok
http://glance.ovirt.org:9292/v2/ - 404

Thanks,  Jack

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Inconsistency beteween hosted engine id and sanlock id

2016-04-12 Thread Martin Sivak
Hi,

thanks for the summary, this is what I was suspecting.

Just a clarification about the hosted engine host-id and lockspace.
Hosted engine has a separate lockspace from VDSM and uses
hosted-engine's host-id there consistently to protect a metadata
whiteboard. It has nothing to do with the VM and there is no conflict
here.


The issue seems to be that the VDSM lockspace is being used when
connect storage domain is called and both hosted engine and
ovirt-engine can call the connect command. Unfortunately hosted engine
does not know the vds_spm_id when mounting the volume for the first
time (even before ovirt-engine VM is started) and uses the host-id for
that.

Now, there is probably no issue when all hosts accessing that storage
domain are hosted engine enabled right from the start as the storage
domain is mounted to all hosts before the engine starts and the
locking uses consistent id (hosted engine host-id).

The problem surfaces on a host where the engine manages to call the
"connect hosted engine storage domain" first. Because engine uses the
vds_spm_id for the requested lease and a collision happens.

I do not see any easy fix at this moment, maybe except telling engine
to use hosted engine id when it tries to connect the hosted engine
storage domain. That feels like a hack, but might work.

There also seems to be a bug for this issue now:
https://bugzilla.redhat.com/show_bug.cgi?id=1322849

Simone/Nir can you please comment on the issue to confirm that our
findings are correct?

Thanks

Regards

--
Martin Sivak
SLA / oVirt

On Tue, Apr 12, 2016 at 4:31 PM, Baptiste Agasse
 wrote:
> Hi all,
>
> Last week we had problem on our ovirt infrastructure. The hosted engine 
> didn't came up after the reboot of the host which hosted it. With the help of 
> some people on #ovirt IRC channel (msivak, nsoffer and some others, thank to 
> all of them) i managed to have my hosted engine up and running, but the 
> underlying problem is still there. I think there is an inconsistency between 
> sanlock ID of the hosts.
>
> Some background:
>
> We installed ovirt in 3.5 on CentOS 7 about 9 month ago. We have one DC with 
> two clusters:
>
> cluster 1: 4 hosts (virt1, virt2, virt3, virt4) that were installed with 
> 'hosted-engine --deploy' so there are capable to run the engine VM.
> cluster 2: 2 hosts (virt6 and virt7) that were installed via the webui, so 
> are 'normal' ovirt hosts.
>
> Since that we have successfully upgraded ovirt to 3.6 and set our cluster to 
> 3.6 compatibility mode.
>
> Some weeks after something broke and the virt4 host rebooted. After some help 
> on the IRC channel, i managed to get the engine vm up and running. After that 
> i dug into the problem that seems to be around the sanlock part.
>
> After explanations, that i understand is:
>
> sanlock manage locks at DC level. there is an hosted_engine lock to manage 
> who run the VM and there is a vdsm level lock on the hosted_engine disk (or 
> any other VM disks) to know who can write to the disk.
>
> The problem in my case is that on some hosts that were installed in 3.5, the 
> hosted_engine ID and the vds_spm_id are not the same, and some other host 
> have it vds_spm_id identical to some other host hosted_engine_id. So in some 
> case, some host can't acquire the lock on some disks and have different IDs 
> in the sanlock space.
>
> Example, im my case:
>
> #
> # For the hosted_engine hosts:
> #
> [root@virt1 ~]# grep host_id /etc/ovirt-hosted-engine/hosted-engine.conf
> host_id=1
>
> [root@virt2 ~]# grep host_id /etc/ovirt-hosted-engine/hosted-engine.conf
> host_id=2
>
> [root@virt3 ~]# grep host_id /etc/ovirt-hosted-engine/hosted-engine.conf
> host_id=3
>
> [root@virt4 ~]# grep host_id /etc/ovirt-hosted-engine/hosted-engine.conf
> host_id=4
>
> #
> # For all hosts, including hosted engine:
> #
> [root@virt1 ~]# sanlock client status
> daemon 3a99892c-5d3a-4d3d-bac7-d35259363c98.virt1
> p -1 helper
> p -1 listener
> p -1 status
> s 
> hosted-engine:1:/var/run/vdsm/storage/377ae8e8-0eeb-4591-b50f-3d21298b4146/607719dd-b71e-4527-814a-964ed0c1f8ea/6a0b878d-fe7e-4fb6-bd5d-1254bebb0ca0:0
> s 
> 295207d7-41ea-4cda-a028-f860c357d46b:1:/dev/295207d7-41ea-4cda-a028-f860c357d46b/ids:0
> s 
> daf1b53c-7e29-4b18-a9e2-910605cc7080:1:/dev/daf1b53c-7e29-4b18-a9e2-910605cc7080/ids:0
> s 
> 680d5ed1-ed70-4340-a430-ddfa39ee3052:1:/dev/680d5ed1-ed70-4340-a430-ddfa39ee3052/ids:0
> s 
> 350e5736-41c0-4017-a8fd-9866edad:1:/dev/350e5736-41c0-4017-a8fd-9866edad/ids:0
> s 
> 377ae8e8-0eeb-4591-b50f-3d21298b4146:1:/dev/377ae8e8-0eeb-4591-b50f-3d21298b4146/ids:0
>
> [root@virt2 ~]# sanlock client status
> daemon 48fe11a1-6c64-4a56-abf0-6f9690e6a8c2.virt2
> p -1 helper
> p -1 listener
> p -1 status
> s 
> hosted-engine:2:/var/run/vdsm/storage/377ae8e8-0eeb-4591-b50f-3d21298b4146/607719dd-b71e-4527-814a-964ed0c1f8ea/6a0b878d-fe7e-4fb6-bd5d-1254bebb0ca0:0
> s 
> 

[ovirt-users] [ANN] oVirt 3.6.5 Third Release Candidate is now available for testing

2016-04-12 Thread Sandro Bonazzola
The oVirt Project is pleased to announce the availability of the Third
Release Candidate of oVirt 3.6.5 for testing, as of April 12th, 2016

This release is available now for:
* Fedora 22
* Red Hat Enterprise Linux 6.7
* CentOS Linux 6.7 (or similar)
* Red Hat Enterprise Linux 7.2 or later
* CentOS Linux (or similar) 7.2 or later

This release supports Hypervisor Hosts running:
* Red Hat Enterprise Linux 7.2 or later
* CentOS Linux (or similar) 7.2 or later
* Fedora 22

This release is also available with experimental support for:
* Debian 8.3 Jessie

This release candidate includes the following updated packages:
* ovirt-engine
* mom
* vdsm
* qemu-kvm-ev
* ovirt-hosted-engine-ha

See the release notes [1] for installation / upgrade instructions and a
list of new features and bugs fixed.

Notes:
* A new oVirt Live ISO is already available [2].
* Mirrors[3] might need up to one day to synchronize.

Additional Resources:
* Read more about the oVirt 3.6.5 release highlights:
http://www.ovirt.org/release/3.6.5/
* Get more oVirt Project updates on Twitter: https://twitter.com/ovirt
* Check out the latest project news on the oVirt blog:
http://www.ovirt.org/blog/

[1] http://www.ovirt.org/release/3.6.5/
[2] http://resources.ovirt.org/pub/ovirt-3.6-pre/iso/
[3] http://www.ovirt.org/Repository_mirrors#Current_mirrors


-- 
Sandro Bonazzola
Better technology. Faster innovation. Powered by community collaboration.
See how it works at redhat.com
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Inconsistency beteween hosted engine id and sanlock id

2016-04-12 Thread Baptiste Agasse
Hi all,

Last week we had problem on our ovirt infrastructure. The hosted engine didn't 
came up after the reboot of the host which hosted it. With the help of some 
people on #ovirt IRC channel (msivak, nsoffer and some others, thank to all of 
them) i managed to have my hosted engine up and running, but the underlying 
problem is still there. I think there is an inconsistency between sanlock ID of 
the hosts.

Some background:

We installed ovirt in 3.5 on CentOS 7 about 9 month ago. We have one DC with 
two clusters:

cluster 1: 4 hosts (virt1, virt2, virt3, virt4) that were installed with 
'hosted-engine --deploy' so there are capable to run the engine VM.
cluster 2: 2 hosts (virt6 and virt7) that were installed via the webui, so are 
'normal' ovirt hosts.

Since that we have successfully upgraded ovirt to 3.6 and set our cluster to 
3.6 compatibility mode.

Some weeks after something broke and the virt4 host rebooted. After some help 
on the IRC channel, i managed to get the engine vm up and running. After that i 
dug into the problem that seems to be around the sanlock part.

After explanations, that i understand is:

sanlock manage locks at DC level. there is an hosted_engine lock to manage who 
run the VM and there is a vdsm level lock on the hosted_engine disk (or any 
other VM disks) to know who can write to the disk.

The problem in my case is that on some hosts that were installed in 3.5, the 
hosted_engine ID and the vds_spm_id are not the same, and some other host have 
it vds_spm_id identical to some other host hosted_engine_id. So in some case, 
some host can't acquire the lock on some disks and have different IDs in the 
sanlock space.

Example, im my case:

#
# For the hosted_engine hosts:
#
[root@virt1 ~]# grep host_id /etc/ovirt-hosted-engine/hosted-engine.conf
host_id=1

[root@virt2 ~]# grep host_id /etc/ovirt-hosted-engine/hosted-engine.conf
host_id=2

[root@virt3 ~]# grep host_id /etc/ovirt-hosted-engine/hosted-engine.conf
host_id=3

[root@virt4 ~]# grep host_id /etc/ovirt-hosted-engine/hosted-engine.conf
host_id=4

#
# For all hosts, including hosted engine:
#
[root@virt1 ~]# sanlock client status   

   
daemon 3a99892c-5d3a-4d3d-bac7-d35259363c98.virt1
p -1 helper
p -1 listener
p -1 status
s 
hosted-engine:1:/var/run/vdsm/storage/377ae8e8-0eeb-4591-b50f-3d21298b4146/607719dd-b71e-4527-814a-964ed0c1f8ea/6a0b878d-fe7e-4fb6-bd5d-1254bebb0ca0:0
s 
295207d7-41ea-4cda-a028-f860c357d46b:1:/dev/295207d7-41ea-4cda-a028-f860c357d46b/ids:0
s 
daf1b53c-7e29-4b18-a9e2-910605cc7080:1:/dev/daf1b53c-7e29-4b18-a9e2-910605cc7080/ids:0
s 
680d5ed1-ed70-4340-a430-ddfa39ee3052:1:/dev/680d5ed1-ed70-4340-a430-ddfa39ee3052/ids:0
s 
350e5736-41c0-4017-a8fd-9866edad:1:/dev/350e5736-41c0-4017-a8fd-9866edad/ids:0
s 
377ae8e8-0eeb-4591-b50f-3d21298b4146:1:/dev/377ae8e8-0eeb-4591-b50f-3d21298b4146/ids:0

[root@virt2 ~]# sanlock client status   

   
daemon 48fe11a1-6c64-4a56-abf0-6f9690e6a8c2.virt2
p -1 helper
p -1 listener
p -1 status
s 
hosted-engine:2:/var/run/vdsm/storage/377ae8e8-0eeb-4591-b50f-3d21298b4146/607719dd-b71e-4527-814a-964ed0c1f8ea/6a0b878d-fe7e-4fb6-bd5d-1254bebb0ca0:0
s 
377ae8e8-0eeb-4591-b50f-3d21298b4146:2:/dev/377ae8e8-0eeb-4591-b50f-3d21298b4146/ids:0
s 
295207d7-41ea-4cda-a028-f860c357d46b:3:/dev/295207d7-41ea-4cda-a028-f860c357d46b/ids:0
s 
daf1b53c-7e29-4b18-a9e2-910605cc7080:3:/dev/daf1b53c-7e29-4b18-a9e2-910605cc7080/ids:0
s 
680d5ed1-ed70-4340-a430-ddfa39ee3052:3:/dev/680d5ed1-ed70-4340-a430-ddfa39ee3052/ids:0
s 
350e5736-41c0-4017-a8fd-9866edad:3:/dev/350e5736-41c0-4017-a8fd-9866edad/ids:0
r 
350e5736-41c0-4017-a8fd-9866edad:SDM:/dev/350e5736-41c0-4017-a8fd-9866edad/leases:1048576:26
 p 9304
r 
377ae8e8-0eeb-4591-b50f-3d21298b4146:d704cf05-e294-4ada-9627-920c9997cf22:/dev/377ae8e8-0eeb-4591-b50f-3d21298b4146/leases:49056:21
 p 32747

[root@virt3 ~]# sanlock client status   

   
daemon 3388d8e5-922d-45ab-8ecb-6e321a7a8a4a.virt3
p -1 helper
p -1 listener
p -1 status
s 
daf1b53c-7e29-4b18-a9e2-910605cc7080:2:/dev/daf1b53c-7e29-4b18-a9e2-910605cc7080/ids:0
s 
680d5ed1-ed70-4340-a430-ddfa39ee3052:2:/dev/680d5ed1-ed70-4340-a430-ddfa39ee3052/ids:0
s 
295207d7-41ea-4cda-a028-f860c357d46b:2:/dev/295207d7-41ea-4cda-a028-f860c357d46b/ids:0
s 
350e5736-41c0-4017-a8fd-9866edad:2:/dev/350e5736-41c0-4017-a8fd-9866edad/ids:0
s 

Re: [ovirt-users] HA agent fails to start

2016-04-12 Thread Simone Tiraboschi
Everything seams fine here,
/etc/ovirt-hosted-engine/hosted-engine.conf seams to be correctly
created with the right name.
Can you please check the latest modification time of your
/etc/ovirt-hosted-engine/hosted-engine.conf~ and compare it with the
setup time?

On Tue, Apr 12, 2016 at 2:34 PM, Richard Neuboeck  wrote:
> On 04/12/2016 11:32 AM, Simone Tiraboschi wrote:
>> On Mon, Apr 11, 2016 at 8:11 AM, Richard Neuboeck  
>> wrote:
>>> Hi oVirt Group,
>>>
>>> in my attempts to get all aspects of oVirt 3.6 up and running I
>>> stumbled upon something I'm not sure how to fix:
>>>
>>> Initially I installed a hosted engine setup. After that I added
>>> another HA host (with hosted-engine --deploy). The host was
>>> registered in the Engine correctly and HA agent came up as expected.
>>>
>>> However if I reboot the second host (through the Engine UI or
>>> manually) HA agent fails to start. The reason seems to be that
>>> /etc/ovirt-hosted-engine/hosted-engine.conf is empty. The backup
>>> file ending with ~ exists though.
>>
>> Can you please attach hosted-engine-setup logs from your additional hosts?
>> AFAIK our code will never take a ~ ending backup of that file.
>
> ovirt-hosted-engine-setup logs from both additional hosts are
> attached to this mail.
>
>>
>>> Here are the log messages from the journal:
>>> Apr 11 07:29:39 cube-two.tbi.univie.ac.at systemd[1]: Starting oVirt
>>> Hosted Engine High Availability Monitoring Agent...
>>> Apr 11 07:29:39 cube-two.tbi.univie.ac.at ovirt-ha-agent[3747]:
>>> INFO:ovirt_hosted_engine_ha.agent.agent.Agent:ovirt-hosted-engine-ha
>>> agent 1.3.5.3-0.0.master started
>>> Apr 11 07:29:39 cube-two.tbi.univie.ac.at ovirt-ha-agent[3747]:
>>> INFO:ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine:Found
>>> certificate common name: cube-two.tbi.univie.ac.at
>>> Apr 11 07:29:39 cube-two.tbi.univie.ac.at ovirt-ha-agent[3747]:
>>> ovirt-ha-agent
>>> ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine ERROR Hosted
>>> Engine is not configured. Shutting down.
>>> Apr 11 07:29:39 cube-two.tbi.univie.ac.at ovirt-ha-agent[3747]:
>>> ERROR:ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine:Hosted
>>> Engine is not configured. Shutting down.
>>> Apr 11 07:29:39 cube-two.tbi.univie.ac.at ovirt-ha-agent[3747]:
>>> INFO:ovirt_hosted_engine_ha.agent.agent.Agent:Agent shutting down
>>> Apr 11 07:29:39 cube-two.tbi.univie.ac.at systemd[1]:
>>> ovirt-ha-agent.service: main process exited, code=exited, status=255/n/a
>>>
>>> If I restore the configuration from the backup file and manually
>>> restart the HA agent it's working properly.
>>>
>>> For testing purposes I added a third HA host which turn out to
>>> behave exactly the same.
>>>
>>> Any help would be appreciated!
>>> Thanks
>>> Cheers
>>> Richard
>>>
>>> --
>>> /dev/null
>>>
>>>
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>>
>
>
> --
> /dev/null
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted engine on gluster problem

2016-04-12 Thread Luiz Claudio Prazeres Goncalves
Hi Sandro, I've been using gluster with 3 external hosts for a while and
things are working pretty well, however this single point of failure looks
like a simple feature to implement,but critical to anyone who wants to use
gluster on production  . This is not hyperconvergency which has other
issues/implications. So , why not have this feature out on 3.6 branch? It
looks like just let vdsm use the 'backupvol-server' option when mounting
the engine domain and make the property tests.

Could you add this feature to the next release of 3.6 branch?

Thanks
Luiz

Em ter, 12 de abr de 2016 05:03, Sandro Bonazzola 
escreveu:

> On Mon, Apr 11, 2016 at 11:44 PM, Bond, Darryl 
> wrote:
>
>> My setup is hyperconverged. I have placed my test results in
>> https://bugzilla.redhat.com/show_bug.cgi?id=1298693
>>
>>
> Ok, so you're aware about the limitation of the single point of failure.
> If you drop the host referenced in hosted engine configuration for the
> initial setup it won't be able to connect to shared storage even if the
> other hosts in the cluster are up since the entry point is down.
> Note that hyperconverged deployment is not supported in 3.6.
>
>
>
>>
>> Short description of setup:
>>
>> 3 hosts with 2 disks each set up with gluster replica 3 across the 6
>> disks volume name hosted-engine.
>>
>> Hostname hosted-storage configured in /etc//hosts to point to the host1.
>>
>> Installed hosted engine on host1 with the hosted engine storage path =
>> hosted-storage:/hosted-engine
>>
>> Install first engine on h1 successful. Hosts h2 and h3 added to the
>> hosted engine. All works fine.
>>
>> Additional storage and non-hosted engine hosts added etc.
>>
>> Additional VMs added to hosted-engine storage (oVirt Reports VM and
>> Cinder VM). Additional VM's are hosted by other storage - cinder and NFS.
>>
>> The system is in production.
>>
>>
>> Engine can be migrated around with the web interface.
>>
>>
>> - 3.6.4 upgrade released, follow the upgrade guide, engine is upgraded
>> first , new Centos kernel requires host reboot.
>>
>> - Engine placed on h2 -  h3 into maintenance (local) upgrade and Reboot
>> h3 - No issues - Local maintenance removed from h3.
>>
>> - Engine placed on h3 -  h2 into maintenance (local) upgrade and Reboot
>> h2 - No issues - Local maintenance removed from h2.
>>
>> - Engine placed on h3 -h1 into mainteance (local) upgrade and reboot h1 -
>> engine crashes and does not start elsewhere, VM(cinder)  on h3 on same
>> gluster volume pauses.
>>
>> - Host 1 takes about 5 minutes to reboot (Enterprise box with all it's
>> normal BIOS probing)
>>
>> - Engine starts after h1 comes back and stabilises
>>
>> - VM(cinder) unpauses itself,  VM(reports) continued fine the whole time.
>> I can do no diagnosis on the 2 VMs as the engine is not available.
>>
>> - Local maintenance removed from h​1
>>
>>
>> I don't believe the issue is with gluster itself as the volume remains
>> accessible on all hosts during this time albeit with a missing server
>> (gluster volume status) as each gluster server is rebooted.
>>
>> Gluster was upgraded as part of the process, no issues were seen here.
>>
>>
>> I have been able to duplicate the issue without the upgrade by following
>> the same sort of timeline.
>>
>>
>> 
>> From: Sandro Bonazzola 
>> Sent: Monday, 11 April 2016 7:11 PM
>> To: Richard Neuboeck; Simone Tiraboschi; Roy Golan; Martin Sivak; Sahina
>> Bose
>> Cc: Bond, Darryl; users
>> Subject: Re: [ovirt-users] Hosted engine on gluster problem
>>
>>
>>
>> On Mon, Apr 11, 2016 at 9:37 AM, Richard Neuboeck > > wrote:
>> Hi Darryl,
>>
>> I'm still experimenting with my oVirt installation so I tried to
>> recreate the problems you've described.
>>
>> My setup has three HA hosts for virtualization and three machines
>> for the gluster replica 3 setup.
>>
>> I manually migrated the Engine from the initial install host (one)
>> to host three. Then shut down host one manually and interrupted the
>> fencing mechanisms so the host stayed down. This didn't bother the
>> Engine VM at all.
>>
>> Did you move the host one to maintenance before shutting down?
>> Or is this a crash recovery test?
>>
>>
>>
>> To make things a bit more challenging I then shut down host three
>> while running the Engine VM. Of course the Engine was down for some
>> time until host two detected the problem. It started the Engine VM
>> and everything seems to be running quite well without the initial
>> install host.
>>
>> Thanks for the feedback!
>>
>>
>>
>> My only problem is that the HA agent on host two and three refuse to
>> start after a reboot due to the fact that the configuration of the
>> hosted engine is missing. I wrote another mail to users@ovirt.org> users@ovirt.org>
>> about that.
>>
>> This is weird. Martin,  Simone can you please investigate on this?
>>
>>
>>
>>
>> Cheers
>> 

Re: [ovirt-users] Looking for Python-SDK documentation

2016-04-12 Thread Alexander Wels
On Monday, April 11, 2016 05:03:28 PM Brett I. Holcomb wrote:
> On Mon, 2016-04-11 at 20:41 +0200, Frank Thommen wrote:
> > On 11.04.16 20:17, Brett I. Holcomb wrote:
> > > On Mon, 2016-04-11 at 14:38 -0300, Amador Pahim wrote:
> > > > On 04/11/2016 02:07 PM, Brett I. Holcomb wrote:
> > > > > On Mon, 2016-04-11 at 17:27 +0200, Frank Thommen wrote:
> > > > > > Dear all,
> > > > > > 
> > > > > > I'm currently lost at finding any documentation about the
> > > > > > Python SDK
> > > > > > (http://www.ovirt.org/develop/release-management/features/inf
> > > > > > ra/python-sdk/)
> > > > > > like provided classes, functions etc..  There are some
> > > > > > examples on the
> > > > > > mentioned page, but I cannot find a complete
> > > > > > documentation.  Our oVirt
> > > > > > server is running CentOS 7 with ovirt-engine-sdk-python
> > > > > > installed.
> > > > > > However there doesn't seem to exist an ovirt-engine-sdk-
> > > > > > python-docs
> > > > > > package and I couldn't find any appropriate link on the oVirt
> > > > > > documentation pages (http://www.ovirt.org/documentation/).
> > > > > > 
> > > > > > Any ideas, where the documentation is available?
> > > > > > 
> > > > > > Cheers
> > > > > > Frank
> > > > > > ___
> > > > > > Users mailing list
> > > > > > Users@ovirt.org 
> > > > > > http://lists.ovirt.org/mailman/listinfo/users
> > > > > > 
> > > > > 
> > > > > Other than what you found I found this but that's all.  Doesn't
> > > > > seem
> > > > > to be much other than examples an the one that shows what to
> > > > > import.
> > > > >  I, too would like to find what your are looking for so I can
> > > > > use it.
> > > > 
> > > > There is this RHEV documentation that can be helpful:
> > > > https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_
> > > > Virtualization/3.3/html/Developer_Guide/chap-
> > > > Python_Quick_Start_Example.html
> > > > 
> > > > > http://www.ovirt.org/develop/api/pythonapi/
> > > > > 
> > > > > 
> > > > > ___
> > > > > Users mailing list
> > > > > Users@ovirt.org 
> > > > > http://lists.ovirt.org/mailman/listinfo/users
> > > > > 
> > > 
> > > Thanks.  That is helpful but as Frank said it would be nice to have
> > > an
> > > api reference for each class and function that gives the
> > > parameters,
> > > return values, and other stuff usually found in an API
> > > reference.  The
> > > examples are helpful but don't give all the information abut a
> > > function
> > > or class.
> > 
> > Thanks to all who answered.  Brett brings it to the point:  All sent 
> > links so far are indeed helpful - thanks a lot - but not the
> > reference I 
> > expected. 
> > https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Virt
> > ualization/3.6/html/Python_SDK_Guide/chap-
> > Python_Reference_Documentation.html#Python_Reference_Documentation 
> > mentions `pydoc`, but this documentation seems to be provided only
> > for 
> > some modules or to be incomplete.  Also for me not being a software 
> > developper and newish to Python, the `pydoc` information is not very 
> > useful.  Where can I e.g. find the documentation for vms.get() and 
> > vms.add() (just to name teo concrete examples)?
> > 
> > Frank
> > ___
> > Users mailing list
> > Users@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> 
> I'm in the same boat as Frank.  I've done programming in various
> languages since Fortran without the numbers  but only when needed
> for my job as an Engineer so I'm not a professional but just trying to
> get a job done.  It would be nice to have a full reference so we know
> what to provide.  When trying to connect with the api I finally figured
> out to use ca_file (like ca-file on the command line).  Raz's reference
> is more complete but still leaves a lot out.  The newer equivalent of
> Raz's reference seems to be http://www.ovirt.org/develop/release-manage
> ment/features/infra/python-sdk/.

The Python/Java/Ruby SDKs are simply language specific wrappers around the REST 
api. So if you want a full list of all the options available check out the 
REST api documentation. You will have to translate a particular REST api 
field/feature to the SDK, but all the SDKs are generated from the REST api 
interface definition so the naming and everything should be the same.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Cleanup Ovirt Node

2016-04-12 Thread Fabian Deutsch
Heya,

I am not sure if you want to uninstall/remove a node - then you need
to clear the storage - or if you want to reset it - in this case I'd
recommend to reinstall.

- fabian

On Mon, Apr 11, 2016 at 3:27 PM, Marcelo Leandro  wrote:
> Good Morning ,
> how do I clear the configurations of oVirt node?
>
> Thanks,
>
> Marcelo Leandro
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>



-- 
Fabian Deutsch 
RHEV Hypervisor
Red Hat
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Multi-node cluster with local storage

2016-04-12 Thread Christophe TREFOIS
Hi Pavel,

engine is running on one of the hosts. We don’t have HA for host, but I think 
it is OK.

We used a dedicated NFS for the hosted-engine on that host.

Best,

Dr Christophe Trefois, Dipl.-Ing.
Technical Specialist / Post-Doc

UNIVERSITÉ DU LUXEMBOURG

LUXEMBOURG CENTRE FOR SYSTEMS BIOMEDICINE
Campus Belval | House of Biomedicine
6, avenue du Swing
L-4367 Belvaux
T: +352 46 66 44 6124
F: +352 46 66 44 6949
http://www.uni.lu/lcsb

[Facebook]  [Twitter] 
   [Google Plus] 
   [Linkedin] 
   [skype] 



This message is confidential and may contain privileged information.
It is intended for the named recipient only.
If you receive it in error please notify me and permanently delete the original 
message and any copies.




On 04 Mar 2016, at 07:46, Pavel Gashev 
> wrote:


Christophe,

Where is the engine running? Did you use one of NFSs for hosted-engine?

Thank you


From: Christophe TREFOIS 
>
Sent: Mar 4, 2016 2:09 AM
To: Pavel Gashev; users@ovirt.org
Subject: RE: Multi-node cluster with local storage

Hi,

We have all nodes like you (5 currently), and have them setup as NFS instead of 
local storage.

You have slight overhead because it's NFS, but not so much, and it allows you 
to migrate VMs and disks, even if it's a manual disk, eg no HA will be possible.

So yeah, to me, it's a good option given your setup.

Best,

> -Original Message-
> From: users-boun...@ovirt.org 
> [mailto:users-boun...@ovirt.org] On Behalf
> Of Pavel Gashev
> Sent: jeudi 3 mars 2016 21:44
> To: users@ovirt.org
> Subject: [ovirt-users] Multi-node cluster with local storage
>
> Hello,
>
> I'd like to ask community, what is the best way to use oVirt in the following
> hardware configuration:
>
> Three servers connected 1GB network. Each server - 32 threads, 256GB RAM,
> 4TB RAID.
>
> Please note that a local storage and an 1GB network is a typical hardware
> configuration for almost any dedicated hosting.
>
> Unfortunately, oVirt doesn't support multi-node local storage clusters.
> And Gluster/CEPH doesn't work well over 1G network. It looks like that the
> only way to use oVirt in a three-node cluster is to share local storages over
> NFS. At least it makes possible to migrate VMs and move disks among
> hardware nodes.
>
> Does somebody have such setup?
>
> Thanks
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] ovirt glusterfs performance

2016-04-12 Thread Niels de Vos
On Tue, Apr 12, 2016 at 11:11:54AM +0200, Roderick Mooi wrote:
> Hi
> 
> > It is not removed. Can you try 'gluster volume set volname 
> > cluster.eager-lock enable`?
> 
> This works. BTW by default this setting is “on”

Thanks for reporting back!

> What’s the difference between “on” and “enable”?

Both are the same, you could also use "yes", "true" and possibly others.

Cheers,
Niels


> 
> Thanks for the clarification.
> 
> Regards,
> 
> Roderick
> 
> > On 06 Apr 2016, at 10:56 AM, Ravishankar N  wrote:
> > 
> > On 04/06/2016 02:08 PM, Roderick Mooi wrote:
> >> Hi Ravi and colleagues
> >> 
> >> (apologies for hijacking this thread but I’m not sure where else to report 
> >> this (and it is related).)
> >> 
> >> With gluster 3.7.10, running
> >> #gluster volume set  group virt
> >> fails with:
> >> volume set: failed: option : eager-lock does not exist
> >> Did you mean eager-lock?
> >> 
> >> I had to remove the eager-lock setting from /var/lib/glusterd/groups/virt 
> >> to get this to work. It seems like setting eager-lock has been removed 
> >> from latest gluster. Is this correct? Either way, is there anything else I 
> >> should do?
> > 
> > It is not removed. Can you try 'gluster volume set volname 
> > cluster.eager-lock enable`?
> > I think the disperse (EC) translator introduced a `disperse.eager-lock` 
> > which is why you would need to mention entire volume option name to avoid 
> > ambiguity.
> > We probably need to fix the virt profile setting to include the entire 
> > name. By the way 'gluster volume set help` should give you the list of all 
> > options.
> > 
> > -Ravi
> > 
> >> 
> >> Cheers,
> >> 
> >> Roderick
> >> 
> >>> On 12 Feb 2016, at 6:18 AM, Ravishankar N  >>> > wrote:
> >>> 
> >>> Hi Bill,
> >>> Can you enable virt-profile setting for your volume and see if that 
> >>> helps? You need to enable this optimization when you create the volume 
> >>> using ovrit, or use the following command for an existing volume:
> >>> 
> >>> #gluster volume set  group virt
> >>> 
> >>> -Ravi
> >>> 
> >>> 
> >>> On 02/12/2016 05:22 AM, Bill James wrote:
>  My apologies, I'm showing how much of a noob I am.
>  Ignore last direct to gluster numbers, as that wasn't really glusterfs.
>  
>  
>  [root@ovirt2 test ~]# mount -t glusterfs ovirt2-ks.test.j2noc.com 
>  :/gv1 /mnt/tmp/
>  [root@ovirt2 test ~]# time dd if=/dev/zero of=/mnt/tmp/testfile2 bs=1M 
>  count=1000 oflag=direct
>  1048576000 bytes (1.0 GB) copied, 65.8596 s, 15.9 MB/s
>  
>  That's more how I expected, it is pointing to glusterfs performance.
>  
>  
>  
>  On 02/11/2016 03:27 PM, Bill James wrote:
> > don't know if it helps, but I ran a few more tests, all from the same 
> > hardware node.
> > 
> > The VM:
> > [root@billjov1 ~]# time dd if=/dev/zero of=/root/testfile bs=1M 
> > count=1000 oflag=direct
> > 1048576000 bytes (1.0 GB) copied, 62.5535 s, 16.8 MB/s
> > 
> > Writing directly to gluster volume:
> > [root@ovirt2 test ~]# time dd if=/dev/zero 
> > of=/gluster-store/brick1/gv1/testfile bs=1M count=1000 oflag=direct
> > 1048576000 bytes (1.0 GB) copied, 9.92048 s, 106 MB/s
> > 
> > 
> > Writing to NFS volume:
> > [root@ovirt2 test ~]# time dd if=/dev/zero of=/mnt/storage/qa/testfile 
> > bs=1M count=1000 oflag=direct
> > 1048576000 bytes (1.0 GB) copied, 10.5776 s, 99.1 MB/s
> > 
> > NFS & Gluster are using the same interface. Tests were not run at same 
> > time.
> > 
> > This would suggest my problem isn't glusterfs, but the VM performance.
> > 
> > 
> > 
> > On 02/11/2016 03:13 PM, Bill James wrote:
> >> xml attached. 
> >> 
> >> 
> >> On 02/11/2016 12:28 PM, Nir Soffer wrote: 
> >>> On Thu, Feb 11, 2016 at 8:27 PM, Bill James  
> >>>  
> >>>  wrote: 
>  thank you for the reply. 
>  
>  We setup gluster using the names associated with  NIC 2 IP. 
>    Brick1: ovirt1-ks.test.j2noc.com 
>  :/gluster-store/brick1/gv1 
>    Brick2: ovirt2-ks.test.j2noc.com 
>  :/gluster-store/brick1/gv1 
>    Brick3: ovirt3-ks.test.j2noc.com 
>  :/gluster-store/brick1/gv1 
>  
>  That's NIC 2's IP. 
>  Using 'iftop -i eno2 -L 5 -t' : 
>  
>  dd if=/dev/zero of=/root/testfile bs=1M count=1000 oflag=direct 
>  1048576000 bytes (1.0 GB) copied, 68.0714 s, 15.4 MB/s 
> >>> Can you share the xml of this vm? You can find it in vdsm log, 
> >>> at the time you start the vm. 
> >>> 
> >>> Or you can do (on the host): 
> >>> 
> >>> # 

[ovirt-users] what is detach_only with VM delete

2016-04-12 Thread Fabrice Bacchella
I'm using the python sdk to detroy VM.

I supposed the param action.vm.disks.detach_only is the same thing that the 
check box "Remove Disk(s)"

But once i try to manage it with:
action_params = params.Action(
vm=params.VM(
disks=params.Disks(
detach_only=detach_only_flag,
),
),
)

The disk is not destroyed, whatever detach_only_flag is True of False. If 
detach_only is not defined in params.Disks, everything is fine, the disk is 
destroyed.

But the included comments in brokers.py says:
[@param action.vm.disks.detach_only: boolean]

Did I miss something ?

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] ovirt glusterfs performance

2016-04-12 Thread Ravishankar N

On 04/12/2016 02:41 PM, Roderick Mooi wrote:

Hi

It is not removed. Can you try 'gluster volume set volname 
cluster.eager-lock enable`?


This works. BTW by default this setting is “on”. What’s the difference 
between “on” and “enable”?


Both are identical. You can use any of the booleans to achieve the same 
effect. {"1", "on", "yes", "true", "enable"} or {"0", "off", "no", 
"false", "disable"}
FYI, the patch http://review.gluster.org/#/c/13958/ to fix this issue 
should make it to glusterfs-3.7.11.

-Ravi


Thanks for the clarification.

Regards,

Roderick

On 06 Apr 2016, at 10:56 AM, Ravishankar N > wrote:


On 04/06/2016 02:08 PM, Roderick Mooi wrote:

Hi Ravi and colleagues

(apologies for hijacking this thread but I’m not sure where else to 
report this (and it is related).)


With gluster 3.7.10, running
#gluster volume set  group virt
fails with:
volume set: failed: option : eager-lock does not exist
Did you mean eager-lock?

I had to remove the eager-lock setting from 
/var/lib/glusterd/groups/virt to get this to work. It seems like 
setting eager-lock has been removed from latest gluster. Is this 
correct? Either way, is there anything else I should do?


It is not removed. Can you try 'gluster volume set volname 
cluster.eager-lock enable`?
I think the disperse (EC) translator introduced a 
`disperse.eager-lock` which is why you would need to mention entire 
volume option name to avoid ambiguity.
We probably need to fix the virt profile setting to include the 
entire name. By the way 'gluster volume set help` should give you the 
list of all options.


-Ravi



Cheers,

Roderick

On 12 Feb 2016, at 6:18 AM, Ravishankar N > wrote:


Hi Bill,
Can you enable virt-profile setting for your volume and see if that 
helps? You need to enable this optimization when you create the 
volume using ovrit, or use the following command for an existing 
volume:


#gluster volume set  group virt

-Ravi


On 02/12/2016 05:22 AM, Bill James wrote:

My apologies, I'm showing how much of a noob I am.
Ignore last direct to gluster numbers, as that wasn't really 
glusterfs.



[root@ovirt2 test ~]# mount -t glusterfs ovirt2-ks.test.j2noc.com 
:/gv1 /mnt/tmp/
[root@ovirt2 test ~]# time dd if=/dev/zero of=/mnt/tmp/testfile2 
bs=1M count=1000 oflag=direct

1048576000 bytes (1.0 GB) copied, 65.8596 s, 15.9 MB/s

That's more how I expected, it is pointing to glusterfs performance.



On 02/11/2016 03:27 PM, Bill James wrote:
don't know if it helps, but I ran a few more tests, all from the 
same hardware node.


The VM:
[root@billjov1 ~]# time dd if=/dev/zero of=/root/testfile bs=1M 
count=1000 oflag=direct

1048576000 bytes (1.0 GB) copied, 62.5535 s, 16.8 MB/s

Writing directly to gluster volume:
[root@ovirt2 test ~]# time dd if=/dev/zero 
of=/gluster-store/brick1/gv1/testfile bs=1M count=1000 oflag=direct

1048576000 bytes (1.0 GB) copied, 9.92048 s, 106 MB/s


Writing to NFS volume:
[root@ovirt2 test ~]# time dd if=/dev/zero 
of=/mnt/storage/qa/testfile bs=1M count=1000 oflag=direct

1048576000 bytes (1.0 GB) copied, 10.5776 s, 99.1 MB/s

NFS & Gluster are using the same interface. Tests were not run at 
same time.


This would suggest my problem isn't glusterfs, but the VM 
performance.




On 02/11/2016 03:13 PM, Bill James wrote:

xml attached.


On 02/11/2016 12:28 PM, Nir Soffer wrote:
On Thu, Feb 11, 2016 at 8:27 PM, Bill James  
wrote:

thank you for the reply.

We setup gluster using the names associated with  NIC 2 IP.
  Brick1: ovirt1-ks.test.j2noc.com 
:/gluster-store/brick1/gv1
  Brick2: ovirt2-ks.test.j2noc.com 
:/gluster-store/brick1/gv1
  Brick3: ovirt3-ks.test.j2noc.com 
:/gluster-store/brick1/gv1


That's NIC 2's IP.
Using 'iftop -i eno2 -L 5 -t' :

dd if=/dev/zero of=/root/testfile bs=1M count=1000 oflag=direct
1048576000 bytes (1.0 GB) copied, 68.0714 s, 15.4 MB/s

Can you share the xml of this vm? You can find it in vdsm log,
at the time you start the vm.

Or you can do (on the host):

# virsh
virsh # list
(username: vdsm@ovirt password: shibboleth)
virsh # dumpxml vm-id


Peak rate (sent/received/total): 281Mb 5.36Mb
282Mb
Cumulative (sent/received/total): 1.96GB 14.6MB
1.97GB

gluster volume info gv1:
  Options Reconfigured:
performance.write-behind-window-size: 4MB
  performance.readdir-ahead: on
  performance.cache-size: 1GB
  performance.write-behind: off

performance.write-behind: off didn't help.
Neither did any other changes I've tried.


There is no VM traffic on this VM right now except my test.



On 02/10/2016 11:55 PM, Nir Soffer wrote:
On Thu, Feb 11, 2016 at 2:42 AM, Ravishankar N 


wrote:

+gluster-users

Does disabling 'performance.write-behind' give a better 
throughput?




On 02/10/2016 

Re: [ovirt-users] HA agent fails to start

2016-04-12 Thread Simone Tiraboschi
On Mon, Apr 11, 2016 at 8:11 AM, Richard Neuboeck  wrote:
> Hi oVirt Group,
>
> in my attempts to get all aspects of oVirt 3.6 up and running I
> stumbled upon something I'm not sure how to fix:
>
> Initially I installed a hosted engine setup. After that I added
> another HA host (with hosted-engine --deploy). The host was
> registered in the Engine correctly and HA agent came up as expected.
>
> However if I reboot the second host (through the Engine UI or
> manually) HA agent fails to start. The reason seems to be that
> /etc/ovirt-hosted-engine/hosted-engine.conf is empty. The backup
> file ending with ~ exists though.

Can you please attach hosted-engine-setup logs from your additional hosts?
AFAIK our code will never take a ~ ending backup of that file.

> Here are the log messages from the journal:
> Apr 11 07:29:39 cube-two.tbi.univie.ac.at systemd[1]: Starting oVirt
> Hosted Engine High Availability Monitoring Agent...
> Apr 11 07:29:39 cube-two.tbi.univie.ac.at ovirt-ha-agent[3747]:
> INFO:ovirt_hosted_engine_ha.agent.agent.Agent:ovirt-hosted-engine-ha
> agent 1.3.5.3-0.0.master started
> Apr 11 07:29:39 cube-two.tbi.univie.ac.at ovirt-ha-agent[3747]:
> INFO:ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine:Found
> certificate common name: cube-two.tbi.univie.ac.at
> Apr 11 07:29:39 cube-two.tbi.univie.ac.at ovirt-ha-agent[3747]:
> ovirt-ha-agent
> ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine ERROR Hosted
> Engine is not configured. Shutting down.
> Apr 11 07:29:39 cube-two.tbi.univie.ac.at ovirt-ha-agent[3747]:
> ERROR:ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine:Hosted
> Engine is not configured. Shutting down.
> Apr 11 07:29:39 cube-two.tbi.univie.ac.at ovirt-ha-agent[3747]:
> INFO:ovirt_hosted_engine_ha.agent.agent.Agent:Agent shutting down
> Apr 11 07:29:39 cube-two.tbi.univie.ac.at systemd[1]:
> ovirt-ha-agent.service: main process exited, code=exited, status=255/n/a
>
> If I restore the configuration from the backup file and manually
> restart the HA agent it's working properly.
>
> For testing purposes I added a third HA host which turn out to
> behave exactly the same.
>
> Any help would be appreciated!
> Thanks
> Cheers
> Richard
>
> --
> /dev/null
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted engine on gluster problem

2016-04-12 Thread Sandro Bonazzola
On Mon, Apr 11, 2016 at 11:44 PM, Bond, Darryl  wrote:

> My setup is hyperconverged. I have placed my test results in
> https://bugzilla.redhat.com/show_bug.cgi?id=1298693
>
>
Ok, so you're aware about the limitation of the single point of failure. If
you drop the host referenced in hosted engine configuration for the initial
setup it won't be able to connect to shared storage even if the other hosts
in the cluster are up since the entry point is down.
Note that hyperconverged deployment is not supported in 3.6.



>
> Short description of setup:
>
> 3 hosts with 2 disks each set up with gluster replica 3 across the 6 disks
> volume name hosted-engine.
>
> Hostname hosted-storage configured in /etc//hosts to point to the host1.
>
> Installed hosted engine on host1 with the hosted engine storage path =
> hosted-storage:/hosted-engine
>
> Install first engine on h1 successful. Hosts h2 and h3 added to the hosted
> engine. All works fine.
>
> Additional storage and non-hosted engine hosts added etc.
>
> Additional VMs added to hosted-engine storage (oVirt Reports VM and Cinder
> VM). Additional VM's are hosted by other storage - cinder and NFS.
>
> The system is in production.
>
>
> Engine can be migrated around with the web interface.
>
>
> - 3.6.4 upgrade released, follow the upgrade guide, engine is upgraded
> first , new Centos kernel requires host reboot.
>
> - Engine placed on h2 -  h3 into maintenance (local) upgrade and Reboot h3
> - No issues - Local maintenance removed from h3.
>
> - Engine placed on h3 -  h2 into maintenance (local) upgrade and Reboot h2
> - No issues - Local maintenance removed from h2.
>
> - Engine placed on h3 -h1 into mainteance (local) upgrade and reboot h1 -
> engine crashes and does not start elsewhere, VM(cinder)  on h3 on same
> gluster volume pauses.
>
> - Host 1 takes about 5 minutes to reboot (Enterprise box with all it's
> normal BIOS probing)
>
> - Engine starts after h1 comes back and stabilises
>
> - VM(cinder) unpauses itself,  VM(reports) continued fine the whole time.
> I can do no diagnosis on the 2 VMs as the engine is not available.
>
> - Local maintenance removed from h​1
>
>
> I don't believe the issue is with gluster itself as the volume remains
> accessible on all hosts during this time albeit with a missing server
> (gluster volume status) as each gluster server is rebooted.
>
> Gluster was upgraded as part of the process, no issues were seen here.
>
>
> I have been able to duplicate the issue without the upgrade by following
> the same sort of timeline.
>
>
> 
> From: Sandro Bonazzola 
> Sent: Monday, 11 April 2016 7:11 PM
> To: Richard Neuboeck; Simone Tiraboschi; Roy Golan; Martin Sivak; Sahina
> Bose
> Cc: Bond, Darryl; users
> Subject: Re: [ovirt-users] Hosted engine on gluster problem
>
>
>
> On Mon, Apr 11, 2016 at 9:37 AM, Richard Neuboeck  > wrote:
> Hi Darryl,
>
> I'm still experimenting with my oVirt installation so I tried to
> recreate the problems you've described.
>
> My setup has three HA hosts for virtualization and three machines
> for the gluster replica 3 setup.
>
> I manually migrated the Engine from the initial install host (one)
> to host three. Then shut down host one manually and interrupted the
> fencing mechanisms so the host stayed down. This didn't bother the
> Engine VM at all.
>
> Did you move the host one to maintenance before shutting down?
> Or is this a crash recovery test?
>
>
>
> To make things a bit more challenging I then shut down host three
> while running the Engine VM. Of course the Engine was down for some
> time until host two detected the problem. It started the Engine VM
> and everything seems to be running quite well without the initial
> install host.
>
> Thanks for the feedback!
>
>
>
> My only problem is that the HA agent on host two and three refuse to
> start after a reboot due to the fact that the configuration of the
> hosted engine is missing. I wrote another mail to users@ovirt.org users@ovirt.org>
> about that.
>
> This is weird. Martin,  Simone can you please investigate on this?
>
>
>
>
> Cheers
> Richard
>
> On 04/08/2016 01:38 AM, Bond, Darryl wrote:
> > There seems to be a pretty severe bug with using hosted engine on
> gluster.
> >
> > If the host that was used as the initial hosted-engine --deploy host
> goes away, the engine VM wil crash and cannot be restarted until the host
> comes back.
>
> is this an Hyperconverged setup?
>
>
> >
> > This is regardless of which host the engine was currently running.
> >
> >
> > The issue seems to be buried in the bowels of VDSM and is not an issue
> with gluster itself.
>
> Sahina, can you please investigate on this?
>
>
> >
> > The gluster filesystem is still accessable from the host that was
> running the engine. The issue has been submitted to bugzilla but the fix is
> some way off (4.1).
> >
> >
> > Can my hosted engine