Re: [ovirt-users] Correct recovery procedure of the oVirt Hosted Engine 4.0

2016-10-05 Thread Simone Tiraboschi
On Thu, Oct 6, 2016 at 7:32 AM,  wrote:

> Hi Simone.
> When can we expect a new version of the engine-backup with built-in
> cleaning helper?
>

That bug is targeted to 4.1


>
> 05.10.2016, 13:52, "Simone Tiraboschi" :
> > On Wed, Oct 5, 2016 at 12:40 PM,  wrote:
> >> Ouch. It is beyond my understanding.
> >>
> >> Thus, it appears that described in the RHV4 guide (
> https://access.redhat.com/documentation/en/red-hat-
> virtualization/4.0/single/self-hosted-engine-guide/#
> sect-Restoring_SHE_bkup) recovery procedure in fact incomplete?
> > Yes, you are right although this is a kind of special case since we are
> moving/restoring to a different storage domain while you are not asked to
> remove the old storage if you are restoring in place.
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Correct recovery procedure of the oVirt Hosted Engine 4.0

2016-10-05 Thread aleksey . maksimov
Hi Simone. 
When can we expect a new version of the engine-backup with built-in cleaning 
helper?

05.10.2016, 13:52, "Simone Tiraboschi" :
> On Wed, Oct 5, 2016 at 12:40 PM,  wrote:
>> Ouch. It is beyond my understanding.
>>
>> Thus, it appears that described in the RHV4 guide 
>> (https://access.redhat.com/documentation/en/red-hat-virtualization/4.0/single/self-hosted-engine-guide/#sect-Restoring_SHE_bkup)
>>  recovery procedure in fact incomplete?
> Yes, you are right although this is a kind of special case since we are 
> moving/restoring to a different storage domain while you are not asked to 
> remove the old storage if you are restoring in place.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Ovirt setup

2016-10-05 Thread Bryan Sockel
Hi,

I am getting an error attempting to install ovirt on a pair of bonded nics.  
The error that we are getting is Cannot acquire nic/bridge address.

We typically run all our servers with an active/backup setup. ___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [Gluster-users] 4.0 - 2nd node fails on deploy

2016-10-05 Thread Joe Julian
"no route to host" is a network problem. Looks like quorum loss is appropriate. 

On October 5, 2016 12:31:18 PM GMT+02:00, Sahina Bose  wrote:
>On Wed, Oct 5, 2016 at 1:56 PM, Jason Jeffrey  wrote:
>
>> HI,
>>
>>
>>
>> Logs attached
>>
>
>Have you probed 2 interfaces for same host, that is - dcasrv02 and
>dcastor02? Does "gluster peer status" understand both names as for same
>host?
>
>From glusterd logs and the mount logs - the connection between the
>peers is
>lost, and quorum is lost, which is reaffirming what Simone said
>earlier.
>Logs seem to indicate network issues - check the direct link setup. See
>below
>
>From mount logs:
>[2016-10-04 17:26:15.718300] E [socket.c:2292:socket_connect_finish]
>0-engine-client-2: connection to 10.100.103.3:24007 failed (No route to
>host)
>[2016-10-04 17:26:15.718345] W [MSGID: 108001]
>[afr-common.c:4379:afr_notify] 0-engine-replicate-0: Client-quorum is
>not
>met
>[2016-10-04 17:26:16.428290] E [socket.c:2292:socket_connect_finish]
>0-engine-client-1: connection to 10.100.101.2:24007 failed (No route to
>host)
>[2016-10-04 17:26:16.428336] E [MSGID: 108006]
>[afr-common.c:4321:afr_notify] 0-engine-replicate-0: All subvolumes are
>down. Going offline until atleast one of them comes back up
>
>And in glusterd logs:
>[2016-10-04 17:24:39.522402] E [socket.c:2292:socket_connect_finish]
>0-management: connection to 10.100.50.82:24007 failed (No route to
>host)
>[2016-10-04 17:24:39.522578] I [MSGID: 106004]
>[glusterd-handler.c:5201:__glusterd_peer_rpc_notify] 0-management: Peer
> (<1e788fc9-dfe9-4753-92c7-76a95c8d0891>), in state Cluster>, has disconnected from glusterd.
>[2016-10-04 17:24:39.523272] C [MSGID: 106002]
>[glusterd-server-quorum.c:351:glusterd_do_volume_quorum_action]
>0-management: Server quorum lost for volume engine. Stopping local
>bricks.
>[2016-10-04 17:24:39.523314] I [MSGID: 106132]
>[glusterd-utils.c:1560:glusterd_service_stop] 0-management: brick
>already
>stopped
>[2016-10-04 17:24:39.526188] E [socket.c:2292:socket_connect_finish]
>0-management: connection to 10.100.103.3:24007 failed (No route to
>host)
>[2016-10-04 17:24:39.526219] I [MSGID: 106004]
>[glusterd-handler.c:5201:__glusterd_peer_rpc_notify] 0-management: Peer
> (<9a9c037e-96cd-4f73-9800-a1df5cdd2818>), in state Cluster>, has disconnected from glusterd.
>
>
>
>> Thanks
>>
>>
>>
>> *From:* Sahina Bose [mailto:sab...@redhat.com]
>> *Sent:* 05 October 2016 08:11
>> *To:* Jason Jeffrey ; gluster-us...@gluster.org;
>> Ravishankar Narayanankutty 
>> *Cc:* Simone Tiraboschi ; users
>
>>
>> *Subject:* Re: [ovirt-users] 4.0 - 2nd node fails on deploy
>>
>>
>>
>> [Adding gluster-users ML]
>>
>> The brick logs are filled with errors :
>> [2016-10-05 19:30:28.659061] E [MSGID: 113077]
>[posix-handle.c:309:posix_handle_pump]
>> 0-engine-posix: malformed internal link /var/run/vdsm/storage/
>>
>0a021563-91b5-4f49-9c6b-fff45e85a025/d84f0551-0f2b-457c-808c-6369c6708d43/
>> 1b5a5e34-818c-4914-8192-2f05733b5583 for /xpool/engine/brick/.
>> glusterfs/b9/8e/b98ed8d2-3bf9-4b11-92fd-ca5324e131a8
>> [2016-10-05 19:30:28.659069] E [MSGID: 113091]
>[posix.c:180:posix_lookup]
>> 0-engine-posix: Failed to create inode handle for path
>> 
>> The message "E [MSGID: 113018] [posix.c:198:posix_lookup]
>0-engine-posix:
>> lstat on null failed" repeated 3 times between [2016-10-05
>19:30:28.656529]
>> and [2016-10-05 19:30:28.659076]
>> [2016-10-05 19:30:28.659087] W [MSGID: 115005]
>> [server-resolve.c:126:resolve_gfid_cbk] 0-engine-server:
>> b98ed8d2-3bf9-4b11-92fd-ca5324e131a8: failed to resolve (Success)
>>
>> - Ravi, the above are from the data brick of the arbiter volume. Can
>you
>> take a look?
>>
>>
>>
>> Jason,
>>
>> Could you also provide the mount logs from the first host
>> (/var/log/glusterfs/rhev-data-center-mnt-glusterSD*engine.log) and
>> glusterd log (/var/log/glusterfs/etc-glusterfs-glusterd.vol.log)
>around
>> the same time frame.
>>
>>
>>
>>
>>
>> On Wed, Oct 5, 2016 at 3:28 AM, Jason Jeffrey 
>wrote:
>>
>> Hi,
>>
>>
>>
>> Servers are powered  off  when I’m not looking at the problem.
>>
>>
>>
>> There may have been instances where all three were not powered on,
>during
>> the same period.
>>
>>
>>
>> Glusterhd log attached, the xpool-engine-brick log is over 1 GB in
>size,
>> I’ve taken a sample of the last  couple days, looks to be highly
>repative.
>>
>>
>>
>> Cheers
>>
>>
>>
>> Jason
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> *From:* Simone Tiraboschi [mailto:stira...@redhat.com]
>> *Sent:* 04 October 2016 16:50
>>
>>
>> *To:* Jason Jeffrey 
>> *Cc:* users 
>> *Subject:* Re: [ovirt-users] 4.0 - 2nd node fails on deploy
>>
>>
>>
>>
>>
>>
>>
>> On Tue, Oct 4, 2016 at 5:22 PM, Jason Jeffrey 
>wrote:
>>
>> Hi,
>>
>>
>>
>> DCASTORXX is a hosts entry for dedicated  direct 10GB links (each
>private
>> /28) between the x3 servers  i.e 1=> 2&3, 2=> 1&3, etc) planned to be
>used
>> solely for storage.
>>
>>
>>
>> I,e
>>
>>
>>
>> 10.100.50.81dcasrv01
>>
>> 10.100.101.1dcastor01
>>
>> 10.100.50.82

Re: [ovirt-users] Change host names/IPs

2016-10-05 Thread Marcin Mirecki
Hello Davide,

No, there is no support for changing host ip.
How urgently do you need this?

Thanks,
Marcin



- Original Message -
> From: "Davide Ferrari" 
> To: "users" 
> Sent: Wednesday, October 5, 2016 5:22:57 PM
> Subject: [ovirt-users] Change host names/IPs
> 
> Hello
> 
> Is there a clean way and possibly without downtime to change the hostname and
> IP addresses of all the hosts in a running oVirt cluster?
> 
> --
> Davide Ferrari
> Senior Systems Engineer
> 
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
> 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Change host names/IPs

2016-10-05 Thread Davide Ferrari
Hello

Is there a clean way and possibly without downtime to change the hostname
and IP addresses of all the hosts in a running oVirt cluster?

-- 
Davide Ferrari
Senior Systems Engineer
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] change broker.conf configuration on shared storage

2016-10-05 Thread emanuel . santosvarina
hi all, 

hmm, broker.conf is now on the shared storage and not replicated on each 
host local file system. how to make changes to the conf, e.g.  the notify 
key? 

thanks, emanuel

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Add extra dnsname for certificate on engine

2016-10-05 Thread Matt .
Hi Didi,

Thanks for the heads up, yes this is very perfect! Looking forward to it!

Cheers overthere!


2016-10-05 12:13 GMT+02:00 Yedidyah Bar David :
> On Mon, Oct 3, 2016 at 5:47 PM, Matt .  wrote:
>> Hi guys,
>>
>>
>> When the engine lives under engine.sub.sub.sub.domain.tld is it
>> possible to add an extra alt hostname to it so we can add a
>> certificate for engine.domain.tld ?
>>
>> I hope someone can point this out.
>
> Can you please check if [1] is enough for your needs? Thanks.
>
> [1] https://bugzilla.redhat.com/show_bug.cgi?id=1325746
>
> Best,
> --
> Didi
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] ETL Service fails

2016-10-05 Thread Andy
Didi,
Thanks for the info,  it appears running the command your provided fixed the 
problem.  Upgrading the appliance could have corrupted the DB, but as I recall 
I didnt receive any errors or nothing failed.  Thanks for all the help guys. 

Andy 


On Wednesday, October 5, 2016 4:33 AM, Yedidyah Bar David  
wrote:
 

 On Mon, Oct 3, 2016 at 12:30 PM, Sandro Bonazzola  wrote:

Adding some relevant people

Are you sure it's not a system/hardware issue? A bad disk, filesystem, network?

If it was something temporary, you might manage to fix it using the hint in the 
error message:

su - postgres -c "psql ovirt_engine_history -c 'REINDEX INDEX 
idx_vm_configuration_version_samples'"

Adding also Eli.

Best,
 

On Sat, Oct 1, 2016 at 6:24 AM, Andy  wrote:

Hello,
I just upgraded to OVIRT 4.04 on the hosted appliance and for some reason I no 
longer see the DHW metrics on the dashboard. I receive a lot of errors, 
obviously, stating "ETL Service has encountered an error, Please consult the 
service log. The research that I have been able to come up with appears to be 
with a FK or a user that no longer exists in the database, however the attached 
logs have zero reference to these errors.

Any help would be appreciated.

Thanks Andy
__ _
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman /listinfo/users





-- 
Sandro Bonazzola
Better technology. Faster innovation. Powered by community collaboration.
See how it works at redhat.com





-- 
Didi


   ___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hosted-engine and GlusterFS on Vlan help

2016-10-05 Thread Sahina Bose
On Tue, Oct 4, 2016 at 9:51 PM, Hanson  wrote:

> Running iperf3 between node1 & node2, I can achieve almost 10gbps without
> ever going out to the gateway...
>
> So switching between port to port on the switch is working properly on the
> vlan.
>
> This must be a problem in the gluster settings? Where do I start
> troubleshooting here?


Is it the gluster traffic that's routed via gateway?

By moving VM, did you mean migrating VM from one host to another? In this
case, there's no storage migration. Why do you think this is a problem in
gluster settings?



>
>
>
> On 10/04/2016 10:38 AM, Hanson wrote:
>
>> Hi Guys,
>>
>> I've converted my lab from using 802.3ad with bonding>bridged vlans to
>> one link with two vlan bridges and am now having traffic jumping to the
>> gateway when moving VM's/ISO/etc.
>>
>> 802.3ad = node1>switch1>node2
>> 801.1q = node1>switch1>gateway>switch1>node2
>>
>> I assume I've setup the same vlan style, though this time I used the gui
>> on the initial host install... setting up the vlans with their parent being
>> eth0.
>>
>> Hosted-engine on deploy then creates ovirtmgmt on top of eth0.11 ...
>>
>> Switch is tagged for vlans 10 & 11. Including a PVID of 11 for good
>> measure. (Gluster is vlan 11)
>>
>> I'd expect the traffic from node to node to be going from port to port
>> like it did in 802.3ad, what have I done wrong or is it using the gui
>> initially?
>>
>> This is how the current setup looks:
>>
>> /var/lib/vdsm/Persistent/netconf/nets/ovirtmgmt:
>> {
>> "ipv6autoconf": false,
>> "nameservers": [],
>> "nic": "eth0",
>> "vlan": 11,
>> "ipaddr": "10.0.3.11",
>> "switch": "legacy",
>> "mtu": 1500,
>> "netmask": "255.255.255.0",
>> "dhcpv6": false,
>> "stp": false,
>> "bridged": true,
>> "gateway": "10.0.3.1",
>> "defaultRoute": true
>> }
>>
>> /etc/sysconfig/network-scripts/ifcfg-ovirtmgmt:
>> # Generated by VDSM version 4.18.13-1.el7.centos
>> DEVICE=ovirtmgmt
>> TYPE=Bridge
>> DELAY=0
>> STP=off
>> ONBOOT=yes
>> IPADDR=10.0.3.11
>> NETMASK=255.255.255.0
>> GATEWAY=10.0.3.1
>> BOOTPROTO=none
>> DEFROUTE=yes
>> NM_CONTROLLED=no
>> IPV6INIT=no
>> VLAN_ID=11
>> MTU=1500
>>
>> Thanks!!
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Correct recovery procedure of the oVirt Hosted Engine 4.0

2016-10-05 Thread Simone Tiraboschi
On Wed, Oct 5, 2016 at 12:40 PM,  wrote:

> Ouch. It is beyond my understanding.
>
> Thus, it appears that described in the RHV4 guide (
> https://access.redhat.com/documentation/en/red-hat-
> virtualization/4.0/single/self-hosted-engine-guide/#
> sect-Restoring_SHE_bkup) recovery procedure in fact incomplete?
>

Yes, you are right although this is a kind of special case since we are
moving/restoring to a different storage domain while you are not asked to
remove the old storage if you are restoring in place.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Correct recovery procedure of the oVirt Hosted Engine 4.0

2016-10-05 Thread aleksey . maksimov
Ouch. It is beyond my understanding.

Thus, it appears that described in the RHV4 guide 
(https://access.redhat.com/documentation/en/red-hat-virtualization/4.0/single/self-hosted-engine-guide/#sect-Restoring_SHE_bkup)
 recovery procedure in fact incomplete?
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] ovirt-ha-agent cpu usage

2016-10-05 Thread Simone Tiraboschi
On Wed, Oct 5, 2016 at 10:34 AM, Nir Soffer  wrote:

> On Wed, Oct 5, 2016 at 10:24 AM, Simone Tiraboschi 
> wrote:
>
>>
>>
>> On Wed, Oct 5, 2016 at 9:17 AM, gregor  wrote:
>>
>>> Hi,
>>>
>>> did you found a solution or cause for this high CPU usage?
>>> I have installed the self hosted engine on another server and there is
>>> no VM running but ovirt-ha-agent uses heavily the CPU.
>>>
>>
>> Yes, it's due to the fact that ovirt-ha-agent periodically reconnects
>> over json rpc and this is CPU intensive since the client has to parse the
>> yaml API specification each time it connects.
>>
>
> Simone, reusing the connection is good idea anyway, but what you describe
> is
> a bug in the client library. The library does *not* need to load and parse
> the
> schema at all for sending requests to vdsm.
>
> The schema is only needed if you want to verify request parameters,
> or provide online help, these are not needed in a client library.
>
> Please file an infra bug about it.
>

Done, https://bugzilla.redhat.com/show_bug.cgi?id=1381899
Thanks.


> Nir
>
>
>> The issue is tracked here:
>> https://bugzilla.redhat.com/show_bug.cgi?id=1349829 - ovirt-ha-agent
>> should reuse json-rpc connections
>> but it depends on:
>> https://bugzilla.redhat.com/show_bug.cgi?id=1376843 - [RFE] Implement a
>> keep-alive with reconnect if needed logic for the python jsonrpc client
>>
>>
>>
>>>
>>> cheers
>>> gregor
>>>
>>> On 08/08/16 15:09, Gianluca Cecchi wrote:
>>> > On Mon, Aug 8, 2016 at 1:03 PM, Roy Golan >> > > wrote:
>>> >
>>> > Does the spikes correlates with info messages on extracting the
>>> ovf?
>>> >
>>> >
>>> >
>>> >
>>> >
>>> >
>>> > yes, it seems so and it happens every 14-15 seconds
>>> >
>>> > These are the lines I see scrolling in agent.log when I notice cpu
>>> > spikes in ovirt-ha-agent...
>>> >
>>> > MainThread::INFO::2016-08-08
>>> > 15:03:07,815::storage_server::212::ovirt_hosted_engine_ha.li
>>> b.storage_server.StorageServer::(connect_storage_server)
>>> > Connecting storage server
>>> > MainThread::INFO::2016-08-08
>>> > 15:03:08,144::storage_server::220::ovirt_hosted_engine_ha.li
>>> b.storage_server.StorageServer::(connect_storage_server)
>>> > Refreshing the storage domain
>>> > MainThread::INFO::2016-08-08
>>> > 15:03:08,705::hosted_engine::685::ovirt_hosted_engine_ha.age
>>> nt.hosted_engine.HostedEngine::(_initialize_storage_images)
>>> > Preparing images
>>> > MainThread::INFO::2016-08-08
>>> > 15:03:08,705::image::126::ovirt_hosted_engine_ha.lib.image.I
>>> mage::(prepare_images)
>>> > Preparing images
>>> > MainThread::INFO::2016-08-08
>>> > 15:03:09,653::hosted_engine::688::ovirt_hosted_engine_ha.age
>>> nt.hosted_engine.HostedEngine::(_initialize_storage_images)
>>> > Reloading vm.conf from the shared storage domain
>>> > MainThread::INFO::2016-08-08
>>> > 15:03:09,653::config::205::ovirt_hosted_engine_ha.agent.host
>>> ed_engine.HostedEngine.config::(refresh_local_conf_file)
>>> > Trying to get a fresher copy of vm configuration from the OVF_STORE
>>> > MainThread::INFO::2016-08-08
>>> > 15:03:09,843::ovf_store::100::ovirt_hosted_engine_ha.lib.ovf
>>> .ovf_store.OVFStore::(scan)
>>> > Found OVF_STORE: imgUUID:223d26c2-1668-493c-a322-8054923d135f,
>>> > volUUID:108a362c-f5a9-440e-8817-1ed8a129afe8
>>> > MainThread::INFO::2016-08-08
>>> > 15:03:10,309::ovf_store::100::ovirt_hosted_engine_ha.lib.ovf
>>> .ovf_store.OVFStore::(scan)
>>> > Found OVF_STORE: imgUUID:12ca2fc6-01f7-41ab-ab22-e75c822ac9b6,
>>> > volUUID:1a18851e-6858-401c-be6e-af14415034b5
>>> > MainThread::INFO::2016-08-08
>>> > 15:03:10,652::ovf_store::109::ovirt_hosted_engine_ha.lib.ovf
>>> .ovf_store.OVFStore::(getEngineVMOVF)
>>> > Extracting Engine VM OVF from the OVF_STORE
>>> > MainThread::INFO::2016-08-08
>>> > 15:03:10,974::ovf_store::116::ovirt_hosted_engine_ha.lib.ovf
>>> .ovf_store.OVFStore::(getEngineVMOVF)
>>> > OVF_STORE volume path:
>>> > /rhev/data-center/mnt/ovirt01.lutwyn.org:_SHE__DOMAIN/31a9e9
>>> fd-8dcb-4475-aac4-09f897ee1b45/images/12ca2fc6-01f7-41ab-ab2
>>> 2-e75c822ac9b6/1a18851e-6858-401c-be6e-af14415034b5
>>> > MainThread::INFO::2016-08-08
>>> > 15:03:11,494::config::225::ovirt_hosted_engine_ha.agent.host
>>> ed_engine.HostedEngine.config::(refresh_local_conf_file)
>>> > Found an OVF for HE VM, trying to convert
>>> > MainThread::INFO::2016-08-08
>>> > 15:03:11,497::config::230::ovirt_hosted_engine_ha.agent.host
>>> ed_engine.HostedEngine.config::(refresh_local_conf_file)
>>> > Got vm.conf from OVF_STORE
>>> > MainThread::INFO::2016-08-08
>>> > 15:03:11,675::hosted_engine::462::ovirt_hosted_engine_ha.age
>>> nt.hosted_engine.HostedEngine::(start_monitoring)
>>> > Current state EngineUp (score: 3400)
>>> >
>>> >
>>> > ___
>>> > Users mailing list
>>> > Users@ovirt.org
>>> > http://lists.ovirt.org/mailman/listinfo/users
>>> >
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lis

Re: [ovirt-users] 4.0 - 2nd node fails on deploy

2016-10-05 Thread Sahina Bose
On Wed, Oct 5, 2016 at 1:56 PM, Jason Jeffrey  wrote:

> HI,
>
>
>
> Logs attached
>

Have you probed 2 interfaces for same host, that is - dcasrv02 and
dcastor02? Does "gluster peer status" understand both names as for same
host?

>From glusterd logs and the mount logs - the connection between the peers is
lost, and quorum is lost, which is reaffirming what Simone said earlier.
Logs seem to indicate network issues - check the direct link setup. See
below

>From mount logs:
[2016-10-04 17:26:15.718300] E [socket.c:2292:socket_connect_finish]
0-engine-client-2: connection to 10.100.103.3:24007 failed (No route to
host)
[2016-10-04 17:26:15.718345] W [MSGID: 108001]
[afr-common.c:4379:afr_notify] 0-engine-replicate-0: Client-quorum is not
met
[2016-10-04 17:26:16.428290] E [socket.c:2292:socket_connect_finish]
0-engine-client-1: connection to 10.100.101.2:24007 failed (No route to
host)
[2016-10-04 17:26:16.428336] E [MSGID: 108006]
[afr-common.c:4321:afr_notify] 0-engine-replicate-0: All subvolumes are
down. Going offline until atleast one of them comes back up

And in glusterd logs:
[2016-10-04 17:24:39.522402] E [socket.c:2292:socket_connect_finish]
0-management: connection to 10.100.50.82:24007 failed (No route to host)
[2016-10-04 17:24:39.522578] I [MSGID: 106004]
[glusterd-handler.c:5201:__glusterd_peer_rpc_notify] 0-management: Peer
 (<1e788fc9-dfe9-4753-92c7-76a95c8d0891>), in state , has disconnected from glusterd.
[2016-10-04 17:24:39.523272] C [MSGID: 106002]
[glusterd-server-quorum.c:351:glusterd_do_volume_quorum_action]
0-management: Server quorum lost for volume engine. Stopping local bricks.
[2016-10-04 17:24:39.523314] I [MSGID: 106132]
[glusterd-utils.c:1560:glusterd_service_stop] 0-management: brick already
stopped
[2016-10-04 17:24:39.526188] E [socket.c:2292:socket_connect_finish]
0-management: connection to 10.100.103.3:24007 failed (No route to host)
[2016-10-04 17:24:39.526219] I [MSGID: 106004]
[glusterd-handler.c:5201:__glusterd_peer_rpc_notify] 0-management: Peer
 (<9a9c037e-96cd-4f73-9800-a1df5cdd2818>), in state , has disconnected from glusterd.



> Thanks
>
>
>
> *From:* Sahina Bose [mailto:sab...@redhat.com]
> *Sent:* 05 October 2016 08:11
> *To:* Jason Jeffrey ; gluster-us...@gluster.org;
> Ravishankar Narayanankutty 
> *Cc:* Simone Tiraboschi ; users 
>
> *Subject:* Re: [ovirt-users] 4.0 - 2nd node fails on deploy
>
>
>
> [Adding gluster-users ML]
>
> The brick logs are filled with errors :
> [2016-10-05 19:30:28.659061] E [MSGID: 113077] 
> [posix-handle.c:309:posix_handle_pump]
> 0-engine-posix: malformed internal link /var/run/vdsm/storage/
> 0a021563-91b5-4f49-9c6b-fff45e85a025/d84f0551-0f2b-457c-808c-6369c6708d43/
> 1b5a5e34-818c-4914-8192-2f05733b5583 for /xpool/engine/brick/.
> glusterfs/b9/8e/b98ed8d2-3bf9-4b11-92fd-ca5324e131a8
> [2016-10-05 19:30:28.659069] E [MSGID: 113091] [posix.c:180:posix_lookup]
> 0-engine-posix: Failed to create inode handle for path
> 
> The message "E [MSGID: 113018] [posix.c:198:posix_lookup] 0-engine-posix:
> lstat on null failed" repeated 3 times between [2016-10-05 19:30:28.656529]
> and [2016-10-05 19:30:28.659076]
> [2016-10-05 19:30:28.659087] W [MSGID: 115005]
> [server-resolve.c:126:resolve_gfid_cbk] 0-engine-server:
> b98ed8d2-3bf9-4b11-92fd-ca5324e131a8: failed to resolve (Success)
>
> - Ravi, the above are from the data brick of the arbiter volume. Can you
> take a look?
>
>
>
> Jason,
>
> Could you also provide the mount logs from the first host
> (/var/log/glusterfs/rhev-data-center-mnt-glusterSD*engine.log) and
> glusterd log (/var/log/glusterfs/etc-glusterfs-glusterd.vol.log) around
> the same time frame.
>
>
>
>
>
> On Wed, Oct 5, 2016 at 3:28 AM, Jason Jeffrey  wrote:
>
> Hi,
>
>
>
> Servers are powered  off  when I’m not looking at the problem.
>
>
>
> There may have been instances where all three were not powered on, during
> the same period.
>
>
>
> Glusterhd log attached, the xpool-engine-brick log is over 1 GB in size,
> I’ve taken a sample of the last  couple days, looks to be highly repative.
>
>
>
> Cheers
>
>
>
> Jason
>
>
>
>
>
>
>
>
>
> *From:* Simone Tiraboschi [mailto:stira...@redhat.com]
> *Sent:* 04 October 2016 16:50
>
>
> *To:* Jason Jeffrey 
> *Cc:* users 
> *Subject:* Re: [ovirt-users] 4.0 - 2nd node fails on deploy
>
>
>
>
>
>
>
> On Tue, Oct 4, 2016 at 5:22 PM, Jason Jeffrey  wrote:
>
> Hi,
>
>
>
> DCASTORXX is a hosts entry for dedicated  direct 10GB links (each private
> /28) between the x3 servers  i.e 1=> 2&3, 2=> 1&3, etc) planned to be used
> solely for storage.
>
>
>
> I,e
>
>
>
> 10.100.50.81dcasrv01
>
> 10.100.101.1dcastor01
>
> 10.100.50.82dcasrv02
>
> 10.100.101.2dcastor02
>
> 10.100.50.83dcasrv03
>
> 10.100.103.3dcastor03
>
>
>
> These were setup with the gluster commands
>
>
>
> · gluster volume create iso replica 3 arbiter 1
> dcastor01:/xpool/iso/brick   dcastor02:/xpool/iso/brick
> dcastor03:/xpool/iso/brick
>
> · gluster volume create export rep

Re: [ovirt-users] Correct recovery procedure of the oVirt Hosted Engine 4.0

2016-10-05 Thread Simone Tiraboschi
On Wed, Oct 5, 2016 at 11:56 AM,  wrote:

> Weird. The RHV4 guides not contain the information that we need to clean
> the database from old storage domain before running the command
> engine-setup.
> What specific actions do we need?
>

You can check this as a reference but take care because it's still not
fully tested:
https://gerrit.ovirt.org/#/c/64966/3/packaging/setup/dbutils/hecleaner_sp.sql


> Eventually I want to get a full recovery plan at the moment for oVirt 4.0.
>
> 05.10.2016, 12:07, "Simone Tiraboschi" :
>
>
>
> On Wed, Oct 5, 2016 at 10:30 AM,  wrote:
>
> Well.
> Then, in the case of conditions:
>
> 1) the vm is not available anymore due to storage corruption
> 2) an empty shared storage is available
> 3) engine backup exists
> 4) all VMs still running on the hosts in the cluster
>
>
> The recovery plan will be like this (as I understand it):
>
>
> 1) On all the hosts (if they are still available):
>
> # service ovirt-ha-broker stop
> # service ovirt-ha-agent stop
> # chkconfig --del ovirt-ha-broker
> # chkconfig --del ovirt-ha-agent
>
>
> 2) On first host (if the original host is not available anymore, provision
> a new host from scratch and proceed on this new host):
>
>   2.1) # hosted-engine --deploy
>
>  ◾use same fqdn you had previously in the HE VM.
>  ◾point to the new shared storage
>  ◾provide the same admin password you used in previous setup
>  ◾install the OS on the vm
>
>
> I'd suggest to use the engine appliance also for this.
> You can just say No when it asks about automatically running engine-setup.
>
>
>  ◾confirm it has been installed
>
>  on Hosted Engine VM:
>
>   a) Install the ovirt-engine rpms on the vm but don't run engine-setup:
>   # yum install http://resources.ovirt.org/pub/yum-repo/ovirt-release40.
> rpm
>   # yum install epel-release
>   # yum install ovirt-engine
>   b) Restore the backup:
>   # engine-backup --mode=restore --file=file_name --log=log_file_name
> --provision-db --provision-dwh-db --restore-permissions
>
>
> In order to let the engine auto-import the new hosted-engine storage
> domain, you have to remove the old one.
> The same for the engine VM. Unfortunately you cannot do that from the
> engine since they are somehow protected to avoid unintentional damages.
> The easiest way is to remove them from the DB before running engine-setup.
> I'm working on a helper utility to make it easiser:
> https://gerrit.ovirt.org/#/c/64966/
> I think I'll integrate it with engine-backup to simply do it with an
> additional CLI flag.
>
>
>   c) Run "engine-setup"
>
>2.2) Open Administration Portal and remove the all old hosts used for
> Hosted Engine
>
>
> Right, we can also integrate this step in the HE cleaning helper.
>
>
>
>2.3) Confirm that the engine has been installed (Return to the host and
> continue the hosted-engine deployment script by selecting option 1) and
> then finish the deploy.
>
>2.4) In Administration Portal activate new host
>
>
> 3) On all additional hosts run "hosted-engine --deploy".
>
>
> I strongly suggest to deploy them from the engine and not from CLI.
> CLI deploy support for additional HE host is deprecated an it will be
> removed in 4.1.
>
>
> Right?
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Add extra dnsname for certificate on engine

2016-10-05 Thread Yedidyah Bar David
On Mon, Oct 3, 2016 at 5:47 PM, Matt .  wrote:
> Hi guys,
>
>
> When the engine lives under engine.sub.sub.sub.domain.tld is it
> possible to add an extra alt hostname to it so we can add a
> certificate for engine.domain.tld ?
>
> I hope someone can point this out.

Can you please check if [1] is enough for your needs? Thanks.

[1] https://bugzilla.redhat.com/show_bug.cgi?id=1325746

Best,
-- 
Didi
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] libvirt-v2v error

2016-10-05 Thread Yaniv Kaul
If you are not using oVirt for this task (which you should, as it hides
all/most/some of those details from you), then I suggest asking in the
libguestfs mailing list - libgues...@redhat.com
Y.

On Wed, Oct 5, 2016 at 9:34 AM, Saman Bandara 
wrote:

> Dear sir,
>
> I'm getting following error while trying to convert a VMWare RHEL6 server
> to kvm.
> Please give any suggestion to resolve this.
>
> [root@kvm16 ~]# virt-v2v ic esx://10.16.32.12/?no_verify=1 -o rhev -os
> 10.16.32.16:/vm-images/export_domain --network rhevm "10.16.32.36-db-
> slcloudcontrol"
> virt-v2v: Failed to connect to qemu:///system: libvirt error code: 45,
> message: authentication failed: Failed to step SASL negotiation: -7
> (SASL(-7): invalid parameter supplied: Unexpectedly missing a prompt result)
>
> --
>
> *Saman K. Bandara,** Database Administrator*
>
> *ShipXpres Technologies (Pvt) Ltd.*
>
> 2300 Marsh Point Road, Suite 101 || Neptune Beach, FL 32266
>
> Phone: +94 71 8135485 <%2B94%20718135485> | +94  <%2B94%20777522730>76
> 6014001
>
> *Company Website * || *LinkedIn
>  *|| *Facebook
> * || *Twitter
> *
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Correct recovery procedure of the oVirt Hosted Engine 4.0

2016-10-05 Thread aleksey . maksimov
Weird. The RHV4 guides not contain the information that we need to clean the database from old storage domain before running the command engine-setup.What specific actions do we need?Eventually I want to get a full recovery plan at the moment for oVirt 4.0. 05.10.2016, 12:07, "Simone Tiraboschi" :  On Wed, Oct 5, 2016 at 10:30 AM,  wrote:Well.Then, in the case of conditions:1) the vm is not available anymore due to storage corruption2) an empty shared storage is available3) engine backup exists4) all VMs still running on the hosts in the clusterThe recovery plan will be like this (as I understand it):1) On all the hosts (if they are still available):# service ovirt-ha-broker stop# service ovirt-ha-agent stop# chkconfig --del ovirt-ha-broker# chkconfig --del ovirt-ha-agent2) On first host (if the original host is not available anymore, provision a new host from scratch and proceed on this new host):  2.1) # hosted-engine --deploy ◾use same fqdn you had previously in the HE VM. ◾point to the new shared storage ◾provide the same admin password you used in previous setup ◾install the OS on the vm I'd suggest to use the engine appliance also for this.You can just say No when it asks about automatically running engine-setup.   ◾confirm it has been installed on Hosted Engine VM:  a) Install the ovirt-engine rpms on the vm but don't run engine-setup:  # yum install http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm  # yum install epel-release  # yum install ovirt-engine  b) Restore the backup:  # engine-backup --mode=restore --file=file_name --log=log_file_name --provision-db --provision-dwh-db --restore-permissions In order to let the engine auto-import the new hosted-engine storage domain, you have to remove the old one.The same for the engine VM. Unfortunately you cannot do that from the engine since they are somehow protected to avoid unintentional damages.The easiest way is to remove them from the DB before running engine-setup.I'm working on a helper utility to make it easiser:https://gerrit.ovirt.org/#/c/64966/I think I'll integrate it with engine-backup to simply do it with an additional CLI flag.    c) Run "engine-setup"   2.2) Open Administration Portal and remove the all old hosts used for Hosted Engine Right, we can also integrate this step in the HE cleaning helper. 2.3) Confirm that the engine has been installed (Return to the host and continue the hosted-engine deployment script by selecting option 1) and then finish the deploy.   2.4) In Administration Portal activate new host3) On all additional hosts run "hosted-engine --deploy". I strongly suggest to deploy them from the engine and not from CLI.CLI deploy support for additional HE host is deprecated an it will be removed in 4.1. Right?___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Unable to find OVF_STORE after recovery / upgrade

2016-10-05 Thread Simone Tiraboschi
On Mon, Oct 3, 2016 at 6:47 PM, Sam Cappello  wrote:

> Hi,
> so i was running a 3.4 hosted engine two node setup on centos 6, had some
> disk issues so i tried to upgrade to centos 7 and follow the path 3.4 > 3.5
> > 3.6 > 4.0.  i screwed up dig time somewhere between 3.6 and 4.0, so i
> wiped the drives, installed a fresh 4.0.3, then created the database and
> restored the 3.6 engine backup before running engine-setup as per the
> docs.   things seemed to work, but i have the the following issues /
> symptoms:
> - ovirt-ha-agent running 100% CPU on both nodes
> - messages in the UI that the Hosted Engine storage Domain isn't active
> and Failed to import the Hosted Engine Storage Domain
> - hosted engine is not visible in the UI
> and the following repeating in the agent.log:
>
> MainThread::INFO::2016-10-03 12:38:27,718::hosted_engine::
> 461::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
> Current state EngineUp (score: 3400)
> MainThread::INFO::2016-10-03 12:38:27,720::hosted_engine::
> 466::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
> Best remote host vmhost1.oracool.net (id: 1, score: 3400)
> MainThread::INFO::2016-10-03 12:38:37,979::states::421::
> ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(consume) Engine
> vm running on localhost
> MainThread::INFO::2016-10-03 12:38:37,985::hosted_engine::
> 612::ovirt_hosted_engine_ha.agent.hosted_engine.
> HostedEngine::(_initialize_vdsm) Initializing VDSM
> MainThread::INFO::2016-10-03 12:38:45,645::hosted_engine::
> 639::ovirt_hosted_engine_ha.agent.hosted_engine.
> HostedEngine::(_initialize_storage_images) Connecting the storage
> MainThread::INFO::2016-10-03 12:38:45,647::storage_server::
> 219::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server)
> Connecting storage server
> MainThread::INFO::2016-10-03 12:39:00,543::storage_server::
> 226::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server)
> Connecting storage server
> MainThread::INFO::2016-10-03 12:39:00,562::storage_server::
> 233::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server)
> Refreshing the storage domain
> MainThread::INFO::2016-10-03 12:39:01,235::hosted_engine::
> 666::ovirt_hosted_engine_ha.agent.hosted_engine.
> HostedEngine::(_initialize_storage_images) Preparing images
> MainThread::INFO::2016-10-03 12:39:01,236::image::126::
> ovirt_hosted_engine_ha.lib.image.Image::(prepare_images) Preparing images
> MainThread::INFO::2016-10-03 12:39:09,295::hosted_engine::
> 669::ovirt_hosted_engine_ha.agent.hosted_engine.
> HostedEngine::(_initialize_storage_images) Reloading vm.conf from the
> shared storage domain
> MainThread::INFO::2016-10-03 12:39:09,296::config::206::
> ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.
> config::(refresh_local_conf_file) Trying to get a fresher copy of vm
> configuration from the OVF_STORE
> MainThread::WARNING::2016-10-03 12:39:16,928::ovf_store::107::
> ovirt_hosted_engine_ha.lib.ovf.ovf_store.OVFStore::(scan) Unable to find
> OVF_STORE
>

The engine will automatically create it once the hosted-engine storage
domain and the engine VM are correctly been imported.


> MainThread::ERROR::2016-10-03 12:39:16,934::config::235::
> ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.
> config::(refresh_local_conf_file) Unable to get vm.conf from OVF_STORE,
> falling back to initial vm.conf
>
> I have searched a bit and not really found a solution, and have come to
> the conclusion that i have made a mess of things, and am wondering if the
> best solution is to export the VMs, and reinstall everything then import
> them back?
> i am using remote  NFS storage.
> if i try and add the hosted engine storage domain it says it is already
> registered.
>

The best option here is to manually remove it from the DB and let the
engine import it again.
I'm working on an helper utility here but it's still not fully tested:
https://gerrit.ovirt.org/#/c/64966/


> i have also upgraded and am now running oVirt Engine Version:
> 4.0.4.4-1.el7.centos
> hosts were installed using ovirt-node.  currently at
> 3.10.0-327.28.3.el7.x86_64
> if a fresh install is best, any advice / pointer to doc that explains best
> way to do this?
> i have not moved my most important server over to this cluster yet so i
> can take some downtime to reinstall.
> thanks!
> sam
>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] libvirt-v2v error

2016-10-05 Thread Saman Bandara
Dear sir,

I'm getting following error while trying to convert a VMWare RHEL6 server
to kvm.
Please give any suggestion to resolve this.

[root@kvm16 ~]# virt-v2v ic esx://10.16.32.12/?no_verify=1 -o rhev -os
10.16.32.16:/vm-images/export_domain --network rhevm
"10.16.32.36-db-slcloudcontrol"
virt-v2v: Failed to connect to qemu:///system: libvirt error code: 45,
message: authentication failed: Failed to step SASL negotiation: -7
(SASL(-7): invalid parameter supplied: Unexpectedly missing a prompt result)

-- 

*Saman K. Bandara,** Database Administrator*

*ShipXpres Technologies (Pvt) Ltd.*

2300 Marsh Point Road, Suite 101 || Neptune Beach, FL 32266

Phone: +94 71 8135485 <%2B94%20718135485> | +94  <%2B94%20777522730>76
6014001

*Company Website * || *LinkedIn
 *|| *Facebook
* || *Twitter
*
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Correct recovery procedure of the oVirt Hosted Engine 4.0

2016-10-05 Thread Simone Tiraboschi
On Wed, Oct 5, 2016 at 10:30 AM,  wrote:

> Well.
> Then, in the case of conditions:
>
> 1) the vm is not available anymore due to storage corruption
> 2) an empty shared storage is available
> 3) engine backup exists
> 4) all VMs still running on the hosts in the cluster
>
>
> The recovery plan will be like this (as I understand it):
>
>
> 1) On all the hosts (if they are still available):
>
> # service ovirt-ha-broker stop
> # service ovirt-ha-agent stop
> # chkconfig --del ovirt-ha-broker
> # chkconfig --del ovirt-ha-agent
>
>
> 2) On first host (if the original host is not available anymore, provision
> a new host from scratch and proceed on this new host):
>
>   2.1) # hosted-engine --deploy
>
>  ◾use same fqdn you had previously in the HE VM.
>  ◾point to the new shared storage
>  ◾provide the same admin password you used in previous setup
>  ◾install the OS on the vm
>

I'd suggest to use the engine appliance also for this.
You can just say No when it asks about automatically running engine-setup.


>  ◾confirm it has been installed
>
>  on Hosted Engine VM:
>
>   a) Install the ovirt-engine rpms on the vm but don't run engine-setup:
>   # yum install http://resources.ovirt.org/pub/yum-repo/ovirt-release40.
> rpm
>   # yum install epel-release
>   # yum install ovirt-engine
>   b) Restore the backup:
>   # engine-backup --mode=restore --file=file_name --log=log_file_name
> --provision-db --provision-dwh-db --restore-permissions
>

In order to let the engine auto-import the new hosted-engine storage
domain, you have to remove the old one.
The same for the engine VM. Unfortunately you cannot do that from the
engine since they are somehow protected to avoid unintentional damages.
The easiest way is to remove them from the DB before running engine-setup.
I'm working on a helper utility to make it easiser:
https://gerrit.ovirt.org/#/c/64966/
I think I'll integrate it with engine-backup to simply do it with an
additional CLI flag.


>   c) Run "engine-setup"
>
>2.2) Open Administration Portal and remove the all old hosts used for
> Hosted Engine
>

Right, we can also integrate this step in the HE cleaning helper.


>
>2.3) Confirm that the engine has been installed (Return to the host and
> continue the hosted-engine deployment script by selecting option 1) and
> then finish the deploy.
>
>2.4) In Administration Portal activate new host
>
>
> 3) On all additional hosts run "hosted-engine --deploy".
>

I strongly suggest to deploy them from the engine and not from CLI.
CLI deploy support for additional HE host is deprecated an it will be
removed in 4.1.


>
>
> Right?
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] ovirt-ha-agent cpu usage

2016-10-05 Thread Nir Soffer
On Wed, Oct 5, 2016 at 10:24 AM, Simone Tiraboschi 
wrote:

>
>
> On Wed, Oct 5, 2016 at 9:17 AM, gregor  wrote:
>
>> Hi,
>>
>> did you found a solution or cause for this high CPU usage?
>> I have installed the self hosted engine on another server and there is
>> no VM running but ovirt-ha-agent uses heavily the CPU.
>>
>
> Yes, it's due to the fact that ovirt-ha-agent periodically reconnects over
> json rpc and this is CPU intensive since the client has to parse the yaml
> API specification each time it connects.
>

Simone, reusing the connection is good idea anyway, but what you describe
is
a bug in the client library. The library does *not* need to load and parse
the
schema at all for sending requests to vdsm.

The schema is only needed if you want to verify request parameters,
or provide online help, these are not needed in a client library.

Please file an infra bug about it.

Nir


> The issue is tracked here:
> https://bugzilla.redhat.com/show_bug.cgi?id=1349829 - ovirt-ha-agent
> should reuse json-rpc connections
> but it depends on:
> https://bugzilla.redhat.com/show_bug.cgi?id=1376843 - [RFE] Implement a
> keep-alive with reconnect if needed logic for the python jsonrpc client
>
>
>
>>
>> cheers
>> gregor
>>
>> On 08/08/16 15:09, Gianluca Cecchi wrote:
>> > On Mon, Aug 8, 2016 at 1:03 PM, Roy Golan > > > wrote:
>> >
>> > Does the spikes correlates with info messages on extracting the ovf?
>> >
>> >
>> >
>> >
>> >
>> >
>> > yes, it seems so and it happens every 14-15 seconds
>> >
>> > These are the lines I see scrolling in agent.log when I notice cpu
>> > spikes in ovirt-ha-agent...
>> >
>> > MainThread::INFO::2016-08-08
>> > 15:03:07,815::storage_server::212::ovirt_hosted_engine_ha.li
>> b.storage_server.StorageServer::(connect_storage_server)
>> > Connecting storage server
>> > MainThread::INFO::2016-08-08
>> > 15:03:08,144::storage_server::220::ovirt_hosted_engine_ha.li
>> b.storage_server.StorageServer::(connect_storage_server)
>> > Refreshing the storage domain
>> > MainThread::INFO::2016-08-08
>> > 15:03:08,705::hosted_engine::685::ovirt_hosted_engine_ha.age
>> nt.hosted_engine.HostedEngine::(_initialize_storage_images)
>> > Preparing images
>> > MainThread::INFO::2016-08-08
>> > 15:03:08,705::image::126::ovirt_hosted_engine_ha.lib.image.
>> Image::(prepare_images)
>> > Preparing images
>> > MainThread::INFO::2016-08-08
>> > 15:03:09,653::hosted_engine::688::ovirt_hosted_engine_ha.age
>> nt.hosted_engine.HostedEngine::(_initialize_storage_images)
>> > Reloading vm.conf from the shared storage domain
>> > MainThread::INFO::2016-08-08
>> > 15:03:09,653::config::205::ovirt_hosted_engine_ha.agent.host
>> ed_engine.HostedEngine.config::(refresh_local_conf_file)
>> > Trying to get a fresher copy of vm configuration from the OVF_STORE
>> > MainThread::INFO::2016-08-08
>> > 15:03:09,843::ovf_store::100::ovirt_hosted_engine_ha.lib.ovf
>> .ovf_store.OVFStore::(scan)
>> > Found OVF_STORE: imgUUID:223d26c2-1668-493c-a322-8054923d135f,
>> > volUUID:108a362c-f5a9-440e-8817-1ed8a129afe8
>> > MainThread::INFO::2016-08-08
>> > 15:03:10,309::ovf_store::100::ovirt_hosted_engine_ha.lib.ovf
>> .ovf_store.OVFStore::(scan)
>> > Found OVF_STORE: imgUUID:12ca2fc6-01f7-41ab-ab22-e75c822ac9b6,
>> > volUUID:1a18851e-6858-401c-be6e-af14415034b5
>> > MainThread::INFO::2016-08-08
>> > 15:03:10,652::ovf_store::109::ovirt_hosted_engine_ha.lib.ovf
>> .ovf_store.OVFStore::(getEngineVMOVF)
>> > Extracting Engine VM OVF from the OVF_STORE
>> > MainThread::INFO::2016-08-08
>> > 15:03:10,974::ovf_store::116::ovirt_hosted_engine_ha.lib.ovf
>> .ovf_store.OVFStore::(getEngineVMOVF)
>> > OVF_STORE volume path:
>> > /rhev/data-center/mnt/ovirt01.lutwyn.org:_SHE__DOMAIN/31a9e9
>> fd-8dcb-4475-aac4-09f897ee1b45/images/12ca2fc6-01f7-41ab-
>> ab22-e75c822ac9b6/1a18851e-6858-401c-be6e-af14415034b5
>> > MainThread::INFO::2016-08-08
>> > 15:03:11,494::config::225::ovirt_hosted_engine_ha.agent.host
>> ed_engine.HostedEngine.config::(refresh_local_conf_file)
>> > Found an OVF for HE VM, trying to convert
>> > MainThread::INFO::2016-08-08
>> > 15:03:11,497::config::230::ovirt_hosted_engine_ha.agent.host
>> ed_engine.HostedEngine.config::(refresh_local_conf_file)
>> > Got vm.conf from OVF_STORE
>> > MainThread::INFO::2016-08-08
>> > 15:03:11,675::hosted_engine::462::ovirt_hosted_engine_ha.age
>> nt.hosted_engine.HostedEngine::(start_monitoring)
>> > Current state EngineUp (score: 3400)
>> >
>> >
>> > ___
>> > Users mailing list
>> > Users@ovirt.org
>> > http://lists.ovirt.org/mailman/listinfo/users
>> >
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org

Re: [ovirt-users] ETL Service fails

2016-10-05 Thread Yedidyah Bar David
On Mon, Oct 3, 2016 at 12:30 PM, Sandro Bonazzola 
wrote:

> Adding some relevant people
>

Are you sure it's not a system/hardware issue? A bad disk, filesystem,
network?

If it was something temporary, you might manage to fix it using the hint in
the error message:

su - postgres -c "psql ovirt_engine_history -c 'REINDEX INDEX
idx_vm_configuration_version_samples'"

Adding also Eli.

Best,


>
> On Sat, Oct 1, 2016 at 6:24 AM, Andy  wrote:
>
>> Hello,
>> I just upgraded to OVIRT 4.04 on the hosted appliance and for some reason
>> I no longer see the DHW metrics on the dashboard. I receive a lot of
>> errors, obviously, stating "ETL Service has encountered an error, Please
>> consult the service log. The research that I have been able to come up with
>> appears to be with a FK or a user that no longer exists in the database,
>> however the attached logs have zero reference to these errors.
>>
>> Any help would be appreciated.
>>
>> Thanks Andy
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>
>
> --
> Sandro Bonazzola
> Better technology. Faster innovation. Powered by community collaboration.
> See how it works at redhat.com
> 
>



-- 
Didi
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Correct recovery procedure of the oVirt Hosted Engine 4.0

2016-10-05 Thread aleksey . maksimov
Well.
Then, in the case of conditions:

1) the vm is not available anymore due to storage corruption
2) an empty shared storage is available
3) engine backup exists
4) all VMs still running on the hosts in the cluster


The recovery plan will be like this (as I understand it):


1) On all the hosts (if they are still available):

# service ovirt-ha-broker stop
# service ovirt-ha-agent stop
# chkconfig --del ovirt-ha-broker
# chkconfig --del ovirt-ha-agent


2) On first host (if the original host is not available anymore, provision a 
new host from scratch and proceed on this new host):

  2.1) # hosted-engine --deploy

 ◾use same fqdn you had previously in the HE VM.
 ◾point to the new shared storage
 ◾provide the same admin password you used in previous setup
 ◾install the OS on the vm
 ◾confirm it has been installed

 on Hosted Engine VM:

  a) Install the ovirt-engine rpms on the vm but don't run engine-setup:
  # yum install http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm
  # yum install epel-release
  # yum install ovirt-engine
  b) Restore the backup:
  # engine-backup --mode=restore --file=file_name --log=log_file_name 
--provision-db --provision-dwh-db --restore-permissions
  c) Run "engine-setup"

   2.2) Open Administration Portal and remove the all old hosts used for Hosted 
Engine

   2.3) Confirm that the engine has been installed (Return to the host and 
continue the hosted-engine deployment script by selecting option 1) and then 
finish the deploy.

   2.4) In Administration Portal activate new host


3) On all additional hosts run "hosted-engine --deploy".


Right?
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Correct recovery procedure of the oVirt Hosted Engine 4.0

2016-10-05 Thread Simone Tiraboschi
On Wed, Oct 5, 2016 at 9:19 AM,  wrote:

> Hello oVirt guru`s!
>
>
> My Hosted Engine VM located on a dedicated LUN FC Storage.
>
> I do daily data backups (on NFS share) with the command:
>
> /usr/bin/engine-backup --mode=backup --scope=all --file=$BcpFileName.xz
> --log=$BcpFileName.log --archive-compressor=xz --files-compressor=None
>
> However, I don't know what would be the correct procedure to recover,
> because in different manuals outline the various steps.
>
> For example, there is information that I have to do configure postgresql
> (with password from file 
> files/etc/ovirt-engine/engine.conf.d/10-setup-database.conf)
> before restoring (engine-backup --mode=restore):
> https://www.ovirt.org/documentation/admin-guide/hosted-engine-backup-and-
> restore/


The recent releases of engine-backup can do that for you.
Adding Didi here.


>
>
> And at the same time, in another document, there are no such steps:
> https://access.redhat.com/documentation/en/red-hat-
> virtualization/4.0/single/self-hosted-engine-guide/#
> sect-Restoring_SHE_bkup
>
> What should be the correct procedure for the recovery of Hosted Engine 4.0
> ?
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] ovirt-ha-agent cpu usage

2016-10-05 Thread Simone Tiraboschi
On Wed, Oct 5, 2016 at 9:17 AM, gregor  wrote:

> Hi,
>
> did you found a solution or cause for this high CPU usage?
> I have installed the self hosted engine on another server and there is
> no VM running but ovirt-ha-agent uses heavily the CPU.
>

Yes, it's due to the fact that ovirt-ha-agent periodically reconnects over
json rpc and this is CPU intensive since the client has to parse the yaml
API specification each time it connects.
The issue is tracked here:
https://bugzilla.redhat.com/show_bug.cgi?id=1349829 - ovirt-ha-agent should
reuse json-rpc connections
but it depends on:
https://bugzilla.redhat.com/show_bug.cgi?id=1376843 - [RFE] Implement a
keep-alive with reconnect if needed logic for the python jsonrpc client


>
> cheers
> gregor
>
> On 08/08/16 15:09, Gianluca Cecchi wrote:
> > On Mon, Aug 8, 2016 at 1:03 PM, Roy Golan  > > wrote:
> >
> > Does the spikes correlates with info messages on extracting the ovf?
> >
> >
> >
> >
> >
> >
> > yes, it seems so and it happens every 14-15 seconds
> >
> > These are the lines I see scrolling in agent.log when I notice cpu
> > spikes in ovirt-ha-agent...
> >
> > MainThread::INFO::2016-08-08
> > 15:03:07,815::storage_server::212::ovirt_hosted_engine_ha.
> lib.storage_server.StorageServer::(connect_storage_server)
> > Connecting storage server
> > MainThread::INFO::2016-08-08
> > 15:03:08,144::storage_server::220::ovirt_hosted_engine_ha.
> lib.storage_server.StorageServer::(connect_storage_server)
> > Refreshing the storage domain
> > MainThread::INFO::2016-08-08
> > 15:03:08,705::hosted_engine::685::ovirt_hosted_engine_ha.
> agent.hosted_engine.HostedEngine::(_initialize_storage_images)
> > Preparing images
> > MainThread::INFO::2016-08-08
> > 15:03:08,705::image::126::ovirt_hosted_engine_ha.lib.
> image.Image::(prepare_images)
> > Preparing images
> > MainThread::INFO::2016-08-08
> > 15:03:09,653::hosted_engine::688::ovirt_hosted_engine_ha.
> agent.hosted_engine.HostedEngine::(_initialize_storage_images)
> > Reloading vm.conf from the shared storage domain
> > MainThread::INFO::2016-08-08
> > 15:03:09,653::config::205::ovirt_hosted_engine_ha.agent.
> hosted_engine.HostedEngine.config::(refresh_local_conf_file)
> > Trying to get a fresher copy of vm configuration from the OVF_STORE
> > MainThread::INFO::2016-08-08
> > 15:03:09,843::ovf_store::100::ovirt_hosted_engine_ha.lib.
> ovf.ovf_store.OVFStore::(scan)
> > Found OVF_STORE: imgUUID:223d26c2-1668-493c-a322-8054923d135f,
> > volUUID:108a362c-f5a9-440e-8817-1ed8a129afe8
> > MainThread::INFO::2016-08-08
> > 15:03:10,309::ovf_store::100::ovirt_hosted_engine_ha.lib.
> ovf.ovf_store.OVFStore::(scan)
> > Found OVF_STORE: imgUUID:12ca2fc6-01f7-41ab-ab22-e75c822ac9b6,
> > volUUID:1a18851e-6858-401c-be6e-af14415034b5
> > MainThread::INFO::2016-08-08
> > 15:03:10,652::ovf_store::109::ovirt_hosted_engine_ha.lib.
> ovf.ovf_store.OVFStore::(getEngineVMOVF)
> > Extracting Engine VM OVF from the OVF_STORE
> > MainThread::INFO::2016-08-08
> > 15:03:10,974::ovf_store::116::ovirt_hosted_engine_ha.lib.
> ovf.ovf_store.OVFStore::(getEngineVMOVF)
> > OVF_STORE volume path:
> > /rhev/data-center/mnt/ovirt01.lutwyn.org:_SHE__DOMAIN/
> 31a9e9fd-8dcb-4475-aac4-09f897ee1b45/images/12ca2fc6-
> 01f7-41ab-ab22-e75c822ac9b6/1a18851e-6858-401c-be6e-af14415034b5
> > MainThread::INFO::2016-08-08
> > 15:03:11,494::config::225::ovirt_hosted_engine_ha.agent.
> hosted_engine.HostedEngine.config::(refresh_local_conf_file)
> > Found an OVF for HE VM, trying to convert
> > MainThread::INFO::2016-08-08
> > 15:03:11,497::config::230::ovirt_hosted_engine_ha.agent.
> hosted_engine.HostedEngine.config::(refresh_local_conf_file)
> > Got vm.conf from OVF_STORE
> > MainThread::INFO::2016-08-08
> > 15:03:11,675::hosted_engine::462::ovirt_hosted_engine_ha.
> agent.hosted_engine.HostedEngine::(start_monitoring)
> > Current state EngineUp (score: 3400)
> >
> >
> > ___
> > Users mailing list
> > Users@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> >
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Correct recovery procedure of the oVirt Hosted Engine 4.0

2016-10-05 Thread aleksey . maksimov
Hello oVirt guru`s!


My Hosted Engine VM located on a dedicated LUN FC Storage.

I do daily data backups (on NFS share) with the command:

/usr/bin/engine-backup --mode=backup --scope=all --file=$BcpFileName.xz 
--log=$BcpFileName.log --archive-compressor=xz --files-compressor=None

However, I don't know what would be the correct procedure to recover, because 
in different manuals outline the various steps.

For example, there is information that I have to do configure postgresql (with 
password from file files/etc/ovirt-engine/engine.conf.d/10-setup-database.conf) 
before restoring (engine-backup --mode=restore):
https://www.ovirt.org/documentation/admin-guide/hosted-engine-backup-and-restore/

And at the same time, in another document, there are no such steps:
https://access.redhat.com/documentation/en/red-hat-virtualization/4.0/single/self-hosted-engine-guide/#sect-Restoring_SHE_bkup

What should be the correct procedure for the recovery of Hosted Engine 4.0 ?
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] ovirt-ha-agent cpu usage

2016-10-05 Thread gregor
Hi,

did you found a solution or cause for this high CPU usage?
I have installed the self hosted engine on another server and there is
no VM running but ovirt-ha-agent uses heavily the CPU.

cheers
gregor

On 08/08/16 15:09, Gianluca Cecchi wrote:
> On Mon, Aug 8, 2016 at 1:03 PM, Roy Golan  > wrote:
> 
> Does the spikes correlates with info messages on extracting the ovf?
> 
> 
> 
> 
> 
> 
> yes, it seems so and it happens every 14-15 seconds
> 
> These are the lines I see scrolling in agent.log when I notice cpu
> spikes in ovirt-ha-agent...
> 
> MainThread::INFO::2016-08-08
> 15:03:07,815::storage_server::212::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server)
> Connecting storage server
> MainThread::INFO::2016-08-08
> 15:03:08,144::storage_server::220::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server)
> Refreshing the storage domain
> MainThread::INFO::2016-08-08
> 15:03:08,705::hosted_engine::685::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_storage_images)
> Preparing images
> MainThread::INFO::2016-08-08
> 15:03:08,705::image::126::ovirt_hosted_engine_ha.lib.image.Image::(prepare_images)
> Preparing images
> MainThread::INFO::2016-08-08
> 15:03:09,653::hosted_engine::688::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_storage_images)
> Reloading vm.conf from the shared storage domain
> MainThread::INFO::2016-08-08
> 15:03:09,653::config::205::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.config::(refresh_local_conf_file)
> Trying to get a fresher copy of vm configuration from the OVF_STORE
> MainThread::INFO::2016-08-08
> 15:03:09,843::ovf_store::100::ovirt_hosted_engine_ha.lib.ovf.ovf_store.OVFStore::(scan)
> Found OVF_STORE: imgUUID:223d26c2-1668-493c-a322-8054923d135f,
> volUUID:108a362c-f5a9-440e-8817-1ed8a129afe8
> MainThread::INFO::2016-08-08
> 15:03:10,309::ovf_store::100::ovirt_hosted_engine_ha.lib.ovf.ovf_store.OVFStore::(scan)
> Found OVF_STORE: imgUUID:12ca2fc6-01f7-41ab-ab22-e75c822ac9b6,
> volUUID:1a18851e-6858-401c-be6e-af14415034b5
> MainThread::INFO::2016-08-08
> 15:03:10,652::ovf_store::109::ovirt_hosted_engine_ha.lib.ovf.ovf_store.OVFStore::(getEngineVMOVF)
> Extracting Engine VM OVF from the OVF_STORE
> MainThread::INFO::2016-08-08
> 15:03:10,974::ovf_store::116::ovirt_hosted_engine_ha.lib.ovf.ovf_store.OVFStore::(getEngineVMOVF)
> OVF_STORE volume path:
> /rhev/data-center/mnt/ovirt01.lutwyn.org:_SHE__DOMAIN/31a9e9fd-8dcb-4475-aac4-09f897ee1b45/images/12ca2fc6-01f7-41ab-ab22-e75c822ac9b6/1a18851e-6858-401c-be6e-af14415034b5
>  
> MainThread::INFO::2016-08-08
> 15:03:11,494::config::225::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.config::(refresh_local_conf_file)
> Found an OVF for HE VM, trying to convert
> MainThread::INFO::2016-08-08
> 15:03:11,497::config::230::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.config::(refresh_local_conf_file)
> Got vm.conf from OVF_STORE
> MainThread::INFO::2016-08-08
> 15:03:11,675::hosted_engine::462::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
> Current state EngineUp (score: 3400)
> 
> 
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
> 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] 4.0 - 2nd node fails on deploy

2016-10-05 Thread Sahina Bose
[Adding gluster-users ML]
The brick logs are filled with errors :
[2016-10-05 19:30:28.659061] E [MSGID: 113077]
[posix-handle.c:309:posix_handle_pump] 0-engine-posix: malformed internal
link
/var/run/vdsm/storage/0a021563-91b5-4f49-9c6b-fff45e85a025/d84f0551-0f2b-457c-808c-6369c6708d43/1b5a5e34-818c-4914-8192-2f05733b5583
for
/xpool/engine/brick/.glusterfs/b9/8e/b98ed8d2-3bf9-4b11-92fd-ca5324e131a8
[2016-10-05 19:30:28.659069] E [MSGID: 113091] [posix.c:180:posix_lookup]
0-engine-posix: Failed to create inode handle for path

The message "E [MSGID: 113018] [posix.c:198:posix_lookup] 0-engine-posix:
lstat on null failed" repeated 3 times between [2016-10-05 19:30:28.656529]
and [2016-10-05 19:30:28.659076]
[2016-10-05 19:30:28.659087] W [MSGID: 115005]
[server-resolve.c:126:resolve_gfid_cbk] 0-engine-server:
b98ed8d2-3bf9-4b11-92fd-ca5324e131a8: failed to resolve (Success)

- Ravi, the above are from the data brick of the arbiter volume. Can you
take a look?

Jason,
Could you also provide the mount logs from the first host
(/var/log/glusterfs/rhev-data-center-mnt-glusterSD*engine.log) and glusterd
log (/var/log/glusterfs/etc-glusterfs-glusterd.vol.log) around the same
time frame.



On Wed, Oct 5, 2016 at 3:28 AM, Jason Jeffrey  wrote:

> Hi,
>
>
>
> Servers are powered  off  when I’m not looking at the problem.
>
>
>
> There may have been instances where all three were not powered on, during
> the same period.
>
>
>
> Glusterhd log attached, the xpool-engine-brick log is over 1 GB in size,
> I’ve taken a sample of the last  couple days, looks to be highly repative.
>
>
>
> Cheers
>
>
>
> Jason
>
>
>
>
>
>
>
>
>
> *From:* Simone Tiraboschi [mailto:stira...@redhat.com]
> *Sent:* 04 October 2016 16:50
>
> *To:* Jason Jeffrey 
> *Cc:* users 
> *Subject:* Re: [ovirt-users] 4.0 - 2nd node fails on deploy
>
>
>
>
>
>
>
> On Tue, Oct 4, 2016 at 5:22 PM, Jason Jeffrey  wrote:
>
> Hi,
>
>
>
> DCASTORXX is a hosts entry for dedicated  direct 10GB links (each private
> /28) between the x3 servers  i.e 1=> 2&3, 2=> 1&3, etc) planned to be used
> solely for storage.
>
>
>
> I,e
>
>
>
> 10.100.50.81dcasrv01
>
> 10.100.101.1dcastor01
>
> 10.100.50.82dcasrv02
>
> 10.100.101.2dcastor02
>
> 10.100.50.83dcasrv03
>
> 10.100.103.3dcastor03
>
>
>
> These were setup with the gluster commands
>
>
>
> · gluster volume create iso replica 3 arbiter 1
> dcastor01:/xpool/iso/brick   dcastor02:/xpool/iso/brick
> dcastor03:/xpool/iso/brick
>
> · gluster volume create export replica 3 arbiter 1
> dcastor02:/xpool/export/brick  dcastor03:/xpool/export/brick
> dcastor01:/xpool/export/brick
>
> · gluster volume create engine replica 3 arbiter 1
> dcastor01:/xpool/engine/brick dcastor02:/xpool/engine/brick
> dcastor03:/xpool/engine/brick
>
> · gluster volume create data replica 3 arbiter 1
> dcastor01:/xpool/data/brick  dcastor03:/xpool/data/brick
> dcastor02:/xpool/data/bricky
>
>
>
>
>
> So yes, DCASRV01 is the server (pri) and have local bricks access through
> DCASTOR01 interface
>
>
>
> Is the issue here not the incorrect soft link ?
>
>
>
> No, this should be fine.
>
>
>
> The issue is that periodically your gluster volume losses its server
> quorum and become unavailable.
>
> It happened more than once from your logs.
>
>
>
> Can you please attach also gluster logs for that volume?
>
>
>
>
>
> lrwxrwxrwx. 1 vdsm kvm  132 Oct  3 17:27 hosted-engine.metadata ->
> /var/run/vdsm/storage/bbb70623-194a-46d2-a164-76a4876ecaaf/fd44dbf9-473a-
> 496a-9996-c8abe3278390/cee9440c-4eb8-453b-bc04-c47e6f9cbc93
>
> [root@dcasrv01 /]# ls -al /var/run/vdsm/storage/bbb70623-194a-46d2-a164-
> 76a4876ecaaf/
>
> ls: cannot access /var/run/vdsm/storage/bbb70623-194a-46d2-a164-76a4876ecaaf/:
> No such file or directory
>
> But the data does exist
>
> [root@dcasrv01 fd44dbf9-473a-496a-9996-c8abe3278390]# ls -al
>
> drwxr-xr-x. 2 vdsm kvm4096 Oct  3 17:17 .
>
> drwxr-xr-x. 6 vdsm kvm4096 Oct  3 17:17 ..
>
> -rw-rw. 2 vdsm kvm 1028096 Oct  3 20:48 cee9440c-4eb8-453b-bc04-
> c47e6f9cbc93
>
> -rw-rw. 2 vdsm kvm 1048576 Oct  3 17:17 cee9440c-4eb8-453b-bc04-
> c47e6f9cbc93.lease
>
> -rw-r--r--. 2 vdsm kvm 283 Oct  3 17:17 
> cee9440c-4eb8-453b-bc04-c47e6f9cbc93.meta
>
>
>
>
> Thanks
>
>
>
> Jason
>
>
>
>
>
>
>
> *From:* Simone Tiraboschi [mailto:stira...@redhat.com]
> *Sent:* 04 October 2016 14:40
>
>
> *To:* Jason Jeffrey 
> *Cc:* users 
> *Subject:* Re: [ovirt-users] 4.0 - 2nd node fails on deploy
>
>
>
>
>
>
>
> On Tue, Oct 4, 2016 at 10:51 AM, Simone Tiraboschi 
> wrote:
>
>
>
>
>
> On Mon, Oct 3, 2016 at 11:56 PM, Jason Jeffrey  wrote:
>
> Hi,
>
>
>
> Another problem has appeared, after rebooting the primary the VM will not
> start.
>
>
>
> Appears the symlink is broken between gluster mount ref and vdsm
>
>
>
> The first host was correctly deployed but it seas that you are facing some
> issue connecting the storage.
>
> Can you please attach vdsm logs and /var/log/messages from the fir