Re: [Users] Failing to attach NFS data storage domain (Ovirt 3.2)

2013-03-20 Thread Limor Gavish
Updating from sanlock-2.6-4.fc18 to sanlock-2.6-7.fc18 fixed it. Huge
thanks!
On Mar 20, 2013 10:24 PM, "Federico Simoncelli"  wrote:

> - Original Message -
> > From: "Limor Gavish" 
> > To: "Federico Simoncelli" 
> > Cc: "Yuval M" , users@ovirt.org, "Nezer Zaidenberg" <
> nzaidenb...@mac.com>, "Eli Mesika"
> > , "Maor Lipchuk" 
> > Sent: Wednesday, March 20, 2013 9:02:35 PM
> > Subject: Re: [Users] Failing to attach NFS data storage domain (Ovirt
> 3.2)
> >
> > Thank you very much for your response.
> >
> >
> > Attached VDSM logs as you requested (The VDSM logs where the NFS
> > domain was created were missing so we had to recreate the NFS
> > domain, therefore the sdUUID has changed).
> > Here is the rest of the commands you asked:
> >
> >
> > [root@bufferoverflow wil]# mount -t nfs
> > bufferoverflow:/home/BO_Ovirt_Storage /mnt/tmp
> > [root@bufferoverflow wil]# cd
> > /mnt/tmp/1083422e-a5db-41b6-b667-b9ef1ef244f0/dom_md/
> > [root@bufferoverflow dom_md]# ls -l
> > total 2052
> > -rw-rw 1 vdsm kvm 1048576 Mar 20 21:46 ids
> > -rw-rw 1 vdsm kvm 0 Mar 20 21:45 inbox
> > -rw-rw 1 vdsm kvm 2097152 Mar 20 21:45 leases
> > -rw-r--r-- 1 vdsm kvm 311 Mar 20 21:45 metadata
> > -rw-rw 1 vdsm kvm 0 Mar 20 21:45 outbox
> > [root@bufferoverflow dom_md]# sanlock direct dump ids
>
> Sorry I should have mentioned that if you use root_squash for your nfs
> share you have to switch to the vdsm user:
>
> (root)# su -s /bin/sh vdsm
> (vdsm)$ cd /mnt/tmp//dom_md/
>
> (vdsm)$ sanlock direct dump ids
> (and now you should be able to see the output)
>
> If the output is still empty then used hexdump -C to inspect it
> (and eventually post it here compressed).
>
> Another important thing that you should check is:
>
> # ps fax | grep sanlock
>
> If the output doesn't look like the following:
>
>  1966 ?SLs0:00 wdmd -G sanlock
>  2036 ?SLsl   0:00 sanlock daemon -U sanlock -G sanlock
>  2037 ?S  0:00  \_ sanlock daemon -U sanlock -G sanlock
>
> Then I suggest you to update sanlock to the latest build:
>
> http://koji.fedoraproject.org/koji/buildinfo?buildID=377815
> (sanlock-2.6-7.fc18)
>
> And eventually if after rebooting the problem persists, please post
> also the sanlock log (/var/log/sanlock.log)
>
> > Please note, the VDSM is running as a system service (it was
> > installed from a package) while ovirt-engine was built from sources
> > and thus is not running as root. Is this an issue?
>
> It shouldn't be.
>
> --
> Federico
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[Users] ovirt-iso-uploader

2013-03-20 Thread Limor Gavish
Hello,

I am running ovirt 3.2.1 (built from sources) on Fedora 18.
It seems that I don't have the ovirt-iso-uploader installed.

[wil@bufferoverflow ovirt-engine]$  ovirt-iso-uploader list
bash: ovirt-iso-uploader: command not found...
[wil@bufferoverflow ovirt-engine]$ sudo yum list installed | grep -i uplo
[wil@bufferoverflow ovirt-engine]$

What is the easiest way to install the ovirt-iso-uploader (preferably using
"yum install")?

Thanks,
Limor G
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] VM crashes and doesn't recover

2013-03-28 Thread Limor Gavish
t;> svdsm
>> > failed [Errno 2] No such file or directory
>> > Mar 24 19:57:52 bufferoverflow vdsm SuperVdsmProxy WARNING Connect to
>> svdsm
>> > failed [Errno 2] No such file or directory
>> > Mar 24 19:57:53 bufferoverflow vdsm SuperVdsmProxy WARNING Connect to
>> svdsm
>> > failed [Errno 2] No such file or directory
>> > Mar 24 19:57:54 bufferoverflow vdsm SuperVdsmProxy WARNING Connect to
>> svdsm
>> > failed [Errno 2] No such file or directory
>> > Mar 24 19:57:55 bufferoverflow vdsm SuperVdsmProxy WARNING Connect to
>> svdsm
>> > failed [Errno 2] No such file or directory
>> > Mar 24 19:57:55 bufferoverflow vdsm Storage.Misc ERROR Panic: Couldn't
>> > connect to supervdsm
>> > Mar 24 19:57:55 bufferoverflow respawn: slave '/usr/share/vdsm/vdsm'
>> died,
>> > respawning slave
>> > Mar 24 19:57:55 bufferoverflow vdsm fileUtils WARNING Dir
>> > /rhev/data-center/mnt already exists
>> > Mar 24 19:57:58 bufferoverflow vdsm vds WARNING Unable to load the json
>> rpc
>> > server module. Please make sure it is installed.
>> > Mar 24 19:57:58 bufferoverflow vdsm vm.Vm WARNING
>> > vmId=`4d3d81b3-d083-4569-acc2-8e631ed51843`::Unknown type found, device:
>> > '{'device': u'unix', 'alias': u'channel0', 'type': u'channel',
>> 'address':
>> > {u'bus': u'0', u'controller': u'0', u'type': u'virtio-serial', u'port':
>> > u'1'}}' found
>> > Mar 24 19:57:58 bufferoverflow vdsm vm.Vm WARNING
>> > vmId=`4d3d81b3-d083-4569-acc2-8e631ed51843`::Unknown type found, device:
>> > '{'device': u'unix', 'alias': u'channel1', 'type': u'channel',
>> 'address':
>> > {u'bus': u'0', u'controller': u'0', u'type': u'virtio-serial', u'port':
>> > u'2'}}' found
>> > Mar 24 19:57:58 bufferoverflow vdsm vm.Vm WARNING
>> > vmId=`4d3d81b3-d083-4569-acc2-8e631ed51843`::_readPauseCode unsupported
>> by
>> > libvirt vm
>> > Mar 24 19:57:58 bufferoverflow kernel: [ 7402.688177] ata1: hard
>> resetting
>> > link
>> > Mar 24 19:57:59 bufferoverflow kernel: [ 7402.994510] ata1: SATA link up
>> > 6.0 Gbps (SStatus 133 SControl 300)
>> > Mar 24 19:57:59 bufferoverflow kernel: [ 7403.005510] ACPI Error: [DSSP]
>> > Namespace lookup failure, AE_NOT_FOUND (20120711/psargs-359)
>> > Mar 24 19:57:59 bufferoverflow kernel: [ 7403.005517] ACPI Error: Method
>> > parse/execution failed [\_SB_.PCI0.SAT0.SPT0._GTF] (Node
>> 880407c74d48),
>> > AE_NOT_FOUND (20120711/psparse-536)
>> > Mar 24 19:57:59 bufferoverflow kernel: [ 7403.015485] ACPI Error: [DSSP]
>> > Namespace lookup failure, AE_NOT_FOUND (20120711/psargs-359)
>> > Mar 24 19:57:59 bufferoverflow kernel: [ 7403.015493] ACPI Error: Method
>> > parse/execution failed [\_SB_.PCI0.SAT0.SPT0._GTF] (Node
>> 880407c74d48),
>> > AE_NOT_FOUND (20120711/psparse-536)
>> > Mar 24 19:57:59 bufferoverflow kernel: [ 7403.016061] ata1.00:
>> configured
>> > for UDMA/133
>> > Mar 24 19:57:59 bufferoverflow kernel: [ 7403.016066] ata1: EH complete
>> > Mar 24 19:58:01 bufferoverflow sanlock[1208]: 2013-03-24 19:58:01+0200
>> 7422
>> > [4759]: 1083422e close_task_aio 0 0x7ff3740008c0 busy
>> > Mar 24 19:58:01 bufferoverflow sanlock[1208]: 2013-03-24 19:58:01+0200
>> 7422
>> > [4759]: 1083422e close_task_aio 1 0x7ff374000910 busy
>> > Mar 24 19:58:01 bufferoverflow sanlock[1208]: 2013-03-24 19:58:01+0200
>> 7422
>> > [4759]: 1083422e close_task_aio 2 0x7ff374000960 busy
>> > Mar 24 19:58:01 bufferoverflow sanlock[1208]: 2013-03-24 19:58:01+0200
>> 7422
>> > [4759]: 1083422e close_task_aio 3 0x7ff3740009b0 busy
>> > Mar 24 19:58:01 bufferoverflow kernel: [ 7405.714145] device-mapper:
>> table:
>> > 253:0: multipath: error getting device
>> > Mar 24 19:58:01 bufferoverflow kernel: [ 7405.714148] device-mapper:
>> ioctl:
>> > error adding target to table
>> > Mar 24 19:58:01 bufferoverflow kernel: [ 7405.715051] device-mapper:
>> table:
>> > 253:0: multipath: error getting device
>> > Mar 24 19:58:01 bufferoverflow kernel: [ 7405.715053] device-mapper:
>> ioctl:
>> > error

Re: [Users] oVirt storage is down and doesn't come up

2013-04-15 Thread Limor Gavish
r a Postgres ui if you have one.
> In the results you will see the storage connection in the format of
> %hostname%:/%mountName%, then in the VDSM server check in the mount list
> that you see that it is mounted, the mount itself should contain a
> directory named as the uuid of the master domain, let me know the result.
>
> Tal.
>
>
>
>
> On 04/12/2013 07:29 PM, Limor Gavish wrote:
>
> Hi,
>
>  For some reason, without doing anything, all the storage domains became
> down and restarting VDSM or the entire machine do not bring it up.
> I am not using lvm
> The following errors appear several times in vdsm.log (full logs are
> attached):
>
>  Thread-22::WARNING::2013-04-12
> 19:00:08,597::lvm::378::Storage.LVM::(_reloadvgs) lvm vgs failed: 5 [] ['
>  Volume group "1083422e-a5db-41b6-b667-b9ef1ef244f0" not found']
> Thread-22::DEBUG::2013-04-12
> 19:00:08,598::lvm::402::OperationMutex::(_reloadvgs) Operation 'lvm reload
> operation' released the operation mutex
> Thread-22::DEBUG::2013-04-12
> 19:00:08,681::resourceManager::615::ResourceManager::(releaseResource)
> Trying to release resource 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3'
> Thread-22::DEBUG::2013-04-12
> 19:00:08,681::resourceManager::634::ResourceManager::(releaseResource)
> Released resource 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3' (0 active
> users)
> Thread-22::DEBUG::2013-04-12
> 19:00:08,681::resourceManager::640::ResourceManager::(releaseResource)
> Resource 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3' is free, finding
> out if anyone is waiting for it.
> Thread-22::DEBUG::2013-04-12
> 19:00:08,682::resourceManager::648::ResourceManager::(releaseResource) No
> one is waiting for resource 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3',
> Clearing records.
>  Thread-22::ERROR::2013-04-12
> 19:00:08,682::task::850::TaskManager.Task::(_setError)
> Task=`e35a22ac-771a-4916-851f-2fe9d60a0ae6`::Unexpected error
> Traceback (most recent call last):
>   File "/usr/share/vdsm/storage/task.py", line 857, in _run
> return fn(*args, **kargs)
>   File "/usr/share/vdsm/logUtils.py", line 45, in wrapper
> res = f(*args, **kwargs)
>   File "/usr/share/vdsm/storage/hsm.py", line 939, in connectStoragePool
> masterVersion, options)
>   File "/usr/share/vdsm/storage/hsm.py", line 986, in _connectStoragePool
> res = pool.connect(hostID, scsiKey, msdUUID, masterVersion)
>   File "/usr/share/vdsm/storage/sp.py", line 695, in connect
> self.__rebuild(msdUUID=msdUUID, masterVersion=masterVersion)
>   File "/usr/share/vdsm/storage/sp.py", line 1232, in __rebuild
> masterVersion=masterVersion)
>   File "/usr/share/vdsm/storage/sp.py", line 1576, in getMasterDomain
>  raise se.StoragePoolMasterNotFound(self.spUUID, msdUUID)
> StoragePoolMasterNotFound: Cannot find master domain:
> 'spUUID=5849b030-626e-47cb-ad90-3ce782d831b3,
> msdUUID=1083422e-a5db-41b6-b667-b9ef1ef244f0'
> Thread-22::DEBUG::2013-04-12
> 19:00:08,685::task::869::TaskManager.Task::(_run)
> Task=`e35a22ac-771a-4916-851f-2fe9d60a0ae6`::Task._run:
> e35a22ac-771a-4916-851f-2fe9d60a0ae6
> ('5849b030-626e-47cb-ad90-3ce782d831b3', 1,
> '5849b030-626e-47cb-ad90-3ce782d831b3',
> '1083422e-a5db-41b6-b667-b9ef1ef244f0', 3942) {} failed - stopping task
> Thread-22::DEBUG::2013-04-12
> 19:00:08,685::task::1194::TaskManager.Task::(stop)
> Task=`e35a22ac-771a-4916-851f-2fe9d60a0ae6`::stopping in state preparing
> (force False)
> Thread-22::DEBUG::2013-04-12
> 19:00:08,685::task::974::TaskManager.Task::(_decref)
> Task=`e35a22ac-771a-4916-851f-2fe9d60a0ae6`::ref 1 aborting True
> Thread-22::INFO::2013-04-12
> 19:00:08,686::task::1151::TaskManager.Task::(prepare)
> Task=`e35a22ac-771a-4916-851f-2fe9d60a0ae6`::aborting: Task is aborted:
> 'Cannot find master domain' - code 304
>
>  *[wil@bufferoverflow ~]$ **sudo vgs --noheadings --units b --nosuffix
> --separator \| -o
> uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free
> *
>   No volume groups found
>
>  *[wil@bufferoverflow ~]$ **mount*
> proc on /proc type proc (rw,nosuid,nodev,noexec,relatime)
> sysfs on /sys type sysfs (rw,nosuid,nodev,noexec,relatime)
> devtmpfs on /dev type devtmpfs
> (rw,nosuid,size=8131256k,nr_inodes=2032814,mode=755)
> securityfs on /sys/kernel/security type securityfs
> (rw,nosuid,nodev,noexec,relatime)
> tmpfs on /dev/shm type tmpfs (rw,nosuid,nodev)
> devpts on /dev/pts type devpts
> (rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000)
> tmpfs on /run type tmpfs (rw,nosuid,nodev,

Re: [Users] Unable to add host in a development setup - "Host kernelpanic.home installation failed. Failed to decode own public key from store 'keys/engine.p12' using alias '1'."

2013-04-28 Thread Limor Gavish
Thank you very much for your prompt response. However, according to "git
rebase" (after "git fetch") there is no new change

[wil@bufferoverflow repository]$ git fetch
https://github.com/alonbl/ovirt-engine.git otopi
>From https://github.com/alonbl/ovirt-engine
 * branchotopi  -> FETCH_HEAD
[wil@bufferoverflow repository]$ git rebase --merge FETCH_HEAD
*Current branch otopi is up to date.*
*
*
The latest change I see in "git log" is

[wil@bufferoverflow repository]$ git show HEAD
commit 99d95891ab0e814159a463d246543b08e41c746d
Author: Alon Bar-Lev 
Date:   Thu Apr 25 12:39:50 2013 +0300

packaging: engine-setup re-write using otopi

Change-Id: I4d5e3ac7927802326e324c9b98ada75b421638aa
Signed-off-by: Alon Bar-Lev 
Signed-off-by: Sandro Bonazzola 
Signed-off-by: Alex Lourie 

If you don't wish to push the change yet, you can send a patch and I will
apply it.


On Mon, Apr 29, 2013 at 1:04 AM, Alon Bar-Lev  wrote:

>
>
> ----- Original Message -
> > From: "Limor Gavish" 
> > To: "Alon Bar-Lev" 
> > Cc: "Yuval M" , users@ovirt.org, "Nezer Zaidenberg" <
> nzaidenb...@mac.com>, "Vered Volansky"
> > 
> > Sent: Sunday, April 28, 2013 10:56:45 PM
> > Subject: Re: [Users] Unable to add host in a development setup - "Host
> kernelpanic.home installation failed. Failed
> > to decode own public key from store 'keys/engine.p12' using alias '1'."
> >
> > Hi,
> >
> > Thank you very much for your help but I do not succeed to start the
> engine
> > after compiling your version.
> >
> > I switched to the otopi branch like you said.
> > git fetch https://github.com/alonbl/ovirt-engine.git otopi
> > git checkout -b otopi FETCH_HEAD
> >
> > I followed the instructions in
> > https://github.com/alonbl/ovirt-engine/blob/otopi/README.developer
> > *So here are the last 3 commands I used:*
> > make clean install-dev PREFIX=$HOME/ovirt-engine
> > $HOME/ovirt-engine/bin/engine-setup-2
> > --jboss-home=$HOME/ovirt-engine/jboss-as-7.1.1.Final
> > $HOME/ovirt-engine/share/ovirt-engine/services/ovirt-engine.py start
> >
> > However, the last command
> > ($HOME/ovirt-engine/share/ovirt-engine/services/ovirt-engine.py start)
> > hangs.
>
> This is expected... it is the service running, Ctrl-C to terminate.
>
> > *In server.log I see the following error:*
> > 2013-04-28 22:13:40,203 ERROR [org.jboss.as] (MSC service thread 1-3)
> > JBAS015875: JBoss AS 7.1.1.Final "Brontes" started (with errors) in
> 2788ms
> > - Started 504 of 5 40 services (1 services failed or missing
> > dependencies, 34 services are passive or on-demand)
> > *In engine.log I see the following error:*
> > 2013-04-28 22:13:40,082 ERROR
> > [org.ovirt.engine.core.utils.ErrorTranslatorImpl] (MSC service thread
> 1-4)
> > File: bundles/VdsmErrors could not be loaded: java.util.Mis
> >  singResourceException: Can't find bundle for base name
> bundles/VdsmErrors,
> > locale en_US
> >
> > Full logs are attached.
>
> The problem is that for some reason the PREFIX is not taken into account.
>
> Related errors:
> 2013-04-28 19:57:46,468 WARN  [org.ovirt.engine.core.utils.LocalConfig]
> (MSC service thread 1-3) The file
> "/usr/share/ovirt-engine/conf/engine.conf.defaults" doesn't exist or isn't
> readable. Will return an empty set of properties.
> 2013-04-28 19:57:46,468 WARN  [org.ovirt.engine.core.utils.LocalConfig]
> (MSC service thread 1-3) The file "/etc/ovirt-engine/engine.conf" doesn't
> exist or isn't readable. Will return an empty set of properties.
>
> It can be due to previous make execution without PREFIX.
> I've committed a change to Makefile to force regenerate the file locations
> when using install-dev.
> Can you please give it a try?
>
> Thanks!
>
> > Thanks,
> > Limor on behalf of Yuval
> >
> >
> > On Sun, Apr 28, 2013 at 2:51 PM, Alon Bar-Lev  wrote:
> >
> > >
> > > Maybe someone can help you in this.
> > >
> > > I prefer people install development mode and use fully functional
> product
> > > within their development environment.
> > >
> > > Alon
> > >
> > > - Original Message -
> > > > From: "Yuval M" 
> > > > To: "Alon Bar-Lev" 
> > > > Cc: users@ovirt.org, "Nezer Zaidenberg" ,
> "Limor
> > > Gavish" , "Vered Volansky"
> > > > 
> > > > Sent: Sunday, 

Re: [Users] ovirt shows vm status as down, but vm is actually up and running

2013-05-03 Thread Limor Gavish
You can try to compress the file:
sudo yum install bzip2
bzip2 vdsm.log


On Fri, May 3, 2013 at 10:08 AM, Pradipta Kumar Banerjee  wrote:

> On 05/03/2013 01:10 AM, Dan Kenigsberg wrote:
> > On Thu, May 02, 2013 at 11:26:41PM +0530, Pradipta Kumar Banerjee wrote:
> >> Hi,
> >> I have a situation where VM is actually up and running (I can access
> its VNC
> >> console), however oVirt engine ui shows the status as down.
> >> I'm using ovirt engine v3.2 from stable repo on F18
> >> Any suggestions to trouble shoot the issue ?
> >
> > That's like the worse thing that can happen - if you ever make decide to
> > start another instance of that VM on another host - so please tread
> > carefully.
> >
> > Does the host that run the VM shows as Up in Engine?
> Yes the host is up
> > What does
> > vdsClient -s 0 list table
> > shows there?
>
> Empty output
> >
> > and
> > virsh -r list
> This lists the VM. Also I'm able to access the console of the VM and work
> on it.
>
> > ?
> >
> > When is the last time that the VM was known to be up by Engine? Do you
> > have vdsm.log from that time?
>
> >
> > It is very important to the project that you collect as much information
> > about the lifecycle of this VM (i.e. was it migrated from somewhere,
> > complete vdsm.log since creation). However, you may have a conflicting
> > interest if you value that VM's data. You should probably shut it down
> > cleanly from within the guest in that case.
> No migration, VM pinned to specific host itself.
> Its a test setup so no worries on data corruption. However I agree its
> important
> to find out the root cause.
>
> Btw I should mention that I'm using ovirt-engine on x86_64 and vdsm on IBM
> Power
> host. Not sure if its a host specific issue.
> I'll see if I can recreate the issue.
>
> vdsm.log file is around 11MB. Is there any place where I can upload it?
>
>
>
> >
> > Dan.
> >
>
>
> --
> Regards,
> Pradipta
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users