On 01/17/2014 12:55 PM, Gianluca Cecchi wrote:
On Fri, Nov 29, 2013 at 11:48 AM, Dan Kenigsberg  wrote:
On Fri, Nov 29, 2013 at 04:04:03PM +0530, Vijay Bellur wrote:


There are two ways in which GlusterFS can be used as a storage domain:

a) Use gluster native/fuse access with POSIXFS

b) Use the gluster native storage domain to bypass fuse (with
libgfapi). We are currently addressing an issue in libvirt
(https://bugzilla.redhat.com/show_bug.cgi?id=1017289) to enable
snapshot support with libgfapi. Once this is addressed, we will have
libgfapi support in the native storage domain.

It won't be as immediate, since there's a required fix on Vdsm's side
(Bug 1022961 - Running a VM from a gluster domain uses mount instead of
gluster URI)

Till then, fuse would
be used with native storage domain. You can find more details about
native storage domain here:

http://www.ovirt.org/Features/GlusterFS_Storage_Domain
rg/mailman/listinfo/users

Hello,
revamping here....
I'm using oVirt 3.3.3 beta1 after upgrade from 3.3.2 on fedora 19
ovirt beta repo
It seems bug referred by Vijay (1017289) is still marked as
"assigned", but actually it is for rhel 6.
Bug referred by Dan (1022961) is marked as "blocked" but I don't see
any particular updates since late november. But it is for rhevm, so I
think it is for rhel 6...
So what is the situation for fedora 19 and oVirt in upcoming 3.3.3?
And upcoming fedora 19/20 and 3.4?


I ask because I see that in my qemu command line generated by oVirt there is:

for virtio (virtio-blk) disk
-drive 
file=/rhev/data-center/mnt/glusterSD/node01.mydomain:gvdata/d0b96d4a-62aa-4e9f-b50e-f7a0cb5be291/images/a5e4f67b-50b5-4740-9990-39deb8812445/53408cb0-bcd4-40de-bc69-89d59b7b5bc2,if=none,id=drive-virtio-disk0,format=raw,serial=a5e4f67b-50b5-4740-9990-39deb8812445,cache=none,werror=stop,rerror=stop,aio=threads

for virtio-scsi disk
-drive 
file=/rhev/data-center/mnt/glusterSD/node01.mydomain:gvdata/d0b96d4a-62aa-4e9f-b50e-f7a0cb5be291/images/c1477133-6b06-480d-a233-1dae08daf8b3/c2a82c64-9dee-42bb-acf2-65b8081f2edf,if=none,id=drive-scsi0-0-0-0,format=raw,serial=c1477133-6b06-480d-a233-1dae08daf8b3,cache=none,werror=stop,rerror=stop,aio=threads

So it is referred as mount point and not gluster:// way
Also, the output of "mount" command on hypervisors shows:

node01.mydomain:gvdata on
/rhev/data-center/mnt/glusterSD/node01.mydomain:gvdata type
fuse.glusterfs 
(rw,relatime,user_id=0,group_id=0,default_permissions,allow_other,max_read=131072)
fusectl on /sys/fs/fuse/connections type fusectl (rw,relatime)

and it seems indeed fuse mount, so not using libgfapi...
output of
"lsof -Pp pid" where pid is the qemu process shows libgfapi....:

qemu-syst 2057 qemu  mem       REG              253,0      99440
         541417 /usr/lib64/libgfapi.so.0.0.0

(btw: strange version 0.0.0 for a release.. not so reassuring ;-)

libgfapi uses a different internal versioning scheme and hence compatibility can be handled. However there is work happening in glusterfs upstream to change the version scheme for libraries and have it compatible with libtool guidelines.

# ll /usr/lib64/libgfapi.so.0*
lrwxrwxrwx. 1 root root    17 Jan  7 12:45 /usr/lib64/libgfapi.so.0 ->
libgfapi.so.0.0.0
-rwxr-xr-x. 1 root root 99440 Jan  3 13:35 /usr/lib64/libgfapi.so.0.0.0
)

At page http://www.gluster.org/category/qemu/ there is a schema about
mount types and benchmarks

1) FUSE mount
2) GlusterFS block driver in QEMU (FUSE bypass)
3) Base (VM image accessed directly from brick)
(
qemu-system-x86_64 –enable-kvm –nographic -smp 4 -m 2048 -drive
file=/test/F17,if=virtio,cache=none => /test is brick directory
)

I have not understood if we are in Base (best performance) or FUSE
(worst performance)


In this particular blog post, base refers to local disk storage and the FUSE performance was measured before a set of optimizations were done in GlusterFS. The "Optimize for Virt" action in oVirt ensures that these optimizations are enabled. Hence, even with FUSE the performance for VM image storage would be somewhere between the worst and best cases.


thanks in advance for clarifications and possible roadmaps...

If you are interested in seeing any new feature related to the oVirt-Gluster integration, please feel free to let us know as part of the GlusterFS 3.6 planning process [1] which is ongoing now.

- Vijay

[1] http://www.gluster.org/community/documentation/index.php/Planning36
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to