cool, thanks.
2013/9/11 John Skinner <john.skin...@appcore.com> > I ran each test independently for each block size. > > iozone -I -i 0 -i 1 -i 2 -r 4k -s 1G > > In order: -I to specify direct-IO, -i 0 to specify write/rewrite, -i 1 to > specify read/re-read, -i 2 to specify random read/write, -r specify block > size, -s to specify file size. > > The report is pretty thorough. What I put in the email was different, I > just took the throughput from each test and put it into a table outside of > the full report. > > ----- Original Message ----- > > From: "Rafael Weingartner" <rafaelweingart...@gmail.com> > To: users@cloudstack.apache.org > Sent: Wednesday, September 11, 2013 4:16:23 PM > Subject: Re: Cloustack 4.1.0 + GlusterFS > > I have never used iozone before, > > How did you get that report? > > I tried: iozone -s 1024m -t 1 -R > > But the report was pretty different from yours. > > > 2013/9/11 John Skinner <john.skin...@appcore.com> > > > I currently have GlusterFS deployed into an 8 node KVM cluster running on > > CloudStack 4.1 for primary storage. Gluster is deployed on 28 1TB drives > > across 2 separate storage appliances using a distributed-replicated > volume > > with the replica set to 2. The storage network is 10Gb copper. > > > > These are the options I have configured for the volume in Gluster, most > of > > them are from a Red Hat document on configuring Red Hat Enterprise > Storage > > for VM hosting: > > > > > > > > performance.io-thread-count: 32 > > performance.cache-size: 1024MB > > performance.write-behind-window-size: 5MB > > performance.write-behind: on > > network.remote-dio: on > > cluster.eager-lock: enable > > performance.stat-prefetch: off > > performance.io-cache: on > > performance.read-ahead: on > > performance.quick-read: on > > > > Here are some of the numbers I was getting when benchmarking the storage > > from the KVM node directly (not a VM) > > > > The below table is in KB/s. The test is single stream 1GB file utilizing > > Direct I/O (no cache). I used iozone to run the benchmark. > > > > Write 4k 45729 > > Read 4k 10189 > > Random Write 4k 31983 > > Random Read 4k 9859 > > Write 16k 182246 > > Read 16k 37146 > > Random Write 16k 113026 > > Random Read 16k 37237 > > Write 64k 420908 > > Read 64k 125315 > > Random Write 64k 383848 > > Random Read 64k 125218 > > Write 256k 567501 > > Read 256k 218413 > > Random Write 256k 508650 > > Random Read 256k 229117 > > > > In the above results, I have the volume mounted to each KVM host as a > FUSE > > glusterfs file system. They are added to CloudStack as a shared mount > > point. In the future it would be great to utilize GlusterFS qemu libvirt > > integration with libgfapi so I could bypass fuse altogether. However, > that > > would require adding that code to CloudStack to support that. > > > > I maybe have 15 or so VMs running from the storage now and it is still > > pretty snappy. Need to do some more testing though and really get it > loaded. > > > > ----- Original Message ----- > > > > > From: "Rafael Weingartner" <rafaelweingart...@gmail.com> > > To: users@cloudstack.apache.org > > Sent: Wednesday, September 11, 2013 8:48:07 AM > > Subject: Re: Cloustack 4.1.0 + GlusterFS > > > > Right now I can think in three main reasons: > > > > The first reason, performance (I do not know Gluster and its performance > > and if the read and write speed are satisfactory). Please if you can > make a > > test, post the results. > > > > Second consistency, I do not know Gluster, but swift that is also a > > Distributed File System is not consistency and they make it pretty clear > on > > their page (http://docs.openstack.org/developer/swift/) > > > > "Swift is a highly available, distributed, eventually consistent > > object/blob store...". > > > > I would not accept to storage my VMs images on a FS that is eventually > > consistent. > > > > Third, network, I haven't used this kind of FS, but I can image that it > > uses a lot of bandwidth to keep synchronizing, managing and securing the > > data. So, managing the networking would be a pain. > > > > > > > > 2013/9/11 Olivier Mauras <oliv...@core-hosting.net> > > > > > > > > > > > Hi, > > > > > > Those thinking that it's not a good idea, do you mind > > > explaining your point of view? > > > GlusterFS seems like the only real > > > alternative to a highly priced SAN for the primary storage... > > > > > > > > > Thanks, > > > Olivier > > > > > > On 2013-09-11 15:08, Rafael Weingartner wrote: > > > > > > > I > > > totally agree with Tomasz, I do not think that using a distributed FS > > > as > > > > primary storage is a good idea, but as a secondary it sounds > > > interesting. > > > > > > > > But, off course you can try *;*) > > > > > > > > 2013/9/11 > > > Shanker Balan <shanker.ba...@shapeblue.com> > > > > > > > >> On 11-Sep-2013, at > > > 5:14 PM, Tomasz Zięba <t.a.zi...@gmail.com [1]> wrote: > > > >> > > > >>> Hi, Some > > > time ago I tried to use GlusterFS as storage for cloudstacka but I > > > noticed that cloudstack uses the default settings for mount command. By > > > default mount command is using the UDP protocol but glusterfs works > > > >> > > > only > > > >> > > > >>> using tcp. I think, if cloudstack developers could add "-o > > > proto=tcp" to code glusterfs should works. For example: /bin/mount -t > > > nfs -o proto=tcp IP:/share /mnt/gluster/ If you are using CitrixXen you > > > should mount the share and make it as SR. For cloudstacka is clear > > > because you should use the option PreSetup when creating > PrimaryStorage. > > > Personally, I doubt that using GlusterFS as a primary storage is a good > > > solution but for secondary storage it should be very usefull. > > > >> And > > > maybe as a Swift backend. -- @shankerbalan M: +91 98860 60539 | O: +91 > > > (80) 67935867 shanker.ba...@shapeblue.com [2] | www.shapeblue.com [3] > | > > > Twitter:@shapeblue ShapeBlue Services India LLP, 22nd floor, Unit > 2201A, > > > World Trade Centre, Bangalore - 560 055 This email and any attachments > > > to it may be confidential and are intended solely for the use of the > > > individual to whom it is addressed. Any views or opinions expressed are > > > solely those of the author and do not necessarily represent those of > > > Shape Blue Ltd or related companies. If you are not the intended > > > recipient of this email, you must neither take any action based upon > its > > > contents, nor copy or show it to anyone. Please contact the sender if > > > you believe you have received this email in error. Shape Blue Ltd is a > > > company incorporated in England & Wales. ShapeBlue Services India LLP > is > > > operated under license from Shape Blue Ltd. ShapeBlue is a registered > > > trademark. > > > > > > > > -- Rafael Weingartner > > > > > > > > > > > > Links: > > > ------ > > > [1] > > > mailto:t.a.zi...@gmail.com > > > [2] mailto:shanker.ba...@shapeblue.com > > > [3] > > > http://www.shapeblue.com > > > > > > > > > > > -- > > Rafael Weingartner > > > > > > > -- > Rafael Weingartner > > -- Rafael Weingartner