Here are the stats with direct io. dd of=ddbenchfile if=/dev/zero bs=8K count=1000000 oflag=direct 8192000000 bytes (8.2 GB) copied, 68.4789 s, 120 MB/s
dd if=ddbenchfile of=/dev/null bs=8K 8192000000 bytes (8.2 GB) copied, 19.7318 s, 415 MB/s These numbers are still over all much faster than when using RADOS bench. The replica is set to 2. The Journals are on the same disk but separate partitions. I kept the block size the same 8K. On Tue, Sep 17, 2013 at 11:37 AM, Campbell, Bill < bcampb...@axcess-financial.com> wrote: > As Gregory mentioned, your 'dd' test looks to be reading from the cache > (you are writing 8GB in, and then reading that 8GB out, so the reads are > all cached reads) so the performance is going to seem good. You can add > the 'oflag=direct' to your dd test to try and get a more accurate reading > from that. > > RADOS performance from what I've seen is largely going to hinge on replica > size and journal location. Are your journals on separate disks or on the > same disk as the OSD? What is the replica size of your pool? > > ------------------------------ > *From: *"Jason Villalta" <ja...@rubixnet.com> > *To: *"Bill Campbell" <bcampb...@axcess-financial.com> > *Cc: *"Gregory Farnum" <g...@inktank.com>, "ceph-users" < > ceph-users@lists.ceph.com> > *Sent: *Tuesday, September 17, 2013 11:31:43 AM > > *Subject: *Re: [ceph-users] Ceph performance with 8K blocks. > > Thanks for you feed back it is helpful. > > I may have been wrong about the default windows block size. What would be > the best tests to compare native performance of the SSD disks at 4K blocks > vs Ceph performance with 4K blocks? It just seems their is a huge > difference in the results. > > > On Tue, Sep 17, 2013 at 10:56 AM, Campbell, Bill < > bcampb...@axcess-financial.com> wrote: > >> Windows default (NTFS) is a 4k block. Are you changing the allocation >> unit to 8k as a default for your configuration? >> >> ------------------------------ >> *From: *"Gregory Farnum" <g...@inktank.com> >> *To: *"Jason Villalta" <ja...@rubixnet.com> >> *Cc: *ceph-users@lists.ceph.com >> *Sent: *Tuesday, September 17, 2013 10:40:09 AM >> *Subject: *Re: [ceph-users] Ceph performance with 8K blocks. >> >> >> Your 8k-block dd test is not nearly the same as your 8k-block rados bench >> or SQL tests. Both rados bench and SQL require the write to be committed to >> disk before moving on to the next one; dd is simply writing into the page >> cache. So you're not going to get 460 or even 273MB/s with sync 8k >> writes regardless of your settings. >> >> However, I think you should be able to tune your OSDs into somewhat >> better numbers -- that rados bench is giving you ~300IOPs on every OSD >> (with a small pipeline!), and an SSD-based daemon should be going faster. >> What kind of logging are you running with and what configs have you set? >> >> Hopefully you can get Mark or Sam or somebody who's done some performance >> tuning to offer some tips as well. :) >> -Greg >> >> On Tuesday, September 17, 2013, Jason Villalta wrote: >> >>> Hello all, >>> I am new to the list. >>> >>> I have a single machines setup for testing Ceph. It has a dual proc 6 >>> cores(12core total) for CPU and 128GB of RAM. I also have 3 Intel 520 >>> 240GB SSDs and an OSD setup on each disk with the OSD and Journal in >>> separate partitions formatted with ext4. >>> >>> My goal here is to prove just how fast Ceph can go and what kind of >>> performance to expect when using it as a back-end storage for virtual >>> machines mostly windows. I would also like to try to understand how it >>> will scale IO by removing one disk of the three and doing the benchmark >>> tests. But that is secondary. So far here are my results. I am aware >>> this is all sequential, I just want to know how fast it can go. >>> >>> DD IO test of SSD disks: I am testing 8K blocks since that is the >>> default block size of windows. >>> dd of=ddbenchfile if=/dev/zero bs=8K count=1000000 >>> 8192000000 bytes (8.2 GB) copied, 17.7953 s, 460 MB/s >>> >>> dd if=ddbenchfile of=/dev/null bs=8K >>> 8192000000 bytes (8.2 GB) copied, 2.94287 s, 2.8 GB/s >>> >>> RADOS bench test with 3 SSD disks and 4MB object size(Default): >>> rados --no-cleanup bench -p pbench 30 write >>> Total writes made: 2061 >>> Write size: 4194304 >>> Bandwidth (MB/sec): 273.004 >>> >>> Stddev Bandwidth: 67.5237 >>> Max bandwidth (MB/sec): 352 >>> Min bandwidth (MB/sec): 0 >>> Average Latency: 0.234199 >>> Stddev Latency: 0.130874 >>> Max latency: 0.867119 >>> Min latency: 0.039318 >>> ----- >>> rados bench -p pbench 30 seq >>> Total reads made: 2061 >>> Read size: 4194304 >>> Bandwidth (MB/sec): 956.466 >>> >>> Average Latency: 0.0666347 >>> Max latency: 0.208986 >>> Min latency: 0.011625 >>> >>> This all looks like I would expect from using three disks. The problems >>> appear to come with the 8K blocks/object size. >>> >>> RADOS bench test with 3 SSD disks and 8K object size(8K blocks): >>> rados --no-cleanup bench -b 8192 -p pbench 30 write >>> Total writes made: 13770 >>> Write size: 8192 >>> Bandwidth (MB/sec): 3.581 >>> >>> Stddev Bandwidth: 1.04405 >>> Max bandwidth (MB/sec): 6.19531 >>> Min bandwidth (MB/sec): 0 >>> Average Latency: 0.0348977 >>> Stddev Latency: 0.0349212 >>> Max latency: 0.326429 >>> Min latency: 0.0019 >>> ------ >>> rados bench -b 8192 -p pbench 30 seq >>> Total reads made: 13770 >>> Read size: 8192 >>> Bandwidth (MB/sec): 52.573 >>> >>> Average Latency: 0.00237483 >>> Max latency: 0.006783 >>> Min latency: 0.000521 >>> >>> So are these performance correct or is this something I missed with the >>> testing procedure? The RADOS bench number with 8K block size are the same >>> we see when testing performance in an VM with SQLIO. Does anyone know of >>> any configure changes that are needed to get the Ceph performance closer to >>> native performance with 8K blocks? >>> >>> Thanks in advance. >>> >>> >>> >>> -- >>> -- >>> *Jason Villalta* >>> Co-founder >>> [image: Inline image 1] >>> 800.799.4407x1230 | www.RubixTechnology.com<http://www.rubixtechnology.com/> >>> >> >> >> -- >> Software Engineer #42 @ http://inktank.com | http://ceph.com >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> >> *NOTICE: Protect the information in this message in accordance with the >> company's security policies. If you received this message in error, >> immediately notify the sender and destroy all copies.* >> >> > > > -- > -- > *Jason Villalta* > Co-founder > [image: Inline image 1] > 800.799.4407x1230 | www.RubixTechnology.com<http://www.rubixtechnology.com/> > > > *NOTICE: Protect the information in this message in accordance with the > company's security policies. If you received this message in error, > immediately notify the sender and destroy all copies.* > > -- -- *Jason Villalta* Co-founder [image: Inline image 1] 800.799.4407x1230 | www.RubixTechnology.com<http://www.rubixtechnology.com/>
<<EmailLogo.png>>
<<EmailLogo.png>>
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com