Hi Christian, Thanks for your reply, here're the test specs: >>> [global] ioengine=libaio runtime=90 direct=1 group_reporting iodepth=16 ramp_time=5 size=1G
[seq_w_4k_20] bs=4k filename=seq_w_4k_20 rw=write numjobs=20 [seq_w_1m_20] bs=1m filename=seq_w_1m_20 rw=write numjobs=20 <<<< Test results: 4k - aggrb=13245KB/s, 1m - aggrb=1102.6MB/s Mount options: ceph-fuse /ceph -m 10.3.138.36:6789 Ceph configurations: >>>> filestore_xattr_use_omap = true auth cluster required = cephx auth service required = cephx auth client required = cephx osd journal size = 128 osd pool default size = 2 osd pool default min size = 1 osd pool default pg num = 512 osd pool default pgp num = 512 osd crush chooseleaf type = 1 <<<< Other configurations are all default. Status: health HEALTH_OK monmap e5: 5 mons at {1= 10.3.138.37:6789/0,2=10.3.138.39:6789/0,3=10.3.138.40:6789/0,4=10.3.138.59:6789/0,GGZ-YG-S0311-PLATFORM-138=10.3.138.36:6789/0 } election epoch 28, quorum 0,1,2,3,4 GGZ-YG-S0311-PLATFORM-138,1,2,3,4 mdsmap e55: 1/1/1 up {0=1=up:active} osdmap e1290: 20 osds: 20 up, 20 in pgmap v7180: 1000 pgs, 2 pools, 14925 MB data, 3851 objects 37827 MB used, 20837 GB / 21991 GB avail 1000 active+clean On Fri, 25 Mar 2016 at 16:44 Christian Balzer <ch...@gol.com> wrote: > > Hello, > > On Fri, 25 Mar 2016 08:11:27 +0000 Zhang Qiang wrote: > > > Hi all, > > > > According to fio, > Exact fio command please. > > >with 4k block size, the sequence write performance of > > my ceph-fuse mount > > Exact mount options, ceph config (RBD cache) please. > > >is just about 20+ M/s, only 200 Mb of 1 Gb full > > duplex NIC outgoing bandwidth was used for maximum. But for 1M block > > size the performance could achieve as high as 1000 M/s, approaching the > > limit of the NIC bandwidth. Why the performance stats differs so mush > > for different block sizes? > That's exactly why. > You can see that with local attached storage as well, many small requests > are slower than large (essential sequential) writes. > Network attached storage in general (latency) and thus Ceph as well (plus > code overhead) amplify that. > > >Can I configure ceph-fuse mount's block size > > for maximum performance? > > > Very little to do with that if you're using sync writes (thus the fio > command line pleasE), if not RBD cache could/should help. > > Christian > > > Basic information about the cluster: 20 OSDs on separate PCIe hard disks > > distributed across 2 servers, each with write performance about 300 M/s; > > 5 MONs; 1 MDS. Ceph version 0.94.6 > > (e832001feaf8c176593e0325c8298e3f16dfb403). > > > > Thanks :) > > > -- > Christian Balzer Network/Systems Engineer > ch...@gol.com Global OnLine Japan/Rakuten Communications > http://www.gol.com/ >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com