Thanks for helping I thought that with CEPH, the more servers you have the better the performance - that is why I am so confused
Also, I tried to add the 4th server ( still no luck - in fact the rado bench output I included was from 4 servers, one OSD on each, bluestore, replication 2 ) Here is the same rados bench but with only 2 servers/ 2 OSDs rados bench -p rbd 120 write --no-cleanup && rados bench -p rbd 120 seq hints = 1 Maintaining 16 concurrent writes of 4194304 bytes to objects of size 4194304 for up to 120 seconds or 0 objects Object prefix: benchmark_data_osd01_384454 sec Cur ops started finished avg MB/s cur MB/s last lat(s) avg lat(s) 0 0 0 0 0 0 - 0 1 16 101 85 339.981 340 0.244679 0.160362 2 16 149 133 265.972 192 0.303997 0.226268 3 16 197 181 241.307 192 0.179013 0.255609 4 16 241 225 224.975 176 0.353464 0.272623 5 16 289 273 218.376 192 0.303821 0.282425 6 16 338 322 214.643 196 0.326009 0.29105 7 16 387 371 211.977 196 0.27048 0.296497 8 16 436 420 209.977 196 0.287188 0.299224 9 16 479 463 205.755 172 0.380512 0.302272 10 16 527 511 204.378 192 0.289532 0.306163 11 16 576 560 203.614 196 0.406783 0.309271 12 16 624 608 202.645 192 0.282266 0.312167 13 16 667 651 200.286 172 0.377555 0.313733 14 16 716 700 199.978 196 0.350938 0.315445 15 16 764 748 199.445 192 0.183931 0.317474 Here is my iperf3 results 0.00-1.00 sec 1.15 GBytes 9.92 Gbits/sec 0 850 KBytes [ 4] 1.00-2.00 sec 1.15 GBytes 9.90 Gbits/sec 0 850 KBytes On Thu, 19 Apr 2018 at 08:28, Marc Roos <m.r...@f1-outsourcing.eu> wrote: > > If I may guess, because with 3 it reads from 3 and with 2 it reads only > from 2. You should be able to verify this with something like dstat -d > -D sda,sdb,sdc,sdd,sde,sdf,sdg not? > > With replication of 2, objects are still being stored among the 3 nodes. > > I am getting with iperf3 on 10Gbit > [ ID] Interval Transfer Bandwidth Retr Cwnd > [ 4] 0.00-10.00 sec 11.5 GBytes 9.89 Gbits/sec 0 1.31 MBytes > [ 4] 10.00-20.00 sec 11.5 GBytes 9.89 Gbits/sec 0 1.79 MBytes > > > > -----Original Message----- > From: Steven Vacaroaia [mailto:ste...@gmail.com] > Sent: donderdag 19 april 2018 14:11 > To: ceph-users > Subject: [ceph-users] ceph luminous 12.2.4 - 2 servers better than 3 ? > > Hi, > > Any idea why 2 servers with one OSD each will provide better performance > than 3 ? > > Servers are identical > Performance is impacted irrespective if I used SSD for WAL/DB or not > Basically, I am getting lots of cur MB/s zero > > Network is separate 10 GB for public and private I tested it with iperf > and I am getting 9.3 Gbs > > I have tried replication by 2 and 3 with same results ( much better for > 2 servers than 3 ) > > reinstalled CEPH multiple times > ceph.conf very simple - no major customization ( see below) I am out of > ideas - any hint will be TRULY appreciated > > Steven > > > > auth_cluster_required = cephx > auth_service_required = cephx > auth_client_required = cephx > > > public_network = 10.10.30.0/24 > cluster_network = 192.168.0.0/24 > > > osd_pool_default_size = 2 > osd_pool_default_min_size = 1 # Allow writing 1 copy in a degraded state > osd_crush_chooseleaf_type = 1 > > > [mon] > mon_allow_pool_delete = true > mon_osd_min_down_reporters = 1 > > [osd] > osd_mkfs_type = xfs > osd_mount_options_xfs = > "rw,noatime,nodiratime,attr2,logbufs=8,logbsize=256k,largeio,inode64,swa > lloc,allocsize=4M" > osd_mkfs_options_xfs = "-f -i size=2048" > bluestore_block_db_size = 32212254720 > bluestore_block_wal_size = 1073741824 > > rados bench -p rbd 120 write --no-cleanup && rados bench -p rbd 120 seq > hints = 1 Maintaining 16 concurrent writes of 4194304 bytes to objects > of size 4194304 for up to 120 seconds or 0 objects Object prefix: > benchmark_data_osd01_383626 > sec Cur ops started finished avg MB/s cur MB/s last lat(s) avg > lat(s) > 0 0 0 0 0 0 - > 0 > 1 16 57 41 163.991 164 0.197929 > 0.065543 > 2 16 57 41 81.992 0 - > 0.065543 > 3 16 67 51 67.9936 20 0.0164632 > 0.249939 > 4 16 67 51 50.9951 0 - > 0.249939 > 5 16 71 55 43.9958 8 0.0171439 > 0.319973 > 6 16 181 165 109.989 440 0.0159057 > 0.563746 > 7 16 182 166 94.8476 4 0.221421 > 0.561684 > 8 16 182 166 82.9917 0 - > 0.561684 > 9 16 240 224 99.5458 116 0.0232989 > 0.638292 > 10 16 264 248 99.1901 96 0.0222669 > 0.583336 > 11 16 264 248 90.1729 0 - > 0.583336 > 12 16 285 269 89.6579 42 0.0165706 > 0.600606 > 13 16 285 269 82.7611 0 - > 0.600606 > 14 16 310 294 83.9918 50 0.0254241 > 0.756351 > > > > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com