I'm using filestore with SSD journal and a 3x replication. I've only noticed the low client IO after luminous upgrade, the actual traffic should be much higher. It had never been that low since my giant deployment (yup, it is a very old cluster)
Regards, Horace Ng ----- Original Message ----- From: "John Spray" <jsp...@redhat.com> To: "horace" <hor...@hkisl.net> Cc: "ceph-users" <ceph-users@lists.ceph.com> Sent: Friday, May 11, 2018 7:04:56 PM Subject: Re: [ceph-users] Inaccurate client io stats On Fri, May 11, 2018 at 4:51 AM, Horace <hor...@hkisl.net> wrote: > Hi everyone, > > I've got a 3-node cluster running without any issue. However, I found out > that since upgraded to luminous, the client io stat is far too way off from > the real one. Have no idea how to troubleshoot this after went through all > the logs. Any help would be appreciated. The ratio from logical IO (from clients) to raw IO (to disks) depends on configuration: - Are you using filestore or bluestore? Any SSD journals? - What replication level is in use? 3x? If you're using filestore, no SSD journals, and 3x journalling, then there will be a factor of six amplification between the client IO and the disk IO. The cluster IO stats do still look rather low though... John > Got more than 10 client hosts connecting to the cluster, running around 300 > VMs. > > ceph version 12.2.4 > > #ceph -s > > cluster: > id: xxxxxxxxxxxxxxx > health: HEALTH_OK > > services: > mon: 3 daemons, quorum ceph0,ceph1,ceph2 > mgr: ceph1(active), standbys: ceph0, ceph2 > osd: 24 osds: 24 up, 24 in > rgw: 1 daemon active > > data: > pools: 17 pools, 956 pgs > objects: 4225k objects, 14495 GB > usage: 43424 GB used, 16231 GB / 59656 GB avail > pgs: 956 active+clean > > io: > client: 123 kB/s rd, 2677 kB/s wr, 38 op/s rd, 278 op/s wr > > (at one of the node) > #atop > > DSK | sdb | busy 42% | read 268 | write 519 | KiB/w > 109 | MBr/s 2.4 | MBw/s 5.6 | avio 5.26 ms | > DSK | sde | busy 26% | read 129 | write 313 | KiB/w > 150 | MBr/s 0.7 | MBw/s 4.6 | avio 5.94 ms | > DSK | sdg | busy 24% | read 90 | write 230 | KiB/w > 86 | MBr/s 0.5 | MBw/s 1.9 | avio 7.50 ms | > DSK | sdf | busy 21% | read 109 | write 148 | KiB/w > 162 | MBr/s 0.8 | MBw/s 2.3 | avio 8.12 ms | > DSK | sdh | busy 19% | read 100 | write 221 | KiB/w > 118 | MBr/s 0.5 | MBw/s 2.5 | avio 5.78 ms | > DSK | sda | busy 18% | read 170 | write 163 | KiB/w > 83 | MBr/s 1.6 | MBw/s 1.3 | avio 5.35 ms | > DSK | sdc | busy 3% | read 0 | write 1545 | KiB/w > 58 | MBr/s 0.0 | MBw/s 8.8 | avio 0.21 ms | > DSK | sdd | busy 3% | read 0 | write 1195 | KiB/w > 57 | MBr/s 0.0 | MBw/s 6.7 | avio 0.24 ms | > > Regards, > Horace Ng > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com