On Mon, Jan 21, 2019 at 11:16 AM Albert Yue <transuranium....@gmail.com> wrote: > > Dear Ceph Users, > > We have set up a cephFS cluster with 6 osd machines, each with 16 8TB > harddisk. Ceph version is luminous 12.2.5. We created one data pool with > these hard disks and created another meta data pool with 3 ssd. We created a > MDS with 65GB cache size. > > But our users are keep complaining that cephFS is too slow. What we observed > is cephFS is fast when we switch to a new MDS instance, once the cache fills > up (which will happen very fast), client became very slow when performing > some basic filesystem operation such as `ls`. >
what's your mds cache config ? > What we know is our user are putting lots of small files into the cephFS, now > there are around 560 Million files. We didn't see high CPU wait on MDS > instance and meta data pool just used around 200MB space. It's unlikely. For output of 'ceph osd df', you should take both both DATA and OMAP into account. > > My question is, what is the relationship between the metadata pool and MDS? > Is this performance issue caused by the hardware behind meta data pool? Why > the meta data pool only used 200MB space, and we saw 3k iops on each of these > three ssds, why can't MDS cache all these 200MB into memory? > > Thanks very much! > > > Best Regards, > > Albert > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com