On Sun, Dec 11, 2016 at 4:38 PM, Mike Miller <millermike...@gmail.com> wrote:
> Hi,
>
> you have given up too early. rsync is not a nice workload for cephfs, in
> particular, most linux kernel clients cephfs will end up caching all
> inodes/dentries. The result is that mds servers crash due to memory
> limitations. And rsync basically screens all inodes/dentries so it is the
> perfect application to gobble up all inode caps.

While historically there have been client bugs that prevented the MDS
from enforcing cache size limits, this is not expected behaviour --
manually calling drop_caches is most definitely a workaround and not
something that I would recommend unless you're stuck with a
known-buggy client version for some reason.

Just felt the need to point that out in case people started picking
this up as a best practice!

Cheers,
John

> We run a cronjob script flush_cache every few (2-5) minutes:
>
> #!/bin/bash
> echo 2 > /proc/sys/vm/drop_caches
>
> on all machines that mount cephfs. There is no performance drop in the
> client machines, but happily, the mds congestion is solved by this.
>
> We also went the rbd way before this, but for large rbd images we much
> prefer cephfs instead.
>
> Regards,
>
> Mike
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to