Hi, Eugen,
I set mds_cache_memory_limit from 1G to 2G, and then to 4G. I still get the
warning
messages, and the messages would disappear in 1 or 2 minutes.
Which version do your kernels use?

Zhenshi Zhou <deader...@gmail.com> 于2018年8月13日周一 下午10:15写道:

> Hi Eugen,
> The command shows "mds_cache_memory_limit": "1073741824".
> And I'll increase the cache size for a try.
>
> Thanks
>
> Eugen Block <ebl...@nde.ag> 于2018年8月13日周一 下午9:48写道:
>
>> Hi,
>>
>> Depending on your kernel (memory leaks with CephFS) increasing the
>> mds_cache_memory_limit could be of help. What is your current setting
>> now?
>>
>> ceph:~ # ceph daemon mds.<MDS> config show | grep mds_cache_memory_limit
>>
>> We had these messages for months, almost every day.
>> It would occur when hourly backup jobs ran and the MDS had to serve an
>> additional client (searching the whole CephFS for changes) besides the
>> existing CephFS clients. First we updated all clients to a more recent
>> kernel version, but the warnings didn't stop. Then we doubled the
>> cache size from 2 GB to 4 GB last week and since then I haven't seen
>> this warning again (for now).
>>
>> Try playing with the cache size to find a setting fitting your needs,
>> but don't forget to monitor your MDS in case something goes wrong.
>>
>> Regards,
>> Eugen
>>
>>
>> Zitat von Wido den Hollander <w...@42on.com>:
>>
>> > On 08/13/2018 01:22 PM, Zhenshi Zhou wrote:
>> >> Hi,
>> >> Recently, the cluster runs healthy, but I get warning messages
>> everyday:
>> >>
>> >
>> > Which version of Ceph? Which version of clients?
>> >
>> > Can you post:
>> >
>> > $ ceph versions
>> > $ ceph features
>> > $ ceph fs status
>> >
>> > Wido
>> >
>> >> 2018-08-13 17:39:23.682213 [INF]  Cluster is now healthy
>> >> 2018-08-13 17:39:23.682144 [INF]  Health check cleared:
>> >> MDS_CLIENT_RECALL (was: 6 clients failing to respond to cache pressure)
>> >> 2018-08-13 17:39:23.052022 [INF]  MDS health message cleared (mds.0):
>> >> Client docker38:docker failing to respond to cache pressure
>> >> 2018-08-13 17:39:23.051979 [INF]  MDS health message cleared (mds.0):
>> >> Client docker73:docker failing to respond to cache pressure
>> >> 2018-08-13 17:39:23.051934 [INF]  MDS health message cleared (mds.0):
>> >> Client docker74:docker failing to respond to cache pressure
>> >> 2018-08-13 17:39:23.051853 [INF]  MDS health message cleared (mds.0):
>> >> Client docker75:docker failing to respond to cache pressure
>> >> 2018-08-13 17:39:23.051815 [INF]  MDS health message cleared (mds.0):
>> >> Client docker27:docker failing to respond to cache pressure
>> >> 2018-08-13 17:39:23.051753 [INF]  MDS health message cleared (mds.0):
>> >> Client docker27 failing to respond to cache pressure
>> >> 2018-08-13 17:38:11.100331 [WRN]  Health check update: 6 clients
>> failing
>> >> to respond to cache pressure (MDS_CLIENT_RECALL)
>> >> 2018-08-13 17:37:39.570014 [WRN]  Health check update: 5 clients
>> failing
>> >> to respond to cache pressure (MDS_CLIENT_RECALL)
>> >> 2018-08-13 17:37:31.099418 [WRN]  Health check update: 3 clients
>> failing
>> >> to respond to cache pressure (MDS_CLIENT_RECALL)
>> >> 2018-08-13 17:36:34.564345 [WRN]  Health check update: 1 clients
>> failing
>> >> to respond to cache pressure (MDS_CLIENT_RECALL)
>> >> 2018-08-13 17:36:27.121891 [WRN]  Health check update: 3 clients
>> failing
>> >> to respond to cache pressure (MDS_CLIENT_RECALL)
>> >> 2018-08-13 17:36:11.967531 [WRN]  Health check update: 5 clients
>> failing
>> >> to respond to cache pressure (MDS_CLIENT_RECALL)
>> >> 2018-08-13 17:35:59.870055 [WRN]  Health check update: 6 clients
>> failing
>> >> to respond to cache pressure (MDS_CLIENT_RECALL)
>> >> 2018-08-13 17:35:47.787323 [WRN]  Health check update: 3 clients
>> failing
>> >> to respond to cache pressure (MDS_CLIENT_RECALL)
>> >> 2018-08-13 17:34:59.435933 [WRN]  Health check failed: 1 clients
>> failing
>> >> to respond to cache pressure (MDS_CLIENT_RECALL)
>> >> 2018-08-13 17:34:59.045510 [WRN]  MDS health message (mds.0): Client
>> >> docker75:docker failing to respond to cache pressure
>> >>
>> >> How can I fix it?
>> >>
>> >>
>> >> _______________________________________________
>> >> ceph-users mailing list
>> >> ceph-users@lists.ceph.com
>> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >>
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@lists.ceph.com
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to