Yes I was also thinking of setting this configuration. But I am just
wondering what are the drawbacks, if there are any? Why is the ceph team
not making this a default setting if it is so good?
-----Original Message-----
To: ceph-users@ceph.io
Subject: [ceph-users] Re: ceph mds slow requests
Hi,
although I've read many threads mentioning these two osd config options
I didn't know what to expect of it. But since you explicitly referred
the slow requests I decided to give it a try and changed 'osd op queue
cut off' to "high" ('osd op queue' was already "wpq"). We've had two
deep-scrub cycles since then, half of our PGs have been deep-scrubbed
and I haven't seen a single slow request.
So thanks again for pointing that out!
Regards,
Eugen
Zitat von Andrej Filipcic <andrej.filip...@ijs.si>:
> Hi,
>
> all our slow request issues were solved with:
> [osd]
> osd op queue = wpq
> osd op queue cut off = high
>
> before we even had several hours old request, since the change it
> rarely gets above 30s even with the heaviest loads, eg >100 iops/hdd
>
> Regards,
> Andrej
>
> On 6/10/20 12:12 PM, Eugen Block wrote:
>> Hi,
>>
>> we have this message almost daily, although in our case it's almost
>> expected. We run a nightly compile job within a cephfs subtree and
>> the OSDs (HDD with rocksDB on SSD) are saturated during those jobs.
>> Also the deep-scrubs which also run during the night have a
>> significant impact and the cluster reports slow requests, but since
>> that happens outside our working hours we can live with it (for now).
>>
>> You write the OSDs are on SSDs, is that true for both data and
>> metadata pool?
>>
>> Regards,
>> Eugen
>>
>>
>> Zitat von locallocal <locallo...@163.com>:
>>
>>> hi,guys.
>>> we have a ceph cluster which version is luminous 12.2.13. and
>>> Recently we encountered a problem.here are some log infomations:
>>>
>>>
>>> 2020-06-08 12:33:52.706070 7f4097e2d700 0 log_channel(cluster) log
>>> [WRN] : slow request 30.518930 seconds old, received at
>>> 2020-06-08 12:33:22.186924:
>>> client_request(client.48978906:941633993 create
>>> #0x100028cab8a/.filename 2020-06-08 12:33:22.197434 caller_uid=0,
>>> caller_gid=0{}) currently submit entry: journal_and_reply ...
>>> 2020-06-08 13:12:17.826727 7f4097e2d700 0 log_channel(cluster) log
>>> [WRN] : slow request 2220.991833 seconds old, received at
>>> 2020-06-08 12:35:16.764233:
>>> client_request(client.42390705:788369155 create
>>> #0x1000224f999/.filename 2020-06-08 12:35:16.774553 caller_uid=0,
>>> caller_gid=0{}) currently submit entry: journal_and_reply
>>>
>>>
>>> it looks like mds can't flush journal to osd of meta pool.but the
>>> osd type is ssd and the load is very low.this problem leads the
>>> client can't mount and the mds can't trim log.
>>> Is there anyone have encountered this problem.Please help!
>>>
>>> _______________________________________________
>>> ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an
>>> email to ceph-users-le...@ceph.io
>>
>>
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an
>> email to ceph-users-le...@ceph.io
>
>
> --
> _____________________________________________________________
> prof. dr. Andrej Filipcic, E-mail: andrej.filip...@ijs.si
> Department of Experimental High Energy Physics - F9
> Jozef Stefan Institute, Jamova 39, P.o.Box 3000
> SI-1001 Ljubljana, Slovenia
> Tel.: +386-1-477-3674 Fax: +386-1-477-3166
> -------------------------------------------------------------
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an
> email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an
email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io