Re: [ceph-users] slow requests on a specific osd

2018-01-15 Thread lists
Hi Wes, On 15-1-2018 20:57, Wes Dillingham wrote: My understanding is that the exact same objects would move back to the OSD if weight went 1 -> 0 -> 1 given the same Cluster state and same object names, CRUSH is deterministic so that would be the almost certain result. Ok, thanks! So this

Re: [ceph-users] slow requests on a specific osd

2018-01-15 Thread Wes Dillingham
My understanding is that the exact same objects would move back to the OSD if weight went 1 -> 0 -> 1 given the same Cluster state and same object names, CRUSH is deterministic so that would be the almost certain result. On Mon, Jan 15, 2018 at 2:46 PM, lists wrote: > Hi Wes, > > On 15-1-2018 20

Re: [ceph-users] slow requests on a specific osd

2018-01-15 Thread lists
Hi Wes, On 15-1-2018 20:32, Wes Dillingham wrote: I dont hear a lot of people discuss using xfs_fsr on OSDs and going over the mailing list history it seems to have been brought up very infrequently and never as a suggestion for regular maintenance. Perhaps its not needed. True, it's just some

Re: [ceph-users] slow requests on a specific osd

2018-01-15 Thread Wes Dillingham
I dont hear a lot of people discuss using xfs_fsr on OSDs and going over the mailing list history it seems to have been brought up very infrequently and never as a suggestion for regular maintenance. Perhaps its not needed. One thing to consider trying, and to rule out something funky with the XFS

[ceph-users] slow requests on a specific osd

2018-01-15 Thread lists
Hi, On our three-node 24 OSDs ceph 10.2.10 cluster, we have started seeing slow requests on a specific OSD, during the the two-hour nightly xfs_fsr run from 05:00 - 07:00. This started after we applied the meltdown patches. The specific osd.10 also has the highest space utilization of all OSD