Re: [ceph-users] (yet another) multi active mds advise needed

2018-05-19 Thread Webert de Souza Lima
Hi Daniel, Thanks for clarifying. I'll have a look at dirfrag option. Regards, Webert Lima Em sáb, 19 de mai de 2018 01:18, Daniel Baumann escreveu: > On 05/19/2018 01:13 AM, Webert de Souza Lima wrote: > > New question: will it make any difference in the balancing if

Re: [ceph-users] Intepreting reason for blocked request

2018-05-19 Thread Bryan Henderson
>>> 2018-05-03 01:56:35.249122 osd.0 192.168.1.16:6800/348 54 : >>> cluster [WRN] slow request 961.557151 seconds old, >>> received at 2018-05-03 01:40:33.689191: >>> pg_query(4.f epoch 490) currently wait for new map >>> > >The OSD is waiting for a new OSD map, which it will get from one

Re: [ceph-users] Multi-MDS Failover

2018-05-19 Thread Blair Bethwaite
On 19 May 2018 at 09:20, Scottix wrote: > It would be nice to have an option to have all IO blocked if it hits a > degraded state until it recovers. Since you are unaware of other MDS state, > seems like that would be tough to do. I agree this would be a nice knob to have

Re: [ceph-users] in retrospect get OSD for "slow requests are blocked" ? / get detailed health status via librados?

2018-05-19 Thread Brad Hubbard
On Sat, May 19, 2018 at 5:01 PM, Uwe Sauter wrote: > The mistery is that these blocked requests occur numerously when at > least > one of the 6 servers is booted with kernel 4.15.17, if all are running > 4.13.16 the number of blocked requests is infrequent

Re: [ceph-users] in retrospect get OSD for "slow requests are blocked" ? / get detailed health status via librados?

2018-05-19 Thread Uwe Sauter
Am 19.05.2018 um 01:45 schrieb Brad Hubbard: On Thu, May 17, 2018 at 6:06 PM, Uwe Sauter wrote: Brad, thanks for the bug report. This is exactly the problem I am having (log-wise). You don't give any indication what version you are running but see