Re: [ceph-users] OSD replacement causes slow requests

2019-07-24 Thread Eugen Block
Hi Wido, thanks for your response. Have you tried to dump the historic slow ops on the OSDs involved to see what is going on? $ ceph daemon osd.X dump_historic_slow_ops Good question, I don't recall doing that. Maybe my colleague did but he's on vacation right now. ;-) But to be clear,

Re: [ceph-users] OSD replacement causes slow requests

2019-07-24 Thread Wido den Hollander
On 7/18/19 12:21 PM, Eugen Block wrote: > Hi list, > > we're facing an unexpected recovery behavior of an upgraded cluster > (Luminous -> Nautilus). > > We added new servers with Nautilus to the existing Luminous cluster, so > we could first replace the MONs step by step. Then we moved the old

[ceph-users] OSD replacement causes slow requests

2019-07-18 Thread Eugen Block
Hi list, we're facing an unexpected recovery behavior of an upgraded cluster (Luminous -> Nautilus). We added new servers with Nautilus to the existing Luminous cluster, so we could first replace the MONs step by step. Then we moved the old servers to a new root in the crush map and then

[ceph-users] OSD replacement

2015-04-14 Thread Corey Kovacs
I am fairly new to ceph and so far things are going great. That said, when I try to replace a failed OSD, I can't seem to get it to use the same OSD id#. I have gotten it to point which a ceph osd create does use the correct id# but when I try to use ceph-deploy to instantiate the replacement, I

Re: [ceph-users] OSD replacement

2015-04-14 Thread Corey Kovacs
Vikhyat, I went through the steps as I did yesterday but with one small change. I was putting the --zap-disk option before the host:disk option. Now it works as expected. I'll try it again with the wrong syntax to see if that's really the problem but it's the only difference between working and

Re: [ceph-users] OSD replacement

2015-04-14 Thread Vikhyat Umrao
Hi, I hope you are following this : http://ceph.com/docs/master/rados/operations/add-or-rm-osds/#removing-osds-manual After removing the osd successfully run the following command : # ceph-deploy --overwrite-conf osd create osd-host:device-path --zap-disk It will give you the same osd id