Now that’s a *very* different question from numbers assigned during an install.

With recent releases instead of going down the full removal litany listed 
below, you can instead down/out the OSD and `destroy` it.  That preserves the 
CRUSH bucket and OSD ID, then when you use ceph-disk, ceph-volume, 
what-have-you to deploy a replacement, you can specify the same and desired OSD 
ID on the commandline.

Note that as of 12.2.2, you’ll want to record and re-set any override reweight 
(manual or reweight-by-utilization) as that usually or never survives.  Also 
note that again as of that release, if the replacement drive is a different 
size, the CRUSH weight is not adjusted, so you may (or may not) want to adjust 
the CRUSH weight.  Slight differences aren’t usually a huge deal; big 
differences can mean you have unused capacity, or overloaded drives.

— Anthony

> 
> Thank you for your answer below.
> 
> I'm not looking to reuse them as much as I am trying to control what unused 
> number is actually used.
> 
> For example if I have 20 osds and 2 have failed...when I replace a disk in 
> one server I don't want it to automatically use the next lowest number for 
> the osd assignment.
> 
> I understand what you mean about not focusing on the osd ids...but my ocd is 
> making me ask the question.
> 
> Thanks,
> Shain
> 
> On 9/11/20, 9:45 AM, "George Shuklin" <george.shuk...@gmail.com> wrote:
> 
>    On 11/09/2020 16:11, Shain Miley wrote:
>> Hello,
>> I have been wondering for quite some time whether or not it is possible to 
>> influence the osd.id numbers that are  assigned during an install.
>> 
>> I have made an attempt to keep our osds in order over the last few years, 
>> but it is a losing battle without having some control over the osd 
>> assignment.
>> 
>> I am currently using ceph-deploy to handle adding nodes to the cluster.
>> 
>    You can reuse osd numbers, but I strongly advice you not to focus on 
>    precise IDs. The reason is that you can have such combination of server 
>    faults, which will swap IDs no matter what.
> 
>    It's a false sense of beauty to have 'ID of OSD match ID in the name of 
>    the server'.
> 
>    How to reuse osd nums?
> 
>    OSD number is used (and should be cleaned if OSD dies) in three places 
>    in Ceph:
> 
>    1) Crush map: ceph osd crush rm osd.x
> 
>    2) osd list: ceph osd rm osd.x
> 
>    3) auth: ceph auth rm osd.x
> 
>    The last one is often forgoten and is a usual reason for ceph-ansible to 
>    fail on new disk in the server.
>    _______________________________________________
>    ceph-users mailing list -- ceph-users@ceph.io
>    To unsubscribe send an email to ceph-users-le...@ceph.io
> 
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to