Now that’s a *very* different question from numbers assigned during an install.
With recent releases instead of going down the full removal litany listed below, you can
instead down/out the OSD and `destroy` it. That preserves the CRUSH bucket and OSD ID,
then when you use ceph-disk, ceph-volume, what-have-you to deploy a replacement, you can
specify the same and desired OSD ID on the commandline.
Note that as of 12.2.2, you’ll want to record and re-set any override reweight (manual or
reweight-by-utilization) as that usually or never survives. Also note that again as of
that release, if the replacement drive is a different size, the CRUSH weight is not
adjusted, so you may (or may not) want to adjust the CRUSH weight. Slight differences
aren’t usually a huge deal; big differences can mean you have unused capacity, or
overloaded drives.
— Anthony
Thank you for your answer below.
I'm not looking to reuse them as much as I am trying to control what unused number is
actually used.
For example if I have 20 osds and 2 have failed...when I replace a disk in one server I
don't want it to automatically use the next lowest number for the osd assignment.
I understand what you mean about not focusing on the osd ids...but my ocd is making me
ask the question.
Thanks,
Shain
On 9/11/20, 9:45 AM, "George Shuklin" <george.shuklin(a)gmail.com> wrote:
On 11/09/2020 16:11, Shain Miley wrote:
Hello,
I have been wondering for quite some time whether or not it is possible to influence the
osd.id numbers that are assigned during an install.
I have made an attempt to keep our osds in order over the last few years, but it is a
losing battle without having some control over the osd assignment.
I am currently using ceph-deploy to handle adding nodes to the cluster.
You can reuse osd numbers, but I strongly advice you not to focus on
precise IDs. The reason is that you can have such combination of server
faults, which will swap IDs no matter what.
It's a false sense of beauty to have 'ID of OSD match ID in the name of
the server'.
How to reuse osd nums?
OSD number is used (and should be cleaned if OSD dies) in three places
in Ceph:
1) Crush map: ceph osd crush rm osd.x
2) osd list: ceph osd rm osd.x
3) auth: ceph auth rm osd.x
The last one is often forgoten and is a usual reason for ceph-ansible to
fail on new disk in the server.
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io