"reweight 0" and "out" are the exact same thing
Paul
--
Paul Emmerich
Looking for help with your Ceph cluster? Contact us at
croit GmbH
Freseniusstr. 31h
81247 München
Tel: +49 89 1896585 90
On Tue, Jun 2, 2020 at 9:30 AM Wido den Hollander <wido(a)42on.com> wrote:
On 6/2/20 5:44 AM, Brent Kennedy wrote:
We are rebuilding servers and before luminous our
process was:
1. Reweight the OSD to 0
2. Wait for rebalance to complete
3. Out the osd
4. Crush remove osd
5. Auth del osd
6. Ceph osd rm #
Seems the luminous documentation says that you should:
1. Out the osd
2. Wait for the cluster rebalance to finish
3. Stop the osd
4. Osd purge #
Is reweighting to 0 no longer suggested?
Side note: I tried our existing process and even after reweight, the
entire
cluster restarted the balance again after step 4
( crush remove osd ) of
the
old process. I should also note, by reweighting
to 0, when I tried to
run
"ceph osd out #", it said it was
already marked out.
I assume the docs are correct, but just want to make sure since
reweighting
had been previously recommended.
The new commands just make it more simple. There are many ways to
accomplish the same goal, but what the docs describe should work in most
scenarios.
Wido
Regards,
-Brent
Existing Clusters:
Test: Nautilus 14.2.2 with 3 osd servers, 1 mon/man, 1 gateway, 2 iscsi
gateways ( all virtual on nvme )
US Production(HDD): Nautilus 14.2.2 with 11 osd servers, 3 mons, 4
gateways,
2 iscsi gateways
UK Production(HDD): Nautilus 14.2.2 with 12 osd servers, 3 mons, 4
gateways
US Production(SSD): Nautilus 14.2.2 with 6 osd servers, 3 mons, 3
gateways,
2 iscsi gateways
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io