Hi Eugen,
What a good coincidence ;-)
So I ran "cephadm ceph-volume lvm list" on the OSD node which I re-instaled and
it saw my osd.2 OSD. So far so good, but the following suggested command does not work as
you can see below:
ubuntu@ceph1f:~$ sudo cephadm deploy --name osd.2 --fsid
91a86f20-8083-40b1-8bf1-fe35fac3d677
Deploy daemon osd.2 ...
Traceback (most recent call last):
File "/usr/local/sbin/cephadm", line 6223, in <module>
r = args.func()
File "/usr/local/sbin/cephadm", line 1440, in _default_image
return func()
File "/usr/local/sbin/cephadm", line 3457, in command_deploy
deploy_daemon(args.fsid, daemon_type, daemon_id, c, uid, gid,
File "/usr/local/sbin/cephadm", line 2193, in deploy_daemon
deploy_daemon_units(fsid, uid, gid, daemon_type, daemon_id, c,
File "/usr/local/sbin/cephadm", line 2255, in deploy_daemon_units
assert osd_fsid
AssertionError
Any ideas what is wrong here?
Regards,
Mabi
‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐
On Thursday, May 27, 2021 12:13 PM, Eugen Block <eblock(a)nde.ag> wrote:
Hi,
I posted a link to the docs [1], [2] just yesterday ;-)
You should see the respective OSD in the output of 'cephadm
ceph-volume lvm list' on that node. You should then be able to get it
back to cephadm with
cephadm deploy --name osd.x
But I haven't tried this yet myself, so please report back if that
works for you.
Regards,
Eugen
[1]
https://tracker.ceph.com/issues/49159
[2]
https://tracker.ceph.com/issues/46691
Zitat von mabi mabi(a)protonmail.ch:
Hello,
I have by mistake re-installed the OS of an OSD node of my Octopus
cluster (managed by cephadm). Luckily the OSD data is on a separate
disk and did not get affected by the re-install.
Now I have the following state:
health: HEALTH_WARN
1 stray daemon(s) not managed by cephadm
1 osds down
1 host (1 osds) down
To fix that I tried to run:
ceph orch daemon add osd ceph1f:/dev/sda
=========================================
Created no osd(s) on host ceph1f; already created?
That did not work, so I tried:
ceph cephadm osd activate ceph1f
=================================
no valid command found; 10 closest matches:
...
Error EINVAL: invalid command
Did not work either. So I wanted to ask how can I "adopt" back an
OSD disk to my cluster?
Thanks for your help.
Regards,
Mabi
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io