You are right, I used the FSID of the OSD and not of the cluster in the deploy command. So
now I tried again with the cluster ID as FSID but still it does not work as you can see
below:
ubuntu@ceph1f:~$ sudo cephadm deploy --name osd.2 --fsid
8d47792c-987d-11eb-9bb6-a5302e00e1fa
Deploy daemon osd.2 ...
Traceback (most recent call last):
File "/usr/local/sbin/cephadm", line 6223, in <module>
r = args.func()
File "/usr/local/sbin/cephadm", line 1440, in _default_image
return func()
File "/usr/local/sbin/cephadm", line 3457, in command_deploy
deploy_daemon(args.fsid, daemon_type, daemon_id, c, uid, gid,
File "/usr/local/sbin/cephadm", line 2193, in deploy_daemon
deploy_daemon_units(fsid, uid, gid, daemon_type, daemon_id, c,
File "/usr/local/sbin/cephadm", line 2255, in deploy_daemon_units
assert osd_fsid
AssertionError
In case that's of any help here is the output of the "cephadm ceph-volume lvm
list" command:
====== osd.2 =======
[block]
/dev/ceph-cca8abe6-cf9b-4c2f-ab81-ae0758585414/osd-block-91a86f20-8083-40b1-8bf1-fe35fac3d677
block device
/dev/ceph-cca8abe6-cf9b-4c2f-ab81-ae0758585414/osd-block-91a86f20-8083-40b1-8bf1-fe35fac3d677
block uuid W3omTg-vami-RB0V-CkVb-cgpb-88Jy-pIK2Tz
cephx lockbox secret
cluster fsid 8d47792c-987d-11eb-9bb6-a5302e00e1fa
cluster name ceph
crush device class None
encrypted 0
osd fsid 91a86f20-8083-40b1-8bf1-fe35fac3d677
osd id 2
osdspec affinity all-available-devices
type block
vdo 0
devices /dev/sda
‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐
On Thursday, May 27, 2021 12:32 PM, Eugen Block <eblock(a)nde.ag> wrote:
ubuntu@ceph1f:~$ sudo cephadm deploy --name osd.2 --fsid
91a86f20-8083-40b1-8bf1-fe35fac3d677
Deploy daemon osd.2 ...
Which fsid is it, the cluster's or the OSD's? According to the
'cephadm deploy' help page it should be the cluster fsid.
Zitat von mabi mabi(a)protonmail.ch:
> Hi Eugen,
> What a good coincidence ;-)
> So I ran "cephadm ceph-volume lvm list" on the OSD node which I
> re-instaled and it saw my osd.2 OSD. So far so good, but the
> following suggested command does not work as you can see below:
ubuntu@ceph1f:~$ sudo cephadm deploy --name osd.2
--fsid
91a86f20-8083-40b1-8bf1-fe35fac3d677
Deploy daemon osd.2 ...
> Traceback (most recent call last):
> File "/usr/local/sbin/cephadm", line 6223, in <module>
> r = args.func()
> File "/usr/local/sbin/cephadm", line 1440, in _default_image
> return func()
> File "/usr/local/sbin/cephadm", line 3457, in command_deploy
> deploy_daemon(args.fsid, daemon_type, daemon_id, c, uid, gid,
> File "/usr/local/sbin/cephadm", line 2193, in deploy_daemon
> deploy_daemon_units(fsid, uid, gid, daemon_type, daemon_id, c,
> File "/usr/local/sbin/cephadm", line 2255, in deploy_daemon_units
> assert osd_fsid
> AssertionError
> Any ideas what is wrong here?
> Regards,
> Mabi
> ‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐
> On Thursday, May 27, 2021 12:13 PM, Eugen Block eblock(a)nde.ag wrote:
>
> > Hi,
> > I posted a link to the docs [1], [2] just yesterday ;-)
> > You should see the respective OSD in the output of 'cephadm
> > ceph-volume lvm list' on that node. You should then be able to get it
> > back to cephadm with
> > cephadm deploy --name osd.x
> > But I haven't tried this yet myself, so please report back if that
> > works for you.
> > Regards,
> > Eugen
> > [1]
https://tracker.ceph.com/issues/49159
> > [2]
https://tracker.ceph.com/issues/46691
> > Zitat von mabi mabi(a)protonmail.ch:
> >
> > > Hello,
> > > I have by mistake re-installed the OS of an OSD node of my Octopus
> > > cluster (managed by cephadm). Luckily the OSD data is on a separate
> > > disk and did not get affected by the re-install.
> > > Now I have the following state:
> > >
> > > health: HEALTH_WARN
> > > 1 stray daemon(s) not managed by cephadm
> > > 1 osds down
> > > 1 host (1 osds) down
> > >
> > >
> > > To fix that I tried to run:
> > > ceph orch daemon add osd ceph1f:/dev/sda
> > > =========================================
> > > Created no osd(s) on host ceph1f; already created?
> > > That did not work, so I tried:
> > > ceph cephadm osd activate ceph1f
> > > =================================
> > > no valid command found; 10 closest matches:
> > > ...
> > > Error EINVAL: invalid command
> > > Did not work either. So I wanted to ask how can I "adopt" back
an
> > > OSD disk to my cluster?
> > > Thanks for your help.
> > > Regards,
> > > Mabi
> > > ceph-users mailing list -- ceph-users(a)ceph.io
> > > To unsubscribe send an email to ceph-users-leave(a)ceph.io
> >
> > ceph-users mailing list -- ceph-users(a)ceph.io
> > To unsubscribe send an email to ceph-users-leave(a)ceph.io
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io