ubuntu@ceph1f:~$ sudo cephadm deploy --name osd.2
--fsid
91a86f20-8083-40b1-8bf1-fe35fac3d677
Deploy daemon osd.2 ...
Which fsid is it, the cluster's or the OSD's? According to the
'cephadm deploy' help page it should be the cluster fsid.
Zitat von mabi <mabi(a)protonmail.ch>ch>:
> Hi Eugen,
>
> What a good coincidence ;-)
>
> So I ran "cephadm ceph-volume lvm list" on the OSD node which I
> re-instaled and it saw my osd.2 OSD. So far so good, but the
> following suggested command does not work as you can see below:
>
ubuntu@ceph1f:~$ sudo cephadm deploy --name osd.2
--fsid
91a86f20-8083-40b1-8bf1-fe35fac3d677
Deploy daemon osd.2 ...
>
> Traceback (most recent call last):
> File "/usr/local/sbin/cephadm", line 6223, in <module>
> r = args.func()
> File "/usr/local/sbin/cephadm", line 1440, in _default_image
> return func()
> File "/usr/local/sbin/cephadm", line 3457, in command_deploy
> deploy_daemon(args.fsid, daemon_type, daemon_id, c, uid, gid,
> File "/usr/local/sbin/cephadm", line 2193, in deploy_daemon
> deploy_daemon_units(fsid, uid, gid, daemon_type, daemon_id, c,
> File "/usr/local/sbin/cephadm", line 2255, in deploy_daemon_units
> assert osd_fsid
> AssertionError
>
> Any ideas what is wrong here?
>
> Regards,
> Mabi
>
> ‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐
> On Thursday, May 27, 2021 12:13 PM, Eugen Block <eblock(a)nde.ag> wrote:
>
>> Hi,
>>
>> I posted a link to the docs [1], [2] just yesterday ;-)
>>
>> You should see the respective OSD in the output of 'cephadm
>> ceph-volume lvm list' on that node. You should then be able to get it
>> back to cephadm with
>>
>> cephadm deploy --name osd.x
>>
>> But I haven't tried this yet myself, so please report back if that
>> works for you.
>>
>> Regards,
>> Eugen
>>
>> [1]
https://tracker.ceph.com/issues/49159
>> [2]
https://tracker.ceph.com/issues/46691
>>
>> Zitat von mabi mabi(a)protonmail.ch:
>>
>> > Hello,
>> > I have by mistake re-installed the OS of an OSD node of my Octopus
>> > cluster (managed by cephadm). Luckily the OSD data is on a separate
>> > disk and did not get affected by the re-install.
>> > Now I have the following state:
>> >
>> > health: HEALTH_WARN
>> > 1 stray daemon(s) not managed by cephadm
>> > 1 osds down
>> > 1 host (1 osds) down
>> >
>> >
>> > To fix that I tried to run:
>> >
>> > ceph orch daemon add osd ceph1f:/dev/sda
>> >
>> > =========================================
>> >
>> > Created no osd(s) on host ceph1f; already created?
>> > That did not work, so I tried:
>> >
>> > ceph cephadm osd activate ceph1f
>> >
>> > =================================
>> >
>> > no valid command found; 10 closest matches:
>> > ...
>> > Error EINVAL: invalid command
>> > Did not work either. So I wanted to ask how can I "adopt" back an
>> > OSD disk to my cluster?
>> > Thanks for your help.
>> > Regards,
>> > Mabi
>> >
>> > ceph-users mailing list -- ceph-users(a)ceph.io
>> > To unsubscribe send an email to ceph-users-leave(a)ceph.io
>>
>> ceph-users mailing list -- ceph-users(a)ceph.io
>> To unsubscribe send an email to ceph-users-leave(a)ceph.io