if you do a mgr failover ("ceph mgr fail") and wait a few minutes do the
issues clear out? I know there's a bug where removed mons get marked as
stray daemons while downsizing by multiple mons at once (cephadm might be
removing them too quickly, not totally sure of the cause) but doing a mgr
failover has always cleared the stray daemon notifications for me. For some
context, what it's listing as stray daemons are roughly what is being
reported in "ceph node ls" that doesn't show up in "ceph orch ps".
The idea
being the orch ps output shows all the daemons cephadm is aware of and
managing while "ceph node ls" are ceph daemons the cluster, but not
necessarily cephadm itself, is aware of. For me, the mon daemons marked
stray were still showing up in that "ceph node ls" output, but doing a mgr
failover would clean that up and then the stray daemon warnings would also
disappear.
On Mon, Jun 12, 2023 at 8:54 AM farhad kh <farhad.khedriyan(a)gmail.com>
wrote:
i deployed the ceph cluster with 8 node (v17.2.6) and
after add all of
hosts, ceph create 5 mon daemon instances
i try decrease that to 3 instance with ` ceph orch apply mon
--placement=label:mon,count:3 it worked, but after that i get error "2
stray daemons not managed by cephadm" .
But every time I tried to deploy and delete other instances, this number
increased Now I have 7 daemon that are not managed by cephadm
How to deal with this issue?
------------------------
[root@opcsdfpsbpp0201 ~]# ceph -s
cluster:
id: 79a2627c-0821-11ee-a494-00505695c58c
health: HEALTH_WARN
16 stray daemon(s) not managed by cephadm
services:
mon: 3 daemons, quorum opcsdfpsbpp0201,opcsdfpsbpp0205,opcsdfpsbpp0203
(age 2m)
mgr: opcsdfpsbpp0201.vttwxa(active, since 27h), standbys:
opcsdfpsbpp0207.kzxepm
mds: 1/1 daemons up, 2 standby
osd: 74 osds: 74 up (since 26h), 74 in (since 26h)
data:
volumes: 1/1 healthy
pools: 6 pools, 6 pgs
objects: 2.10k objects, 8.1 GiB
usage: 28 GiB used, 148 TiB / 148 TiB avail
pgs: 6 active+clean
io:
client: 426 B/s rd, 0 op/s rd, 0 op/s wr
[root@opcsdfpsbpp0201 ~]# ceph health detail
HEALTH_WARN 16 stray daemon(s) not managed by cephadm
[WRN] CEPHADM_STRAY_DAEMON: 16 stray daemon(s) not managed by cephadm
stray daemon mon.opcsdfpsbpp0207 on host opcsdfpsbpp0203 not managed by
cephadm
stray daemon mon.opcsdfpsbpp0209 on host opcsdfpsbpp0203 not managed by
cephadm
stray daemon mon.opcsdfpsbpp0211 on host opcsdfpsbpp0203 not managed by
cephadm
stray daemon mon.opcsdfpsbpp0213 on host opcsdfpsbpp0203 not managed by
cephadm
stray daemon mon.opcsdfpsbpp0207 on host opcsdfpsbpp0205 not managed by
cephadm
stray daemon mon.opcsdfpsbpp0209 on host opcsdfpsbpp0205 not managed by
cephadm
stray daemon mon.opcsdfpsbpp0211 on host opcsdfpsbpp0205 not managed by
cephadm
stray daemon mon.opcsdfpsbpp0213 on host opcsdfpsbpp0205 not managed by
cephadm
stray daemon mon.opcsdfpsbpp0213 on host opcsdfpsbpp0207 not managed by
cephadm
stray daemon mon.opcsdfpsbpp0207 on host opcsdfpsbpp0209 not managed by
cephadm
stray daemon mon.opcsdfpsbpp0209 on host opcsdfpsbpp0209 not managed by
cephadm
stray daemon mon.opcsdfpsbpp0209 on host opcsdfpsbpp0211 not managed by
cephadm
stray daemon mon.opcsdfpsbpp0215 on host opcsdfpsbpp0211 not managed by
cephadm
stray daemon mon.opcsdfpsbpp0211 on host opcsdfpsbpp0213 not managed by
cephadm
stray daemon mon.opcsdfpsbpp0209 on host opcsdfpsbpp0215 not managed by
cephadm
stray daemon mon.opcsdfpsbpp0213 on host opcsdfpsbpp0215 not managed by
cephadm
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io