I indeed have very small osds, but shortly I will be able to test ceph on
much larger osds.
However, looking in syslog I found this
Oct 23 23:45:14 ceph0 bash[2265]: debug 2020-10-23T21:45:14.918+0000
7fe8de931700 -1 mgr load Failed to construct class in 'cephadm'
Oct 23 23:45:14 ceph0 bash[2265]: debug 2020-10-23T21:45:14.918+0000
7fe8de931700 -1 mgr load Traceback (most recent call last):
Oct 23 23:45:14 ceph0 bash[2265]: File
"/usr/share/ceph/mgr/cephadm/module.py", line 325, in __init__
Oct 23 23:45:14 ceph0 bash[2265]: self.rm_util.load_from_store()
Oct 23 23:45:14 ceph0 bash[2265]: File
"/usr/share/ceph/mgr/cephadm/services/osd.py", line 465, in load_from_store
Oct 23 23:45:14 ceph0 bash[2265]: osd_obj =
OSD.from_json(json.loads(osd), ctx=self)
Oct 23 23:45:14 ceph0 bash[2265]: File
"/lib64/python3.6/json/__init__.py", line 348, in loads
Oct 23 23:45:14 ceph0 bash[2265]: 'not
{!r}'.format(s.__class__.__name__))
Oct 23 23:45:14 ceph0 bash[2265]: TypeError: the JSON object must be str,
bytes or bytearray, not 'dict'
Oct 23 23:45:14 ceph0 bash[2265]: debug 2020-10-23T21:45:14.918+0000
7fe8de931700 -1 mgr operator() Failed to run module in active mode
('cephadm')
Oct 23 23:45:14 ceph0 bash[2265]: debug 2020-10-23T21:45:14.922+0000
7fe8de931700 0 [crash DEBUG root] setting log level based on debug_mgr:
WARNING (1/5)
[...]
Oct 23 23:45:33 ceph0 bash[2265]: debug 2020-10-23T21:45:33.477+0000
7fe8cf913700 -1 no module 'cephadm'
Before this, cephadm seemed to be operating normally.
Attached is a larger portion of the relevant log.
Il giorno lun 26 ott 2020 alle ore 09:03 Eugen Block <eblock(a)nde.ag> ha
scritto:
Interesting, what do you see in the MGR logs, there
should be
something in there.
Zitat von Marco Venuti <afm.itunev(a)gmail.com>om>:
Yes, this is the status
# ceph -s
cluster:
id: ab471d92-14a2-11eb-ad67-525400bbdc0d
health: HEALTH_OK
services:
mon: 5 daemons, quorum ceph0.starfleet.sns.it
,ceph1,ceph3,ceph5,ceph4
(age 104m)
mgr: ceph1.jxmtpn(active, since 17m), standbys:
ceph0.starfleet.sns.it.clzhjp
mds: starfs:1 {0=starfs.ceph4.kqwkdc=up:active} 1 up:standby
osd: 12 osds: 10 up (since 103m), 10 in (since 2d)
task status:
scrub status:
mds.starfs.ceph4.kqwkdc: idle
data:
pools: 4 pools, 97 pgs
objects: 10.95k objects, 3.6 GiB
usage: 23 GiB used, 39 GiB / 62 GiB avail
pgs: 97 active+clean
Il giorno dom 25 ott 2020 alle ore 21:02 Eugen Block <eblock(a)nde.ag> ha
scritto:
> Is one of the MGRs up? What is the ceph status?
>
>
> Zitat von Marco Venuti <afm.itunev(a)gmail.com>om>:
>
> > Hi,
> > I'm experimenting ceph on a (small) test cluster. I'm using version
> 15.2.5
> > deployed with cephadm.
> > I was trying to do some "disaster" testing, such as wiping a disk in
> order
> > to simulate a hardware failure, destroy the osd and recreate it, all
of
> > which I managed to do successfully.
> > However, a few hours after this test, the orchestrator failed with no
> > apparent reason. I tried to disable and reenable cephadm, but with no
> luck:
> >
> > # ceph orch ls
> > Error ENOENT: No orchestrator configured (try `ceph orch set backend`)
> > # ceph orch set backend cephadm
> > Error ENOENT: Module not found
> >
> > What could have happened? Is there some way to reenable cephadm?
> >
> > Thanks,
> > Marco
> > _______________________________________________
> > ceph-users mailing list -- ceph-users(a)ceph.io
> > To unsubscribe send an email to ceph-users-leave(a)ceph.io
>
>
>
>