Hello,
i get:
[WRN] CEPHADM_HOST_CHECK_FAILED: 6 hosts fail cephadm check
host ceph01 failed check: Failed to connect to ceph01 (ceph01).
Check that the host is reachable and accepts connections using the
cephadm SSH key
you may want to run:
ssh -F =(ceph cephadm get-ssh-config) -i =(ceph
config-key get mgr/cephadm/ssh_identity_key) root@ceph01
host ceph02 failed
check: Failed to connect to ceph02 (10.10.1.2).
Check that the host is reachable and accepts connections using the
cephadm SSH key
you may want to run:
ssh -F =(ceph cephadm get-ssh-config) -i =(ceph
config-key get mgr/cephadm/ssh_identity_key) root@ceph02
host ceph03 failed
check: Failed to connect to ceph03 (10.10.1.3).
Check that the host is reachable and accepts connections using the
cephadm SSH key
you may want to run:
ssh -F =(ceph cephadm get-ssh-config) -i =(ceph
config-key get mgr/cephadm/ssh_identity_key) root@ceph03
host ceph04 failed
check: Failed to connect to ceph04 (10.10.1.4).
Check that the host is reachable and accepts connections using the
cephadm SSH key
you may want to run:
ssh -F =(ceph cephadm get-ssh-config) -i =(ceph
config-key get mgr/cephadm/ssh_identity_key) root@ceph04
host ceph05 failed
check: Failed to connect to ceph05 (10.10.1.5).
Check that the host is reachable and accepts connections using the
cephadm SSH key
you may want to run:
ssh -F =(ceph cephadm get-ssh-config) -i =(ceph
config-key get mgr/cephadm/ssh_identity_key) root@ceph05
host ceph06 failed
check: Failed to connect to ceph06 (10.10.1.6).
Check that the host is reachable and accepts connections using the
cephadm SSH key
on ceph01 i run:
ceph cephadm get-ssh-config > /tmp/ceph.conf
ceph config-key get mgr/cephadm/ssh_identity_key > /tmp/ceph.key
chmod 600 /tmp/ceph.key
ssh -F /tmp/ceph.conf -i /tmp/ceph.key root@ceph01 (which works)
So i can not understand the errors above.
root@ceph01:~# ceph versions
{
"mon": {
"ceph version 15.2.1
(9fd2f65f91d9246fae2c841a6222d34d121680ee) octopus (stable)": 3
},
"mgr": {
"ceph version 15.2.1
(9fd2f65f91d9246fae2c841a6222d34d121680ee) octopus (stable)": 3
},
"osd": {
"ceph version 15.2.1
(9fd2f65f91d9246fae2c841a6222d34d121680ee) octopus (stable)": 56
},
"mds": {
"ceph version 15.2.1
(9fd2f65f91d9246fae2c841a6222d34d121680ee) octopus (stable)": 1
},
"overall": {
"ceph version 15.2.1
(9fd2f65f91d9246fae2c841a6222d34d121680ee) octopus (stable)": 63
}
}
root@ceph01:~# dpkg -l |grep ceph
ii ceph-base 15.2.4-1~bpo10+1
amd64 common ceph daemon libraries and management tools
ii ceph-common 15.2.4-1~bpo10+1
amd64 common utilities to mount and interact with a ceph
storage cluster
ii ceph-deploy 2.0.1
all Ceph-deploy is an easy to use configuration tool
ii ceph-fuse 15.2.4-1~bpo10+1
amd64 FUSE-based client for the Ceph distributed file system
ii ceph-grafana-dashboards 15.2.4-1~bpo10+1
all grafana dashboards for the ceph dashboard
ii ceph-mds 15.2.4-1~bpo10+1
amd64 metadata server for the ceph distributed file system
ii ceph-mgr 15.2.4-1~bpo10+1
amd64 manager for the ceph distributed storage system
ii ceph-mgr-cephadm 15.2.4-1~bpo10+1
all cephadm orchestrator module for ceph-mgr
ii ceph-mgr-dashboard 15.2.4-1~bpo10+1
all dashboard module for ceph-mgr
ii ceph-mgr-diskprediction-cloud 15.2.4-1~bpo10+1
all diskprediction-cloud module for ceph-mgr
ii ceph-mgr-diskprediction-local 15.2.4-1~bpo10+1
all diskprediction-local module for ceph-mgr
ii ceph-mgr-k8sevents 15.2.4-1~bpo10+1
all kubernetes events module for ceph-mgr
ii ceph-mgr-modules-core 15.2.4-1~bpo10+1
all ceph manager modules which are always enabled
ii ceph-mon 15.2.4-1~bpo10+1
amd64 monitor server for the ceph storage system
ii ceph-osd 15.2.4-1~bpo10+1
amd64 OSD server for the ceph storage system
ii cephadm 15.2.4-1~bpo10+1
amd64 cephadm utility to bootstrap ceph daemons with systemd
and containers
ii libcephfs1 10.2.11-2
amd64 Ceph distributed file system client library
ii libcephfs2 15.2.4-1~bpo10+1
amd64 Ceph distributed file system client library
ii python-ceph-argparse 14.2.8-1
all Python 2 utility libraries for Ceph CLI
ii python3-ceph-argparse 15.2.4-1~bpo10+1
all Python 3 utility libraries for Ceph CLI
ii python3-ceph-common 15.2.4-1~bpo10+1
all Python 3 utility libraries for Ceph
ii python3-cephfs 15.2.4-1~bpo10+1
amd64 Python 3 libraries for the Ceph libcephfs library
root@ceph01:~# ceph -s
cluster:
id: 5436dd5d-83d4-4dc8-a93b-60ab5db145df
health: HEALTH_WARN
6 hosts fail cephadm check
failed to probe daemons or devices
7 nearfull osd(s)
Reduced data availability: 1 pg inactive
Low space hindering backfill (add storage if this doesn't
resolve itself): 26 pgs backfill_toofull
Degraded data redundancy: 202495/33226941 objects degraded
(0.609%), 26 pgs degraded, 26 pgs undersized
3 pool(s) nearfull
services:
mon: 3 daemons, quorum ceph01,ceph02,ceph03 (age 39m)
mgr: ceph02(active, since 77m), standbys: ceph03, ceph01
mds: 2 up:standby
osd: 61 osds: 56 up (since 41m), 55 in (since 41m); 27 remapped pgs
data:
pools: 3 pools, 2049 pgs
objects: 11.08M objects, 37 TiB
usage: 113 TiB used, 28 TiB / 141 TiB avail
pgs: 0.049% pgs not active
202495/33226941 objects degraded (0.609%)
9238/33226941 objects misplaced (0.028%)
1025 active+clean
887 active+clean+snaptrim_wait
110 active+clean+snaptrim
25 active+undersized+degraded+remapped+backfill_toofull
1 undersized+degraded+remapped+backfill_toofull+peered
1 active+remapped+backfilling
io:
client: 1.0 KiB/s rd, 140 KiB/s wr, 0 op/s rd, 1 op/s wr
recovery: 30 MiB/s, 8 objects/s
I already restarted the mgr on ceph02.
Thanks,
Michael
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io
--
SUSE Software Solutions Germany GmbH, Maxfeldstr. 5, 90409 Nürnberg, Germany
(HRB 36809, AG Nürnberg). Geschäftsführer: Felix Imendörffer