output from "ceph -s "
cluster:
id: 7c138e13-7b98-4309-b591-d4091a1742b4
health: HEALTH_WARN
Degraded data redundancy: 1141587/7723191 objects degraded
(14.781%), 15 pgs degraded, 16 pgs undersized
services:
mon: 1 daemons, quorum mon01
mgr: mon01(active)
mds: cephfs-tst-1/1/1 up {0=mon01=up:active}
osd: 3 osds: 3 up, 3 in; 16 remapped pgs
data:
pools: 2 pools, 64 pgs
objects: 2.57 M objects, 59 GiB
usage: 190 GiB used, 5.3 TiB / 5.5 TiB avail
pgs: 1141587/7723191 objects degraded (14.781%)
48 active+clean
15 active+undersized+degraded+remapped+backfill_wait
1 active+undersized+remapped+backfilling
io:
recovery: 0 B/s, 10 objects/s
output from "ceph osd tree"
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-1 5.45819 root default
-3 1.81940 host test-node1
0 hdd 1.81940 osd.0 up 1.00000 1.00000
-5 1.81940 host test-node2
1 hdd 1.81940 osd.1 up 1.00000 1.00000
-7 1.81940 host test-node3
2 hdd 1.81940 osd.2 up 1.00000 1.00000
failure domain not configured yet, setup is 3 OSD node each with a single
disk, 1 node with mon&mds&mgr running.
the cluster was healthy until I run a script for creating multiple folders.
regards
Amudhan
On Thu, Aug 29, 2019 at 5:33 PM Heðin Ejdesgaard Møller <hej(a)synack.fo>
wrote:
In adition to ceph -s, could you provide the output
of
ceph osd tree
and specify what your failure domain is ?
/Heðin
On hós, 2019-08-29 at 13:55 +0200, Janne Johansson wrote:
Den tors 29 aug. 2019 kl 13:50 skrev Amudhan P <amudhan83(a)gmail.com>om>:
Hi,
I am using ceph version 13.2.6 (mimic) on test setup trying with
cephfs.
my ceph health status showing warning .
"ceph health"
HEALTH_WARN Degraded data redundancy: 1197023/7723191 objects
degraded (15.499%)
"ceph health detail"
HEALTH_WARN Degraded data redundancy: 1197128/7723191 objects
degraded (15.500%)
PG_DEGRADED Degraded data redundancy: 1197128/7723191 objects
degraded (15.500%)
pg 2.0 is stuck undersized for 1076.454929, current state
active+undersized+
pg 2.2 is stuck undersized for 1076.456639, current state
active+undersized+
How does "ceph -s" look?
It should have more info on what else is wrong.
--
May the most significant bit of your life be positive.
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io