Hello everybody,
I have a Ceph cluster (14.2.9) running with 18 X 600GB HDD OSDs. There are
three pools (size:3, pg_num:64) with an image size of 200GB on each, and
there are 6 servers connected to these images via iSCSI and storing about
20 VMs on them. Here is the output of "ceph df":
POOLS:
POOL ID STORED OBJECTS USED
%USED MAX AVAIL
cephfs_data 1 0 B 0 0 B
0 0 B
cephfs_metadata 2 17 KiB 22 1.5 MiB
100.00 0 B
defaults.rgw.buckets.data 3 0 B 0 0 B
0 0 B
defaults.rgw.buckets.index 4 0 B 0 0 B
0 0 B
.rgw.root 5 2.0 KiB 5 960 KiB
100.00 0 B
default.rgw.control 6 0 B 8 0 B
0 0 B
default.rgw.meta 7 393 B 2 384 KiB
100.00 0 B
default.rgw.log 8 0 B 207 0 B
0 0 B
rbd 9 150 GiB 38.46k 450 GiB
100.00 0 B
rbd3 13 270 GiB 69.24k 811 GiB
100.00 0 B
rbd2 14 150 GiB 38.52k 451 GiB
100.00 0 B
Based on this, I expect about 1.7 TB RAW capacity usage, BUT it is
currently about 9TBs!
RAW STORAGE:
CLASS SIZE AVAIL USED RAW USED %RAW USED
hdd 9.8 TiB 870 GiB 9.0 TiB 9.0 TiB 91.35
TOTAL 9.8 TiB 870 GiB 9.0 TiB 9.0 TiB 91.35
And the cluster is down because there is very little capacity remaining. I
wonder what makes this and how can I get it fixed. Your help is much
appreciated
Thank You
Show replies by date