Home directories probably means lots of small objects. Default minimum allocation size of BlueStore on HDD is 64 kiB, so there's a lot of overhead for everything smaller;

Details: google bluestore min alloc size, can only be changed during OSD creation

Paul

--
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90


On Fri, Dec 6, 2019 at 12:57 PM Jochen Schulz <schulz@math.uni-goettingen.de> wrote:
Hi!

Thank you!
The output of both commands are below.
I still dont understand why there are 21T used data (because 5.5T*3 =
16.5T != 21T) and why there seems to be only 4.5 T MAX AVAIL, but the
osd output tells we have 25T free space.


$ sudo ceph df
RAW STORAGE:
    CLASS     SIZE        AVAIL       USED        RAW USED     %RAW USED
    hdd        45 TiB      24 TiB      21 TiB       21 TiB         46.33
    ssd       596 GiB     524 GiB     1.7 GiB       72 GiB         12.09
    TOTAL      46 TiB      25 TiB      21 TiB       21 TiB         45.89

POOLS:
    POOL                ID     STORED      OBJECTS     USED        %USED
    MAX AVAIL
    images               8     149 GiB      38.30k     354 GiB      2.52
      4.5 TiB
    cephfs_data          9     5.5 TiB      26.61M      20 TiB     60.36
      4.5 TiB
    cephfs_metadata     10      12 GiB       3.17M      13 GiB      2.57
      164 GiB


$ sudo ceph osd df
ID CLASS WEIGHT  REWEIGHT SIZE    RAW USE DATA    OMAP    META    AVAIL
   %USE  VAR  PGS STATUS
 0   hdd 0.89000  1.00000 931 GiB 456 GiB 453 GiB 136 MiB 3.5 GiB  475
GiB 49.01 1.07 103     up
 1   hdd 0.89000  1.00000 931 GiB 495 GiB 491 GiB 100 MiB 3.9 GiB  436
GiB 53.14 1.16  89     up
 4   hdd 0.89000  1.00000 931 GiB 345 GiB 342 GiB 108 MiB 2.9 GiB  586
GiB 37.05 0.81  87     up
 5   hdd 0.89000  1.00000 931 GiB 521 GiB 517 GiB 108 MiB 4.1 GiB  410
GiB 55.96 1.22  98     up
 6   hdd 0.89000  1.00000 931 GiB 367 GiB 364 GiB  95 MiB 3.2 GiB  564
GiB 39.44 0.86  95     up
 7   hdd 0.89000  1.00000 931 GiB 540 GiB 536 GiB  77 MiB 3.7 GiB  392
GiB 57.96 1.26 111     up
20   hdd 0.89000  1.00000 931 GiB 382 GiB 378 GiB  60 MiB 3.3 GiB  550
GiB 40.96 0.89  85     up
23   hdd 1.81929  1.00000 1.8 TiB 706 GiB 701 GiB 113 MiB 4.9 GiB  1.1
TiB 37.92 0.83 182     up
44   hdd 0.89000  1.00000 931 GiB 468 GiB 465 GiB  34 MiB 3.3 GiB  463
GiB 50.29 1.10  93     up
45   hdd 1.78999  1.00000 1.8 TiB 882 GiB 875 GiB 138 MiB 6.3 GiB  981
GiB 47.33 1.03 179     up
46   hdd 1.78999  1.00000 1.8 TiB 910 GiB 903 GiB 127 MiB 6.4 GiB  953
GiB 48.83 1.06 192     up
22   ssd 0.11639  1.00000 119 GiB  15 GiB 357 MiB  12 GiB 2.8 GiB  104
GiB 12.61 0.27 315     up
12   hdd 0.89000  1.00000 931 GiB 499 GiB 494 GiB  64 MiB 4.5 GiB  432
GiB 53.57 1.17 116     up
13   hdd 0.89000  1.00000 931 GiB 536 GiB 532 GiB  48 MiB 4.4 GiB  395
GiB 57.59 1.26 109     up
30   hdd 0.89000  1.00000 931 GiB 510 GiB 506 GiB  33 MiB 3.9 GiB  421
GiB 54.80 1.19 100     up
32   hdd 0.89000  1.00000 931 GiB 495 GiB 491 GiB  56 MiB 4.1 GiB  436
GiB 53.17 1.16 101     up
33   hdd 0.89000  1.00000 931 GiB 333 GiB 330 GiB  56 MiB 3.1 GiB  598
GiB 35.80 0.78  82     up
15   ssd 0.11639  1.00000 119 GiB  14 GiB 336 MiB  11 GiB 2.9 GiB  105
GiB 12.13 0.26 305     up
17   hdd 0.89000  1.00000 931 GiB 577 GiB 573 GiB  77 MiB 4.4 GiB  354
GiB 61.99 1.35  97     up
18   hdd 0.89000  1.00000 931 GiB 413 GiB 409 GiB  70 MiB 4.0 GiB  518
GiB 44.34 0.97  95     up
19   hdd 1.81879  1.00000 1.8 TiB 895 GiB 889 GiB 144 MiB 5.6 GiB  967
GiB 48.06 1.05 184     up
21   hdd 0.89000  1.00000 931 GiB 360 GiB 357 GiB  60 MiB 3.4 GiB  570
GiB 38.72 0.84 100     up
31   hdd 0.90909  1.00000 931 GiB 508 GiB 505 GiB  80 MiB 3.5 GiB  423
GiB 54.58 1.19 102     up
25   ssd 0.11639  1.00000 119 GiB  14 GiB 339 MiB  11 GiB 2.7 GiB  105
GiB 11.86 0.26 310     up
 8   hdd 0.89000  1.00000 931 GiB 359 GiB 356 GiB  72 MiB 3.1 GiB  572
GiB 38.55 0.84  80     up
 9   hdd 0.89000  1.00000 931 GiB 376 GiB 373 GiB  42 MiB 3.0 GiB  555
GiB 40.39 0.88  87     up
24   hdd 0.89000  1.00000 931 GiB 342 GiB 339 GiB  70 MiB 2.8 GiB  590
GiB 36.69 0.80  78     up
26   hdd 1.78999  1.00000 1.8 TiB 921 GiB 915 GiB 129 MiB 6.1 GiB  942
GiB 49.45 1.08 177     up
27   hdd 1.78999  1.00000 1.8 TiB 891 GiB 885 GiB 125 MiB 5.7 GiB  972
GiB 47.82 1.04 208     up
35   hdd 1.81929  1.00000 1.8 TiB 819 GiB 814 GiB 110 MiB 5.3 GiB  1.0
TiB 43.99 0.96 184     up
29   ssd 0.11638  1.00000 119 GiB  15 GiB 339 MiB  11 GiB 2.9 GiB  105
GiB 12.25 0.27 311     up
14   hdd 1.81929  1.00000 1.8 TiB 889 GiB 884 GiB 1.9 MiB 4.3 GiB  974
GiB 47.70 1.04 162     up
28   hdd 1.81929  1.00000 1.8 TiB 609 GiB 606 GiB 3.4 MiB 2.9 GiB  1.2
TiB 32.67 0.71 169     up
36   hdd 1.36429  1.00000 1.4 TiB 591 GiB 589 GiB 2.0 MiB 2.9 GiB  806
GiB 42.34 0.92 135     up
37   hdd 1.81929  1.00000 1.8 TiB 840 GiB 836 GiB 2.5 MiB 4.1 GiB 1023
GiB 45.12 0.98 177     up
38   hdd 1.81929  1.00000 1.8 TiB 914 GiB 909 GiB 2.2 MiB 4.4 GiB  949
GiB 49.05 1.07 182     up
 2   hdd 0.89000  1.00000 931 GiB 449 GiB 445 GiB  60 MiB 3.4 GiB  482
GiB 48.19 1.05  89     up
 3   hdd 0.89000  1.00000 931 GiB 333 GiB 329 GiB  75 MiB 3.0 GiB  598
GiB 35.73 0.78  86     up
10   hdd 0.89000  1.00000 931 GiB 416 GiB 412 GiB  54 MiB 3.8 GiB  515
GiB 44.68 0.97 100     up
11   hdd 1.81879  1.00000 1.8 TiB 803 GiB 798 GiB  70 MiB 5.3 GiB  1.0
TiB 43.14 0.94 188     up
34   hdd 0.90909  1.00000 931 GiB 605 GiB 600 GiB  64 MiB 4.2 GiB  326
GiB 64.95 1.42 106     up
16   ssd 0.11639  1.00000 119 GiB  14 GiB 322 MiB  10 GiB 3.1 GiB  105
GiB 11.59 0.25 295     up
                    TOTAL  46 TiB  21 TiB  21 TiB  59 GiB 165 GiB   25
TiB 45.89
MIN/MAX VAR: 0.25/1.42  STDDEV: 13.76




On 06.12.19 12:27, Aleksey Gutikov wrote:
> On 6.12.19 13:29, Jochen Schulz wrote:
>> Hi!
>>
>> We have a ceph cluster with 42 OSD in production as a server providing
>> mainly home-directories of users. Ceph is 14.2.4 nautilus.
>>
>> We have 3 pools. One images (for rbd images) a cephfs_metadata and a
>> cephfs_data pool.
>>
>> Our raw data is about 5.6T. All pools have replica size 3 and there are
>> only very little snapshots in the rbd images pool, the cephfspool doesnt
>> use snapshots.
>>
>> How is it possible that the status tells us, that 21T/46T  is used,
>> because thats much more than 3 times the raw size.
>>
>> Also, to make that more confusing, there as at least half of the cluster
>> free, and we get pg backfill_toofull after we added some OSDs lately.
>> The ceph dashboard tells aus the pool ist 82 % full and has only 4.5 T
>> free.
>>
>> The autoscale module seems to calculate the 20T times 3 for the space
>> needed and thus has wrong numbers (see below).
>>
>> Status of the cluster is added below too.
>>
>> how can these size/capacity numbers be explained?
>> and, would be there a recommendation to change something?
>>
>> Thank you in advance!
>>
>> best
>> Jochen
>>
>>
>> # ceph -s
>>
>>   cluster:
>>      id:     2b16167f-3f33-4580-a0e9-7a71978f403d
>>      health: HEALTH_ERR
>>              Degraded data redundancy (low space): 1 pg backfill_toofull
>>              1 subtrees have overcommitted pool target_size_bytes
>>              1 subtrees have overcommitted pool target_size_ratio
>>              2 pools have too many placement groups
>>
>>    services:
>>      mon: 4 daemons, quorum jade,assam,matcha,jasmine (age 2d)
>>      mgr: earl(active, since 24h), standbys: assam
>>      mds: cephfs:1 {0=assam=up:active} 1 up:standby
>>      osd: 42 osds: 42 up (since 106m), 42 in (since 115m); 30 remapped
>> pgs
>>
>>    data:
>>      pools:   3 pools, 2048 pgs
>>      objects: 29.80M objects, 5.6 TiB
>>      usage:   21 TiB used, 25 TiB / 46 TiB avail
>>      pgs:     1164396/89411013 objects misplaced (1.302%)
>>               2018 active+clean
>>               22   active+remapped+backfill_wait
>>               7    active+remapped+backfilling
>>               1    active+remapped+backfill_wait+backfill_toofull
>>
>>    io:
>>      client:   1.7 KiB/s rd, 516 KiB/s wr, 0 op/s rd, 28 op/s wr
>>      recovery: 9.2 MiB/s, 41 objects/s
>>
>>
>> # ceph osd pool autoscale-status
>>   POOL               SIZE  TARGET SIZE  RATE  RAW CAPACITY   RATIO
>> TARGET RATIO  BIAS  PG_NUM  NEW PG_NUM  AUTOSCALE
>>   images           354.2G                3.0        46100G  0.0231
>>           1.0    1024          32  warn
>>   cephfs_metadata  13260M                3.0        595.7G  0.0652
>>           1.0     512           8  warn
>>   cephfs_data      20802G                3.0        46100G  1.3537
>>           1.0     512              warn
>>
>>
>>
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@ceph.io
>> To unsubscribe send an email to ceph-users-leave@ceph.io
>>
>
> Please, provide output of ceph df and ceph osd df - that should explain
> both questions (21T and 82%).
>
>

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-leave@ceph.io