Paul, thank you!
Do you mean that value?
total_space 75.3TiB
Could you tell me where I can read about algorithm of calculating of worst
case scenario?
*rados df out:*
POOL_NAME USED OBJECTS CLONES COPIES MISSING_ON_PRIMARY UNFOUND
DEGRADED RD_OPS RD WR_OPS WR
ala01vf01p01 16.7TiB 4411305 58090 8822610 0 0
0 2819034334 421TiB 931369992 1.43PiB
ala01vf01p02 1.35TiB 358796 0 717592 0 0
0 1109215060 57.5TiB 1286147031 38.5TiB
cachepool 1.60TiB 420974 5 841948 0 0
0 21677037783 2.61PiB 16048037966 2.06PiB
total_objects 5191075
total_used 39.3TiB
total_avail 36.0TiB
*total_space 75.3TiB*
вт, 1 окт. 2019 г. в 01:46, Paul Emmerich <paul.emmerich(a)croit.io>io>:
ceph df shows a worst-case estimate based on current
data
distribution, check "rados df" for more "raw" counts
Paul
--
Paul Emmerich
Looking for help with your Ceph cluster? Contact us at
https://croit.io
croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90
On Mon, Sep 30, 2019 at 7:49 PM Ilmir Mulyukov <ilmir.mulyukov(a)gmail.com>
wrote:
Hello!
We have a small proxmox farm with
ceph consisting of three nodes.
Each node has 6 disks each with a capacity of 4 TB.
A only one pool has been created on these disks.
Size 2/1.
In theory, this pool should have a capacity: 32.74 TB
But the ceph df command returns only: 22.4 TB (USED + MAX AVAIL)(16.7 +
5.7)
How to explain this difference?
ceph version is: 12.2.12-pve1
ceph df command out:
POOLS:
NAME ID QUOTA OBJECTS QUOTA BYTES USED
%USED MAX
AVAIL OBJECTS DIRTY READ WRITE
RAW USED
ala01vf01p01 7 N/A N/A
16.7TiB
74.53 5.70TiB 4411119 4.41M 2.62GiB
887MiB
33.4TiB
crush map:
host n01vf01 {
id -3 # do not change unnecessarily
id -4 class hdd # do not change unnecessarily
id -18 class nvme # do not change unnecessarily
# weight 22.014
alg straw2
hash 0 # rjenkins1
item osd.0 weight 3.669
item osd.13 weight 3.669
item osd.14 weight 3.669
item osd.15 weight 3.669
item osd.16 weight 3.669
item osd.17 weight 3.669
}
host n02vf01 {
id -5 # do not change unnecessarily
id -6 class hdd # do not change unnecessarily
id -19 class nvme # do not change unnecessarily
# weight 22.014
alg straw2
hash 0 # rjenkins1
item osd.1 weight 3.669
item osd.8 weight 3.669
item osd.9 weight 3.669
item osd.10 weight 3.669
item osd.11 weight 3.669
item osd.12 weight 3.669
}
host n04vf01 {
id -34 # do not change unnecessarily
id -35 class hdd # do not change unnecessarily
id -36 class nvme # do not change unnecessarily
# weight 22.014
alg straw2
hash 0 # rjenkins1
item osd.7 weight 3.669
item osd.27 weight 3.669
item osd.24 weight 3.669
item osd.25 weight 3.669
item osd.26 weight 3.669
item osd.28 weight 3.669
}
root default {
id -1 # do not change unnecessarily
id -2 class hdd # do not change unnecessarily
id -21 class nvme # do not change unnecessarily
# weight 66.042
alg straw2
hash 0 # rjenkins1
item n01vf01 weight 22.014
item n02vf01 weight 22.014
item n04vf01 weight 22.014
}
rule replicated_rule {
id 0
type replicated
min_size 1
max_size 10
step take default
step chooseleaf firstn 0 type host
step emit
}
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io