Sorry for replying late :(. And thanks for the tips.
This is a fresh cluster. And I didn’t think data distribution would be a problem. Is this
normal?
Below is the ceph osd df output. The related pool is hdd only (prod.rgw.buckets.data). I
guess there is variance but I couldn’t get the reason. Is it because of PG number which I
get help from pg-calculator <https://ceph.io/pgcalc/>? Or is this expected ceph
behaviour?
#ceph osd df
https://pastebin.ubuntu.com/p/ZmQZsGYpr7/
<https://pastebin.ubuntu.com/p/7C9zpXYntR/>
I am also sharing related cluster information. Any suggestion would be appreciated.
#ceph df
https://pastebin.ubuntu.com/p/sXpf99zhnV/
<https://pastebin.ubuntu.com/p/sXpf99zhnV/>
#ceph detail df
https://pastebin.ubuntu.com/p/dwvwBnnBmv/
<https://pastebin.ubuntu.com/p/dwvwBnnBmv/>
#ceph osd pool ls detail
https://pastebin.ubuntu.com/p/c2KQD5CGMV/
<https://pastebin.ubuntu.com/p/c2KQD5CGMV/>
#crush rules
https://pastebin.ubuntu.com/p/X6WsZhV3Zz/
<https://pastebin.ubuntu.com/p/X6WsZhV3Zz/>
Thanks.
On 26 Jan 2021, at 11:18, Anthony D'Atri
<anthony.datri(a)gmail.com> wrote:
ceph osd df | sort -nk8
On Jan 25, 2021, at 11:22 PM, George Yil
<georgeyil75(a)gmail.com> wrote:
Hi,
I have a ceph nautilus (14.2.9) cluster with 10 nodes. Each node has
19x16TB disks attached.
I created radosgw pools. secondaryzone.rgw.buckets.data pool is configured
as EC 8+2 (jerasure).
ceph df shows 2.1PiB MAX AVAIL space.
Then I configured radosgw as a secondary zone and 100TiB of S3 data is
replicated.
But weirdly enough ceph df shows 1.8PiB MAX AVAIL for the same pool. But
there is only 100TiB of written data. ceph df also confirms it. I can not
figure out where 200TiB capacity is gone.
Would someone please tell me what I am missing?
Thanks.
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io