Eugen,
I didn't really think my cluster was eating itself, but I also didn't want
to be in denial.
Regarding the autoscaler, I really thought that it only went up - I didn't
expect that it would decrease the number of PGs. Plus, I thought I had it
turned off. I see now that it's off globally but enabled for this
particular pool. Also, I see that the target PG count is lower than the
current.
I guess you learn something new every day.
-Dave
--
Dave Hall
Binghamton University
kdhall(a)binghamton.edu
607-760-2328 (Cell)
607-777-4641 (Office)
On Mon, Mar 29, 2021 at 7:52 AM Eugen Block <eblock(a)nde.ag> wrote:
Hi,
that sounds like the pg_autoscaler is doing its work. Check with:
ceph osd pool autoscale-status
I don't think ceph is eating itself or that you're losing data. ;-)
Zitat von Dave Hall <kdhall(a)binghamton.edu>du>:
Hello,
About 3 weeks ago I added a node and increased the number of OSDs in my
cluster from 24 to 32, and then marked one old OSD down because it was
frequently crashing. .
After adding the new OSDs the PG count jumped fairly dramatically, but
ever
since, amidst a continuous low level of
rebalancing, the number of PGs
has
gradually decreased to less by 25% from it's
max value. Although I don't
have specific notes, my perception is that the current number of PGs is
actually lower than it was before I added OSDs.
So what's going on here? It is possible to imagine that my cluster is
slowly eating itself, and that I'm about to lose 200TB of data. It's also
possible to imagine that this is all due to the gradual optimization of
the
pools.
Note that the primary pool is an EC 8 + 2 containing about 124TB.
Thanks.
-Dave
--
Dave Hall
Binghamton University
kdhall(a)binghamton.edu
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io