All,
In looking at the options for setting the default pg autoscale option, I
notice that there is a global option setting and a per-pool option
setting. It seems that the options at the pool level are off, warn, and
on. The same, I assume for the global setting.
Is there a way to get rid of the per-pool setting and set the pool to honor
the global setting? I think I'm looking for 'off, warn, on, or global'.
It seems that once the per-pool option is set for all of one's pools, the
global value is irrelevant. This also implies that in a circumstance where
one would want to temporarily suspend autoscaling it would be required to
modify the setting for each pool and then to modify it back afterward.
Thoughts?
Thanks
-Dave
--
Dave Hall
Binghamton University
kdhall(a)binghamton.edu
On Mon, Mar 29, 2021 at 1:44 PM Anthony D'Atri <anthony.datri(a)gmail.com>
wrote:
Yes the PG autoscalar has a way of reducing PG
count way too far. There’s
a claim that it’s better in Pacific, but I tend to recommend disabling it
and calculating / setting pg_num manually.
On Mar 29, 2021, at 9:06 AM, Dave Hall
<kdhall(a)binghamton.edu> wrote:
Eugen,
I didn't really think my cluster was eating itself, but I also didn't
want
to be in denial.
Regarding the autoscaler, I really thought that it only went up - I
didn't
expect that it would decrease the number of PGs.
Plus, I thought I had
it
turned off. I see now that it's off globally
but enabled for this
particular pool. Also, I see that the target PG count is lower than the
current.
I guess you learn something new every day.
-Dave
--
Dave Hall
Binghamton University
kdhall(a)binghamton.edu
607-760-2328 (Cell)
607-777-4641 (Office)
On Mon, Mar 29, 2021 at 7:52 AM Eugen Block <eblock(a)nde.ag> wrote:
> Hi,
>
> that sounds like the pg_autoscaler is doing its work. Check with:
>
> ceph osd pool autoscale-status
>
> I don't think ceph is eating itself or that you're losing data. ;-)
>
>
> Zitat von Dave Hall <kdhall(a)binghamton.edu>du>:
>
>> Hello,
>>
>> About 3 weeks ago I added a node and increased the number of OSDs in my
>> cluster from 24 to 32, and then marked one old OSD down because it was
>> frequently crashing. .
>>
>> After adding the new OSDs the PG count jumped fairly dramatically, but
> ever
>> since, amidst a continuous low level of rebalancing, the number of PGs
> has
>> gradually decreased to less by 25% from it's max value. Although I
don't
>> have specific notes, my perception is
that the current number of PGs is
>> actually lower than it was before I added OSDs.
>>
>> So what's going on here? It is possible to imagine that my cluster is
>> slowly eating itself, and that I'm about to lose 200TB of data. It's
also
>
possible to imagine that this is all due to the gradual optimization of
the
> pools.
>
> Note that the primary pool is an EC 8 + 2 containing about 124TB.
>
> Thanks.
>
> -Dave
>
> --
> Dave Hall
> Binghamton University
> kdhall(a)binghamton.edu
> _______________________________________________
> ceph-users mailing list -- ceph-users(a)ceph.io
> To unsubscribe send an email to ceph-users-leave(a)ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io