We got the same problem today while we were adding memory to OSD nodes, and it decreased
monitor’s performance a lot. I noticed that the db kept increasing after an OSD is
shutdown, so I guess that it is caused by the warning reports collected by mgr insights
module. When I disabled the mgr insights module, the db size reduced to 1xx MB from 3x
GB.
2020年3月12日 下午2:44,Hartwig Hauschild
<ml-ceph(a)hauschild.it> 写道:
Am 10.03.2020 schrieb Wido den Hollander:
On 3/10/20 10:48 AM, Hartwig Hauschild wrote:
> Hi,
>
> I've done a bit more testing ...
>
> Am 05.03.2020 schrieb Hartwig Hauschild:
>> Hi,
>>
[ snipped ]
I've
read somewhere in the docs that I should provide ample space (tens of
GB) for the store.db, found on the ML and Bugtracker that ~100GB might not
be a bad idea and that large clusters may require space on order of
magnitude greater.
Is there some sort of formula I can use to approximate the space required?
I don't know about a formula, but make sure you have enough space. MONs
are dedicated nodes in most production environments, so I usually
install a 400 ~ 1000GB SSD just to make sure they don't run out of space.
That seems fair.
Also: is the db supposed to grow this fast in Nautilus when it did not do
that in Luminous? Is that behaviour configurable somewhere?
The MONs need to cache the OSDMaps when not all PGs are active+clean
thus their database grows.
You can compact RocksDB in the meantime, but it won't last for ever.
Just make sure the MONs have enough space.
Do you happen to know if that behaved differently in previous releases? I'm
just asking because I have not found anything about this yet and may need to
explain that it's different now.
--
Cheers,
Hardy
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io