On 31/08/2020 12:31, Igor Fedotov wrote:
Hi Wido,
'b' prefix relates to free list manager which keeps all the free extents
for main device in a bitmap. Its records have fixed size hence you can
easily estimate the overall size for these type of data.
Yes, so I figured.
But I doubt it takes that much. I presume that DB just
lacks the proper
compaction. Which could happen eventually but looks like you interrupted
the process by going offline.
May be try manual compaction with ceph-kvstore-tool?
This cluster is suffering from a lot of spillovers. So we tested with
marking one OSD as out.
After being marked as out it still had this large DB. A compact didn't
work, the RocksDB database just stayed so large.
New OSDs coming into the cluster aren't suffering from this and they
have a RocksDB of a couple of MB in size.
Old OSDs installed with Luminous and now upgraded to Nautilus are
suffering from this.
It kind of seems like that garbage data stays behind in RocksDB which is
never clean up.
Wido
>
> Thanks,
>
> Igor
>
>
>
> On 8/31/2020 10:57 AM, Wido den Hollander wrote:
>> Hello,
>>
>> On a Nautilus 14.2.8 cluster I am seeing large RocksDB database with
>> many slow DB bytes in use.
>>
>> To investigate this further I marked one OSD as out and waited for the
>> all the backfilling to complete.
>>
>> Once the backfilling was completed I exported BlueFS and investigated
>> the RocksDB using 'ceph-kvstore-tool'. This resulted in 22GB of data.
>>
>> Listing all the keys in the RocksDB shows me there are 747.000 keys in
>> the DB. A small portion are osdmaps, but the biggest amount are keys
>> prefixed with 'b'.
>>
>> I dumped the stats of the RocksDB and this shows me:
>>
>> L1: 1/0: 439.32 KB
>> L2: 1/0: 2.65 MB
>> L3: 5/0: 14.36 MB
>> L4: 127/0: 7.22 GB
>> L5: 217/0: 13.73 GB
>> Sum: 351/0: 20.98 GB
>>
>> So there is almost 21GB of data in this RocksDB database. Why? Where
>> is this coming from?
>>
>> Throughout this cluster OSDs are suffering from many slow bytes used
>> and I can't figure out why.
>>
>> Has anybody seen this or has a clue on what is going on?
>>
>> I have an external copy of this RocksDB database to do investigations on.
>>
>> Thank you,
>>
>> Wido
>> _______________________________________________
>> ceph-users mailing list -- ceph-users(a)ceph.io
>> To unsubscribe send an email to ceph-users-leave(a)ceph.io