On Fri, Oct 20, 2023, 8:11 AM Zakhar Kirpichenko
<zakhar(a)gmail.com> wrote:
Thank you, Igor.
It is somewhat disappointing that fixing this bug in Pacific has such a
low
priority, considering its impact on existing clusters.
Unfortunately, the hard truth here is that Pacific (stable) was released
over 30 months ago. It has had a good run for a freely distributed product,
and there's only so much time you can dedicate to backporting bugfixes --
it claws time away from other forward-thinking initiatives.
Speaking from someone who's been at the helm of production clusters, I
know Ceph upgrades can be an experience and it's frustrating to hear, but
you have to jump sometime...
Regards,
Tyler
On Fri, 20 Oct 2023 at 14:46, Igor Fedotov
<igor.fedotov(a)croit.io> wrote:
Hi Zakhar,
Definitely we expect one more (and apparently the last) Pacific minor
release. There is no specific date yet though - the plans are to release
Quincy and Reef minor releases prior to it. Hopefully to be done before
the
Christmas/New Year.
Meanwhile you might want to workaround the issue by tuning
bluestore_volume_selection_policy. Unfortunately most likely my original
proposal to set it to rocksdb_original wouldn't work in this case so you
better try "fit_to_fast" mode. This should be coupled with enabling
'level_compaction_dynamic_level_bytes' mode in RocksDB - there is pretty
good spec on applying this mode to BlueStore attached to
https://github.com/ceph/ceph/pull/37156.
Thanks,
Igor
On 20/10/2023 06:03, Zakhar Kirpichenko wrote:
Igor, I noticed that there's no roadmap for the next 16.2.x release.
May I
ask what time frame we are looking at with
regards to a possible fix?
We're experiencing several OSD crashes caused by this issue per day.
/Z
On Mon, 16 Oct 2023 at 14:19, Igor Fedotov <igor.fedotov(a)croit.io>
wrote:
> That's true.
> On 16/10/2023 14:13, Zakhar Kirpichenko wrote:
>
> Many thanks, Igor. I found previously submitted bug reports and
> subscribed to them. My understanding is that the issue is going to be
fixed
> in the next Pacific minor release.
>
> /Z
>
> On Mon, 16 Oct 2023 at 14:03, Igor Fedotov <igor.fedotov(a)croit.io>
wrote:
>
>> Hi Zakhar,
>>
>> please see my reply for the post on the similar issue at:
>>
>>
https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/message/YNJ35HXN4H…
>>
>>
>> Thanks,
>>
>> Igor
>>
>> On 16/10/2023 09:26, Zakhar Kirpichenko wrote:
>> > Hi,
>> >
>> > After upgrading to Ceph 16.2.14 we had several OSD crashes
>> > in bstore_kv_sync thread:
>> >
>> >
>> > 1. "assert_thread_name": "bstore_kv_sync",
>> > 2. "backtrace": [
>> > 3. "/lib64/libpthread.so.0(+0x12cf0) [0x7ff2f6750cf0]",
>> > 4. "gsignal()",
>> > 5. "abort()",
>> > 6. "(ceph::__ceph_assert_fail(char const*, char const*, int,
char
>> > const*)+0x1a9)
[0x564dc5f87d0b]",
>> > 7. "/usr/bin/ceph-osd(+0x584ed4) [0x564dc5f87ed4]",
>> > 8. "(RocksDBBlueFSVolumeSelector::sub_usage(void*,
bluefs_fnode_t
>> > const&)+0x15e)
[0x564dc6604a9e]",
>> > 9. "(BlueFS::_flush_range_F(BlueFS::FileWriter*, unsigned long,
>> unsigned
>> > long)+0x77d) [0x564dc66951cd]",
>> > 10. "(BlueFS::_flush_F(BlueFS::FileWriter*, bool, bool*)+0x90)
>> > [0x564dc6695670]",
>> > 11. "(BlueFS::fsync(BlueFS::FileWriter*)+0x18b)
[0x564dc66b1a6b]",
>> > 12.
"(BlueRocksWritableFile::Sync()+0x18) [0x564dc66c1768]",
>> > 13.
"(rocksdb::LegacyWritableFileWrapper::Sync(rocksdb::IOOptions
>> > const&,
rocksdb::IODebugContext*)+0x1f) [0x564dc6b6496f]",
>> > 14. "(rocksdb::WritableFileWriter::SyncInternal(bool)+0x402)
>> > [0x564dc6c761c2]",
>> > 15. "(rocksdb::WritableFileWriter::Sync(bool)+0x88)
>> [0x564dc6c77808]",
>> > 16.
"(rocksdb::DBImpl::WriteToWAL(rocksdb::WriteThread::WriteGroup
>> > const&,
rocksdb::log::Writer*, unsigned long*, bool, bool,
unsigned
>> > long)+0x309)
[0x564dc6b780c9]",
>> > 17. "(rocksdb::DBImpl::WriteImpl(rocksdb::WriteOptions const&,
>> > rocksdb::WriteBatch*, rocksdb::WriteCallback*, unsigned long*,
>> unsigned
>> > long, bool, unsigned long*, unsigned long,
>> > rocksdb::PreReleaseCallback*)+0x2629) [0x564dc6b80c69]",
>> > 18. "(rocksdb::DBImpl::Write(rocksdb::WriteOptions const&,
>> > rocksdb::WriteBatch*)+0x21) [0x564dc6b80e61]",
>> > 19. "(RocksDBStore::submit_common(rocksdb::WriteOptions&,
>> > std::shared_ptr<KeyValueDB::TransactionImpl>)+0x84)
>> [0x564dc6b1f644]",
>> > 20.
>>
"(RocksDBStore::submit_transaction_sync(std::shared_ptr<KeyValueDB::TransactionImpl>)+0x9a)
>> > [0x564dc6b2004a]",
>> > 21. "(BlueStore::_kv_sync_thread()+0x30d8) [0x564dc6602ec8]",
>> > 22. "(BlueStore::KVSyncThread::entry()+0x11)
[0x564dc662ab61]",
>> > 23. "/lib64/libpthread.so.0(+0x81ca) [0x7ff2f67461ca]",
>> > 24. "clone()"
>> > 25. ],
>> >
>> >
>> > I am attaching two instances of crash info for further reference:
>> >
https://pastebin.com/E6myaHNU
>> >
>> > OSD configuration is rather simple and close to default:
>> >
>> > osd.6 dev bluestore_cache_size_hdd
4294967296
>> >
osd.6 dev
>> > bluestore_cache_size_ssd 4294967296
>> > osd advanced debug_rocksdb
>> > 1/5
>> osd
>> > advanced osd_max_backfills 2
>> > osd basic
>> > osd_memory_target 17179869184
>> > osd advanced osd_recovery_max_active
>> > 2
osd
>> > advanced osd_scrub_sleep
0.100000
>> > osd advanced
>> > rbd_balance_parent_reads false
>> >
>> > debug_rocksdb is a recent change, otherwise this configuration has
been
>> > running without issues for months.
The crashes happened on two
>> different
>> > hosts with identical hardware, the hosts and storage (NVME DB/WAL,
HDD
>> > block) don't exhibit any issues.
We have not experienced such
crashes
>> with
>> > Ceph < 16.2.14.
>> >
>> > Is this a known issue, or should I open a bug report?
>> >
>> > Best regards,
>> > Zakhar
>> > _______________________________________________
>> > ceph-users mailing list -- ceph-users(a)ceph.io
>> > To unsubscribe send an email to ceph-users-leave(a)ceph.io
>>
>
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io