Hi Ilya,
Thanks for the reply. Here's is the output:
root@zephir:~# rbd status ceph-dev
Watchers:
watcher=192.168.1.1:0/338620854 client.19264246
cookie=18446462598732840969
root@zephir:~# rbd snap create ceph-dev@backup --debug-ms 1 --debug-rbd 20
2023-04-17T18:23:16.211+0200 7f5e05dca4c0 1 Processor -- start
2023-04-17T18:23:16.211+0200 7f5e05dca4c0 1 -- start start
2023-04-17T18:23:16.211+0200 7f5e05dca4c0 1 --2- >> [v2:
192.168.1.1:3300/0,v1:192.168.1.1:6789/0] conn(0x5558b4e3c260
0x5558b4e3c630 unknown :-1 s=NONE pgs=0 cs=0 l=0 rev1=0 crypto rx=0 tx=0
comp rx=0 tx=0).connect
2023-04-17T18:23:16.211+0200 7f5e05dca4c0 1 --2- >> [v2:
192.168.1.10:3300/0,v1:192.168.1.10:6789/0] conn(0x5558b4e3cc00
0x5558b4e452d0 unknown :-1 s=NONE pgs=0 cs=0 l=0 rev1=0 crypto rx=0 tx=0
comp rx=0 tx=0).connect
2023-04-17T18:23:16.211+0200 7f5e05dca4c0 1 --2- >> [v2:
192.168.43.208:3300/0,v1:192.168.43.208:6789/0] conn(0x5558b4e45810
0x5558b4e47bf0 unknown :-1 s=NONE pgs=0 cs=0 l=0 rev1=0 crypto rx=0 tx=0
comp rx=0 tx=0).connect
2023-04-17T18:23:16.211+0200 7f5e05dca4c0 1 -- --> [v2:
192.168.1.1:3300/0,v1:192.168.1.1:6789/0] -- mon_getmap magic: 0 v1 --
0x5558b4d29b70 con 0x5558b4e3c260
2023-04-17T18:23:16.211+0200 7f5e05dca4c0 1 -- --> [v2:
192.168.1.10:3300/0,v1:192.168.1.10:6789/0] -- mon_getmap magic: 0 v1 --
0x5558b4d231b0 con 0x5558b4e3cc00
2023-04-17T18:23:16.211+0200 7f5e05dca4c0 1 -- --> [v2:
192.168.43.208:3300/0,v1:192.168.43.208:6789/0] -- mon_getmap magic: 0 v1
-- 0x5558b4cbf680 con 0x5558b4e45810
2023-04-17T18:23:16.211+0200 7f5e04103700 1 --2- >> [v2:
192.168.1.1:3300/0,v1:192.168.1.1:6789/0] conn(0x5558b4e3c260
0x5558b4e3c630 unknown :-1 s=BANNER_CONNECTING pgs=0 cs=0 l=0 rev1=0 crypto
rx=0 tx=0 comp rx=0 tx=0)._handle_peer_banner_payload supporte
d=3 required=0
2023-04-17T18:23:16.211+0200 7f5e04103700 1 --2- >> [v2:
192.168.1.1:3300/0,v1:192.168.1.1:6789/0] conn(0x5558b4e3c260
0x5558b4e3c630 unknown :-1 s=HELLO_CONNECTING pgs=0 cs=0 l=0 rev1=1 crypto
rx=0 tx=0 comp rx=0 tx=0).handle_hello peer v2:192.168.1.1:3300
/0 says I am v2:192.168.1.1:42714/0 (socket says 192.168.1.1:42714)
2023-04-17T18:23:16.211+0200 7f5e04103700 1 -- 192.168.1.1:0/1614127865
learned_addr learned my addr 192.168.1.1:0/1614127865 (peer_addr_for_me v2:
192.168.1.1:0/0)
2023-04-17T18:23:16.211+0200 7f5e03902700 1 --2- 192.168.1.1:0/1614127865
> [v2:192.168.1.10:3300/0,v1:192.168.1.10:6789/0]
conn(0x5558b4e3cc00
0x5558b4e452d0 unknown :-1 s=BANNER_CONNECTING pgs=0 cs=0 l=0
rev1=0 crypto
rx=0 tx=0 comp rx=0 tx=0)._handle_pe
er_banner_payload supported=3 required=0
2023-04-17T18:23:16.211+0200 7f5e04103700 1 -- 192.168.1.1:0/1614127865 >>
[v2:192.168.43.208:3300/0,v1:192.168.43.208:6789/0] conn(0x5558b4e45810
msgr2=0x5558b4e47bf0 unknown :-1 s=STATE_CONNECTING_RE l=0).mark_down
2023-04-17T18:23:16.211+0200 7f5e04103700 1 --2- 192.168.1.1:0/1614127865
>
[v2:192.168.43.208:3300/0,v1:192.168.43.208:6789/0] conn(0x5558b4e45810
0x5558b4e47bf0 unknown :-1 s=START_CONNECT pgs=0 cs=0 l=0 rev1=0 crypto
rx=0 tx=0 comp rx=0 tx=0).stop
2023-04-17T18:23:16.211+0200 7f5e04103700 1 -- 192.168.1.1:0/1614127865 >>
[v2:192.168.1.10:3300/0,v1:192.168.1.10:6789/0] conn(0x5558b4e3cc00
msgr2=0x5558b4e452d0 unknown :-1 s=STATE_CONNECTION_ESTABLISHED
l=0).mark_down
2023-04-17T18:23:16.211+0200 7f5e04103700 1 --2- 192.168.1.1:0/1614127865
> [v2:192.168.1.10:3300/0,v1:192.168.1.10:6789/0]
conn(0x5558b4e3cc00
0x5558b4e452d0 unknown :-1 s=AUTH_CONNECTING pgs=0 cs=0 l=0
rev1=1 crypto
rx=0 tx=0 comp rx=0 tx=0).stop
2023-04-17T18:23:16.211+0200 7f5e04103700 1 -- 192.168.1.1:0/1614127865
--> [v2:192.168.1.1:3300/0,v1:192.168.1.1:6789/0] --
mon_subscribe({config=0+,monmap=0+}) v3 -- 0x5558b4cd6d60 con
0x5558b4e3c260
2023-04-17T18:23:16.211+0200 7f5e04103700 1 --2- 192.168.1.1:0/1614127865
> [v2:192.168.1.1:3300/0,v1:192.168.1.1:6789/0]
conn(0x5558b4e3c260
0x5558b4e3c630 secure :-1 s=READY pgs=355 cs=0 l=1 rev1=1
crypto
rx=0x7f5df400a700 tx=0x7f5df4005b10 comp rx=0 tx=
0).ready entity=mon.0 client_cookie=2d9464291c26a0e7 server_cookie=0
in_seq=0 out_seq=0
2023-04-17T18:23:16.211+0200 7f5e03101700 1 -- 192.168.1.1:0/1614127865
<== mon.0 v2:192.168.1.1:3300/0 1 ==== mon_map magic: 0 v1 ==== 467+0+0
(secure 0 0 0) 0x7f5df40089b0 con 0x5558b4e3c260
2023-04-17T18:23:16.211+0200 7f5e03101700 1 -- 192.168.1.1:0/1614127865
<== mon.0 v2:192.168.1.1:3300/0 2 ==== config(39 keys) v1 ==== 1461+0+0
(secure 0 0 0) 0x7f5df4008b10 con 0x5558b4e3c260
2023-04-17T18:23:16.211+0200 7f5e03101700 1 -- 192.168.1.1:0/1614127865
<== mon.0 v2:192.168.1.1:3300/0 3 ==== mon_map magic: 0 v1 ==== 467+0+0
(secure 0 0 0) 0x7f5df4011e60 con 0x5558b4e3c260
Creating snap: 10% complete...failed.
rbd: failed to create snapshot: (95) Operation not supported
root@zephir:~#
root@zephir:~# rbd config image list ceph-dev
There are 78 values:
Name Value Source
rbd_atime_update_interval 60 config
rbd_balance_parent_reads false config
rbd_balance_snap_reads false config
rbd_blkin_trace_all false config
rbd_blocklist_expire_seconds 0 config
rbd_blocklist_on_break_lock true config
rbd_cache true config
rbd_cache_block_writes_upfront false config
rbd_cache_max_dirty 25165824 config
rbd_cache_max_dirty_age 1.000000 config
rbd_cache_max_dirty_object 0 config
rbd_cache_policy writearound config
rbd_cache_size 33554432 config
rbd_cache_target_dirty 16777216 config
rbd_cache_writethrough_until_flush false config
rbd_clone_copy_on_read false config
rbd_compression_hint none config
rbd_concurrent_management_ops 10 config
rbd_default_map_options config
rbd_default_snapshot_quiesce_mode required config
rbd_disable_zero_copy_writes true config
rbd_discard_granularity_bytes 65536 config
rbd_enable_alloc_hint true config
rbd_invalidate_object_map_on_timeout true config
rbd_io_scheduler simple config
rbd_io_scheduler_simple_max_delay 0 config
rbd_journal_commit_age 5.000000 config
rbd_journal_max_concurrent_object_sets 0 config
rbd_journal_max_payload_bytes 16384 config
rbd_journal_object_flush_age 0.000000 config
rbd_journal_object_flush_bytes 1048576 config
rbd_journal_object_flush_interval 0 config
rbd_journal_object_max_in_flight_appends 0 config
rbd_journal_object_writethrough_until_flush true config
rbd_localize_parent_reads false config
rbd_localize_snap_reads false config
rbd_mirroring_delete_delay 0 config
rbd_mirroring_max_mirroring_snapshots 5 config
rbd_mirroring_replay_delay 0 config
rbd_mirroring_resync_after_disconnect false config
rbd_move_parent_to_trash_on_remove false config
rbd_move_to_trash_on_remove true config
rbd_move_to_trash_on_remove_expire_seconds 0 config
rbd_mtime_update_interval 60 config
rbd_non_blocking_aio true config
rbd_parent_cache_enabled false config
rbd_persistent_cache_mode disabled config
rbd_persistent_cache_path /tmp config
rbd_persistent_cache_size 1073741824 config
rbd_plugins config
rbd_qos_bps_burst 0 config
rbd_qos_bps_burst_seconds 1 config
rbd_qos_bps_limit 0 config
rbd_qos_exclude_ops 0 config
rbd_qos_iops_burst 0 config
rbd_qos_iops_burst_seconds 1 config
rbd_qos_iops_limit 0 config
rbd_qos_read_bps_burst 0 config
rbd_qos_read_bps_burst_seconds 1 config
rbd_qos_read_bps_limit 0 config
rbd_qos_read_iops_burst 0 config
rbd_qos_read_iops_burst_seconds 1 config
rbd_qos_read_iops_limit 0 config
rbd_qos_schedule_tick_min 50 config
rbd_qos_write_bps_burst 0 config
rbd_qos_write_bps_burst_seconds 1 config
rbd_qos_write_bps_limit 0 config
rbd_qos_write_iops_burst 0 config
rbd_qos_write_iops_burst_seconds 1 config
rbd_qos_write_iops_limit 0 config
rbd_quiesce_notification_attempts 10 config
rbd_read_from_replica_policy default config
rbd_readahead_disable_after_bytes 52428800 config
rbd_readahead_max_bytes 524288 config
rbd_readahead_trigger_requests 10 config
rbd_request_timed_out_seconds 30 config
rbd_skip_partial_discard true config
rbd_sparse_read_threshold_bytes 65536 config
Cheers
Reto Gysi
Am Mo., 17. Apr. 2023 um 17:31 Uhr schrieb Ilya Dryomov <idryomov(a)gmail.com
:
On Mon, Apr 17, 2023 at 2:01 PM Reto Gysi
<rlgysi(a)gmail.com> wrote:
Dear Ceph Users,
After upgrading from version 17.2.5 to 17.2.6 I no longer seem to be able
to create snapshots
of images that have an erasure coded datapool.
root@zephir:~# rbd snap create ceph-dev@backup_20230417
Creating snap: 10% complete...failed.
rbd: failed to create snapshot: (95) Operation not supported
root@zephir:~# rbd info ceph-dev
rbd image 'ceph-dev':
size 10 GiB in 2560 objects
order 22 (4 MiB objects)
snapshot_count: 11
id: d2f3d287f13c7b
data_pool: ecpool_hdd
block_name_prefix: rbd_data.7.d2f3d287f13c7b
format: 2
features: layering, exclusive-lock, object-map, fast-diff,
deep-flatten, data-pool
op_features:
flags:
create_timestamp: Wed Nov 23 17:01:03 2022
access_timestamp: Sun Apr 16 17:20:58 2023
modify_timestamp: Wed Nov 23 17:01:03 2022
root@zephir:~#
Before the upgrade I was able to create snapshots of this pool:
SNAPID NAME SIZE PROTECTED
TIMESTAMP
1538 ceph-dev_2023-03-05T02:00:09.030+01:00 10 GiB Sun Mar
5 02:00:14 2023
1545 ceph-dev_2023-03-06T02:00:03.832+01:00 10 GiB Mon Mar
6 02:00:05 2023
1903 ceph-dev_2023-04-05T03:22:01.315+02:00 10 GiB Wed Apr
5 03:22:02 2023
1909 ceph-dev_2023-04-05T03:35:56.748+02:00 10 GiB Wed Apr
5 03:35:57 2023
1915 ceph-dev_2023-04-05T03:37:23.778+02:00 10 GiB Wed Apr
5 03:37:24 2023
1930 ceph-dev_2023-04-06T02:00:06.159+02:00 10 GiB Thu Apr
6 02:00:07 2023
1940 ceph-dev_2023-04-07T02:00:05.913+02:00 10 GiB Fri Apr
7 02:00:06 2023
1952 ceph-dev_2023-04-08T02:00:06.534+02:00 10 GiB Sat Apr
8 02:00:07 2023
1964 ceph-dev_2023-04-09T02:00:06.430+02:00 10 GiB Sun Apr
9 02:00:07 2023
2003 ceph-dev_2023-04-11T02:00:09.750+02:00 10 GiB Tue Apr
11 02:00:10 2023
2014 ceph-dev_2023-04-12T02:00:09.528+02:00 10 GiB Wed Apr
12 02:00:10 2023
root@zephir:~#
I have looked through the release notes of 17.2.6 but couldn't find
anything obvious regarding rbd and ec pools.
Does anyone else have this problem?
Do I need to change some config setting, or was this feature disabled or
is
it a bug?
Hi Reto,
Nothing was disabled and no config changes are expected. This should just
work.
What is the output of "rbd status" for that image?
Can you reproduce with "--debug-ms 1 --debug-rbd 20" appended to "rbd
snap create" command and attach a file with the output?
Thanks,
Ilya