I've just tried this on 17.2.6 and it worked fine....
On 17/04/2023 12:57, Reto Gysi wrote:
Dear Ceph Users,
After upgrading from version 17.2.5 to 17.2.6 I no longer seem to be able
to create snapshots
of images that have an erasure coded datapool.
root@zephir:~# rbd snap create ceph-dev@backup_20230417
Creating snap: 10% complete...failed.
rbd: failed to create snapshot: (95) Operation not supported
root@zephir:~# rbd info ceph-dev
rbd image 'ceph-dev':
size 10 GiB in 2560 objects
order 22 (4 MiB objects)
snapshot_count: 11
id: d2f3d287f13c7b
data_pool: ecpool_hdd
block_name_prefix: rbd_data.7.d2f3d287f13c7b
format: 2
features: layering, exclusive-lock, object-map, fast-diff,
deep-flatten, data-pool
op_features:
flags:
create_timestamp: Wed Nov 23 17:01:03 2022
access_timestamp: Sun Apr 16 17:20:58 2023
modify_timestamp: Wed Nov 23 17:01:03 2022
root@zephir:~#
Before the upgrade I was able to create snapshots of this pool:
SNAPID NAME SIZE PROTECTED
TIMESTAMP
1538 ceph-dev_2023-03-05T02:00:09.030+01:00 10 GiB Sun Mar
5 02:00:14 2023
1545 ceph-dev_2023-03-06T02:00:03.832+01:00 10 GiB Mon Mar
6 02:00:05 2023
1903 ceph-dev_2023-04-05T03:22:01.315+02:00 10 GiB Wed Apr
5 03:22:02 2023
1909 ceph-dev_2023-04-05T03:35:56.748+02:00 10 GiB Wed Apr
5 03:35:57 2023
1915 ceph-dev_2023-04-05T03:37:23.778+02:00 10 GiB Wed Apr
5 03:37:24 2023
1930 ceph-dev_2023-04-06T02:00:06.159+02:00 10 GiB Thu Apr
6 02:00:07 2023
1940 ceph-dev_2023-04-07T02:00:05.913+02:00 10 GiB Fri Apr
7 02:00:06 2023
1952 ceph-dev_2023-04-08T02:00:06.534+02:00 10 GiB Sat Apr
8 02:00:07 2023
1964 ceph-dev_2023-04-09T02:00:06.430+02:00 10 GiB Sun Apr
9 02:00:07 2023
2003 ceph-dev_2023-04-11T02:00:09.750+02:00 10 GiB Tue Apr
11 02:00:10 2023
2014 ceph-dev_2023-04-12T02:00:09.528+02:00 10 GiB Wed Apr
12 02:00:10 2023
root@zephir:~#
I have looked through the release notes of 17.2.6 but couldn't find
anything obvious regarding rbd and ec pools.
Does anyone else have this problem?
Do I need to change some config setting, or was this feature disabled or is
it a bug?
Ceph version info:
root@zephir:~# ceph orch upgrade check --ceph_version 17.2.6
{
"needs_update": {},
"non_ceph_image_daemons": [
"promtail.debian",
"node-exporter.debian",
"promtail.zephir",
"grafana.zephir",
"node-exporter.zephir",
"prometheus.zephir",
"loki.zephir",
"alertmanager.zephir"
],
"target_digest": "
quay.io/ceph/ceph@sha256:1161e35e4e02cf377c93b913ce78773f8413f5a8d7c5eaee4b4773a4f9dd6635",
"target_id":
"9cea3956c04b2d889b91b58f957577fcb4eacd3852df073e3e2567f159fcdbf8",
"target_name": "quay.io/ceph/ceph:v17.2.6",
"target_version": "ceph version 17.2.6
(d7ff0d10654d2280e08f1ab989c7cdf3064446a5) quincy (stable)",
"up_to_date": [
"iscsi.rbd.debian.ijztzu",
"mds.jellyfin.debian.dcsocv",
"mon.debian",
"osd.13",
"osd.6",
"mds.backups.debian.runngh",
"mds.rgysi.debian.uhgqen",
"crash.debian",
"mgr.debian.sookxi",
"iscsi.rbd.zephir.viqahd",
"osd.1",
"mds.jellyfin.zephir.iqywsn",
"osd.12",
"osd.7",
"osd.2",
"crash.zephir",
"rgw.default.zephir.jqmick",
"mds.backups.zephir.ygigch",
"osd.0",
"osd.4",
"mon.zephir",
"mgr.zephir.enywvy",
"mds.rgysi.zephir.diylss",
"osd.3",
"osd.10",
"osd.5",
"osd.8",
"osd.11"
]
}
root@zephir:~#
root@zephir:~# rbd --version
ceph version 17.2.6 (d7ff0d10654d2280e08f1ab989c7cdf3064446a5) quincy
(stable)
root@zephir:~#
Cheers
Reto Gysi
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io