Perhaps setting the object-map feature on the image, and/or running rbd object-map
rebuild? Though I suspect that might perform an equivalent process and take just as
long?
On Dec 15, 2020, at 11:49 PM, 胡 玮文
<huww98(a)outlook.com> wrote:
Hi Andre,
I once faced the same problem. It turns out that ceph need to scan every object in the
image when deleting it, if object map is not enabled. This will take years on such a huge
image. I ended up deleted the whole pool to get rid of the huge image.
Maybe you can scan all the objects in the pool and remove all objects relevant to this
image manually, but I don’t know how.
在 2020年12月16日,15:07,Andre Gebers
<andre.gebers(a)gridqube.com> 写道:
Hi,
I'm running a 15.2.4 test cluster in a rook-ceph environment. The cluster is
reporting HEALTH_OK but it seems it is stuck removing an image. Last section of 'ceph
status' output:
progress:
Removing image replicapool/43def5e07bf47 from trash (6h)
[............................] (remaining: 32y)
This is now going for a couple of weeks and I was wondering if there is a way to speed it
up? The cluster doesn't seem to be doing much judging from the system load.
I've created this largish image to test what is possible with the setup but how do I
get it out of the trash now?
# rbd info --image-id 43def5e07bf47 -p replicapool
rbd image 'csi-vol-cfaa1b00-1711-11eb-b9c9-2aa51e1e24e5':
size 1 EiB in 274877906944 objects
order 22 (4 MiB objects)
snapshot_count: 0
id: 43def5e07bf47
block_name_prefix: rbd_data.43def5e07bf47
format: 2
features: layering
op_features:
flags:
create_timestamp: Sun Oct 25 22:31:23 2020
access_timestamp: Sun Oct 25 22:31:23 2020
modify_timestamp: Sun Oct 25 22:31:23 2020
Any pointers how to resolve this issue are much appreciated.
Regards
Andre
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io