Perhaps setting the object-map feature on the image, and/or running rbd object-map rebuild? Though I suspect that might perform an equivalent process and take just as long?
> On Dec 15, 2020, at 11:49 PM, 胡 玮文 <huw...@outlook.com> wrote: > > Hi Andre, > > I once faced the same problem. It turns out that ceph need to scan every > object in the image when deleting it, if object map is not enabled. This will > take years on such a huge image. I ended up deleted the whole pool to get rid > of the huge image. > > Maybe you can scan all the objects in the pool and remove all objects > relevant to this image manually, but I don’t know how. > >> 在 2020年12月16日,15:07,Andre Gebers <andre.geb...@gridqube.com> 写道: >> >> Hi, >> >> I'm running a 15.2.4 test cluster in a rook-ceph environment. The cluster is >> reporting HEALTH_OK but it seems it is stuck removing an image. Last section >> of 'ceph status' output: >> >> progress: >> Removing image replicapool/43def5e07bf47 from trash (6h) >> [............................] (remaining: 32y) >> >> This is now going for a couple of weeks and I was wondering if there is a >> way to speed it up? The cluster doesn't seem to be doing much judging from >> the system load. >> >> I've created this largish image to test what is possible with the setup but >> how do I get it out of the trash now? >> >> # rbd info --image-id 43def5e07bf47 -p replicapool >> rbd image 'csi-vol-cfaa1b00-1711-11eb-b9c9-2aa51e1e24e5': >> size 1 EiB in 274877906944 objects >> order 22 (4 MiB objects) >> snapshot_count: 0 >> id: 43def5e07bf47 >> block_name_prefix: rbd_data.43def5e07bf47 >> format: 2 >> features: layering >> op_features: >> flags: >> create_timestamp: Sun Oct 25 22:31:23 2020 >> access_timestamp: Sun Oct 25 22:31:23 2020 >> modify_timestamp: Sun Oct 25 22:31:23 2020 >> >> Any pointers how to resolve this issue are much appreciated. >> >> Regards >> Andre >> _______________________________________________ >> ceph-users mailing list -- ceph-users@ceph.io >> To unsubscribe send an email to ceph-users-le...@ceph.io > _______________________________________________ > ceph-users mailing list -- ceph-users@ceph.io > To unsubscribe send an email to ceph-users-le...@ceph.io _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io