Hi Andre,

I once faced the same problem. It turns out that ceph need to scan every object 
in the image when deleting it, if object map is not enabled. This will take 
years on such a huge image. I ended up deleted the whole pool to get rid of the 
huge image.

Maybe you can scan all the objects in the pool and remove all objects relevant 
to this image manually, but I don’t know how.

> 在 2020年12月16日,15:07,Andre Gebers <andre.geb...@gridqube.com> 写道:
> 
> Hi,
> 
> I'm running a 15.2.4 test cluster in a rook-ceph environment. The cluster is 
> reporting HEALTH_OK but it seems it is stuck removing an image. Last section 
> of 'ceph status' output:
> 
>  progress:
>    Removing image replicapool/43def5e07bf47 from trash (6h)
>      [............................] (remaining: 32y)
> 
> This is now going for a couple of weeks and I was wondering if there is a way 
> to speed it up? The cluster doesn't seem to be doing much judging from the 
> system load.
> 
> I've created this largish image to test what is possible with the setup but 
> how do I get it out of the trash now?
> 
> # rbd info --image-id 43def5e07bf47 -p replicapool
> rbd image 'csi-vol-cfaa1b00-1711-11eb-b9c9-2aa51e1e24e5':
>        size 1 EiB in 274877906944 objects
>        order 22 (4 MiB objects)
>        snapshot_count: 0
>        id: 43def5e07bf47
>        block_name_prefix: rbd_data.43def5e07bf47
>        format: 2
>        features: layering
>        op_features:
>        flags:
>        create_timestamp: Sun Oct 25 22:31:23 2020
>        access_timestamp: Sun Oct 25 22:31:23 2020
>        modify_timestamp: Sun Oct 25 22:31:23 2020
> 
> Any pointers how to resolve this issue are much appreciated.
> 
> Regards
> Andre
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to