Re: [ceph-users] Deleting an rbd image hangs

2018-05-09 Thread David Turner
Yeah, I was about to suggest looking up all currently existing rbd IDs and
snapshot IDs, compare to rados ls and remove the objects that exist for
rbds and snapshots not reported by the cluster.

On Wed, May 9, 2018, 8:38 AM Jason Dillaman  wrote:

> On Tue, May 8, 2018 at 2:31 PM,   wrote:
> > Hello Jason,
> >
> >
> > Am 8. Mai 2018 15:30:34 MESZ schrieb Jason Dillaman  >:
> >>Perhaps the image had associated snapshots? Deleting the object
> >>doesn't delete the associated snapshots so those objects will remain
> >>until the snapshot is removed. However, if you have removed the RBD
> >>header, the snapshot id is now gone.
> >>
> >
> > Hmm... that makes me curious...
> >
> > So when i have a vm-image (rbd) on ceph and am doing  One or more
> Snapshots from this Image i *must have* to delete the snapshot(s) at
> First completely before i delete the origin Image?
>
> Yup, the rbd CLI (and related librbd API helpers for removing images)
> will not let you delete an image that has snapshots for this very
> reason.
>
> > How can we then get rid of this orphaned objects when we accidentaly
> have deleted the origin Image First?
>
> Unfortunately, there isn't any available CLI tooling to let if you
> delete a "self-managed" snapshot even if you could determine the
> correct snapshot ids that are no longer in-use. If you are comfortable
> building a custom C/C++ librados application to clean up your pool,
> you could first generate a list of all known in-use snapshot IDs by
> collecting all the "snapshot_XYZ" keys on all "rbd_header.ABC" objects
> in the pool (where XYZ is the snapshot id in hex and ABC is each
> image's unique id) and cross-referencing them w/ the pool's
> "removed_snaps" output from "ceph osd pool ls detail".
>
> > Thanks if you have a Bit of Time to clarify me/us :)
> >
> > - Mehmet
> >
> >>On Tue, May 8, 2018 at 12:29 AM, Eugen Block  wrote:
> >>> Hi,
> >>>
> >>> I have a similar issue and would also need some advice how to get rid
> >>of the
> >>> already deleted files.
> >>>
> >>> Ceph is our OpenStack backend and there was a nova clone without
> >>parent
> >>> information. Apparently, the base image had been deleted without a
> >>warning
> >>> or anything although there were existing clones.
> >>> Anyway, I tried to delete the respective rbd_data and _header files
> >>as
> >>> described in [1]. There were about 700 objects to be deleted, but 255
> >>> objects remained according to the 'rados -p pool ls' command. The
> >>attempt to
> >>> delete the rest (again) resulted (and still results) in "No such file
> >>or
> >>> directory". After about half an hour later one more object vanished
> >>> (rbd_header file), there are now still 254 objects left in the pool.
> >>First I
> >>> thought maybe Ceph will cleanup itself, it just takes some time, but
> >>this
> >>> was weeks ago and the number of objects has not changed since then.
> >>>
> >>> I would really appreciate any help.
> >>>
> >>> Regards,
> >>> Eugen
> >>>
> >>>
> >>> Zitat von Jan Marquardt :
> >>>
> >>>
>  Am 30.04.18 um 09:26 schrieb Jan Marquardt:
> >
> > Am 27.04.18 um 20:48 schrieb David Turner:
> >>
> >> This old [1] blog post about removing super large RBDs is not
> >>relevant
> >> if you're using object map on the RBDs, however it's method to
> >>manually
> >> delete an RBD is still valid.  You can see if this works for you
> >>to
> >> manually remove the problem RBD you're having.
> >
> >
> > I followed the instructions, but it seems that 'rados -p rbd ls |
> >>grep
> > '^rbd_data.221bf2eb141f2.' | xargs -n 200  rados -p rbd rm' gets
> >>stuck,
> > too. It's running since Friday and still not finished. The rbd
> >>image
> > is/was about 1 TB large.
> >
> > Until now the only output was:
> > error removing rbd>rbd_data.221bf2eb141f2.51d2: (2) No
> >>such
> > file or directory
> > error removing rbd>rbd_data.221bf2eb141f2.e3f2: (2) No
> >>such
> > file or directory
> 
> 
>  I am still trying to get rid of this. 'rados -p rbd ls' still shows
> >>a
>  lot of objects beginning with rbd_data.221bf2eb141f2, but if I try
> >>to
>  delete them with 'rados -p rbd rm ' it says 'No such file or
>  directory'. This is not the behaviour I'd expect. Any ideas?
> 
>  Besides this rbd_data.221bf2eb141f2.00016379 is still
> >>causing
>  the OSDs crashing, which leaves the cluster unusable for us at the
>  moment. Even if it's just a proof of concept, I'd like to get this
> >>fixed
>  without destroying the whole cluster.
> 
> >>
> >> [1]
> >>http://cephnotes.ksperis.com/blog/2014/07/04/remove-big-rbd-image
> >>
> >> On Thu, Apr 26, 2018 at 9:25 AM Jan Marquardt  >> > wrote:
> >>
> >> Hi,
> >>
> >> I am 

Re: [ceph-users] Deleting an rbd image hangs

2018-05-09 Thread Jason Dillaman
On Tue, May 8, 2018 at 2:31 PM,   wrote:
> Hello Jason,
>
>
> Am 8. Mai 2018 15:30:34 MESZ schrieb Jason Dillaman :
>>Perhaps the image had associated snapshots? Deleting the object
>>doesn't delete the associated snapshots so those objects will remain
>>until the snapshot is removed. However, if you have removed the RBD
>>header, the snapshot id is now gone.
>>
>
> Hmm... that makes me curious...
>
> So when i have a vm-image (rbd) on ceph and am doing  One or more Snapshots 
> from this Image i *must have* to delete the snapshot(s) at First 
> completely before i delete the origin Image?

Yup, the rbd CLI (and related librbd API helpers for removing images)
will not let you delete an image that has snapshots for this very
reason.

> How can we then get rid of this orphaned objects when we accidentaly have 
> deleted the origin Image First?

Unfortunately, there isn't any available CLI tooling to let if you
delete a "self-managed" snapshot even if you could determine the
correct snapshot ids that are no longer in-use. If you are comfortable
building a custom C/C++ librados application to clean up your pool,
you could first generate a list of all known in-use snapshot IDs by
collecting all the "snapshot_XYZ" keys on all "rbd_header.ABC" objects
in the pool (where XYZ is the snapshot id in hex and ABC is each
image's unique id) and cross-referencing them w/ the pool's
"removed_snaps" output from "ceph osd pool ls detail".

> Thanks if you have a Bit of Time to clarify me/us :)
>
> - Mehmet
>
>>On Tue, May 8, 2018 at 12:29 AM, Eugen Block  wrote:
>>> Hi,
>>>
>>> I have a similar issue and would also need some advice how to get rid
>>of the
>>> already deleted files.
>>>
>>> Ceph is our OpenStack backend and there was a nova clone without
>>parent
>>> information. Apparently, the base image had been deleted without a
>>warning
>>> or anything although there were existing clones.
>>> Anyway, I tried to delete the respective rbd_data and _header files
>>as
>>> described in [1]. There were about 700 objects to be deleted, but 255
>>> objects remained according to the 'rados -p pool ls' command. The
>>attempt to
>>> delete the rest (again) resulted (and still results) in "No such file
>>or
>>> directory". After about half an hour later one more object vanished
>>> (rbd_header file), there are now still 254 objects left in the pool.
>>First I
>>> thought maybe Ceph will cleanup itself, it just takes some time, but
>>this
>>> was weeks ago and the number of objects has not changed since then.
>>>
>>> I would really appreciate any help.
>>>
>>> Regards,
>>> Eugen
>>>
>>>
>>> Zitat von Jan Marquardt :
>>>
>>>
 Am 30.04.18 um 09:26 schrieb Jan Marquardt:
>
> Am 27.04.18 um 20:48 schrieb David Turner:
>>
>> This old [1] blog post about removing super large RBDs is not
>>relevant
>> if you're using object map on the RBDs, however it's method to
>>manually
>> delete an RBD is still valid.  You can see if this works for you
>>to
>> manually remove the problem RBD you're having.
>
>
> I followed the instructions, but it seems that 'rados -p rbd ls |
>>grep
> '^rbd_data.221bf2eb141f2.' | xargs -n 200  rados -p rbd rm' gets
>>stuck,
> too. It's running since Friday and still not finished. The rbd
>>image
> is/was about 1 TB large.
>
> Until now the only output was:
> error removing rbd>rbd_data.221bf2eb141f2.51d2: (2) No
>>such
> file or directory
> error removing rbd>rbd_data.221bf2eb141f2.e3f2: (2) No
>>such
> file or directory


 I am still trying to get rid of this. 'rados -p rbd ls' still shows
>>a
 lot of objects beginning with rbd_data.221bf2eb141f2, but if I try
>>to
 delete them with 'rados -p rbd rm ' it says 'No such file or
 directory'. This is not the behaviour I'd expect. Any ideas?

 Besides this rbd_data.221bf2eb141f2.00016379 is still
>>causing
 the OSDs crashing, which leaves the cluster unusable for us at the
 moment. Even if it's just a proof of concept, I'd like to get this
>>fixed
 without destroying the whole cluster.

>>
>> [1]
>>http://cephnotes.ksperis.com/blog/2014/07/04/remove-big-rbd-image
>>
>> On Thu, Apr 26, 2018 at 9:25 AM Jan Marquardt > > wrote:
>>
>> Hi,
>>
>> I am currently trying to delete an rbd image which is
>>seemingly
>> causing
>> our OSDs to crash, but it always gets stuck at 3%.
>>
>> root@ceph4:~# rbd rm noc_tobedeleted
>> Removing image: 3% complete...
>>
>> Is there any way to force the deletion? Any other advices?
>>
>> Best Regards
>>
>> Jan




 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 

Re: [ceph-users] Deleting an rbd image hangs

2018-05-08 Thread ceph
Hello Jason,
  

Am 8. Mai 2018 15:30:34 MESZ schrieb Jason Dillaman :
>Perhaps the image had associated snapshots? Deleting the object
>doesn't delete the associated snapshots so those objects will remain
>until the snapshot is removed. However, if you have removed the RBD
>header, the snapshot id is now gone.
>

Hmm... that makes me curious...

So when i have a vm-image (rbd) on ceph and am doing  One or more Snapshots 
from this Image i *must have* to delete the snapshot(s) at First completely 
before i delete the origin Image?

How can we then get rid of this orphaned objects when we accidentaly have 
deleted the origin Image First?

Thanks if you have a Bit of Time to clarify me/us :)

- Mehmet

>On Tue, May 8, 2018 at 12:29 AM, Eugen Block  wrote:
>> Hi,
>>
>> I have a similar issue and would also need some advice how to get rid
>of the
>> already deleted files.
>>
>> Ceph is our OpenStack backend and there was a nova clone without
>parent
>> information. Apparently, the base image had been deleted without a
>warning
>> or anything although there were existing clones.
>> Anyway, I tried to delete the respective rbd_data and _header files
>as
>> described in [1]. There were about 700 objects to be deleted, but 255
>> objects remained according to the 'rados -p pool ls' command. The
>attempt to
>> delete the rest (again) resulted (and still results) in "No such file
>or
>> directory". After about half an hour later one more object vanished
>> (rbd_header file), there are now still 254 objects left in the pool.
>First I
>> thought maybe Ceph will cleanup itself, it just takes some time, but
>this
>> was weeks ago and the number of objects has not changed since then.
>>
>> I would really appreciate any help.
>>
>> Regards,
>> Eugen
>>
>>
>> Zitat von Jan Marquardt :
>>
>>
>>> Am 30.04.18 um 09:26 schrieb Jan Marquardt:

 Am 27.04.18 um 20:48 schrieb David Turner:
>
> This old [1] blog post about removing super large RBDs is not
>relevant
> if you're using object map on the RBDs, however it's method to
>manually
> delete an RBD is still valid.  You can see if this works for you
>to
> manually remove the problem RBD you're having.


 I followed the instructions, but it seems that 'rados -p rbd ls |
>grep
 '^rbd_data.221bf2eb141f2.' | xargs -n 200  rados -p rbd rm' gets
>stuck,
 too. It's running since Friday and still not finished. The rbd
>image
 is/was about 1 TB large.

 Until now the only output was:
 error removing rbd>rbd_data.221bf2eb141f2.51d2: (2) No
>such
 file or directory
 error removing rbd>rbd_data.221bf2eb141f2.e3f2: (2) No
>such
 file or directory
>>>
>>>
>>> I am still trying to get rid of this. 'rados -p rbd ls' still shows
>a
>>> lot of objects beginning with rbd_data.221bf2eb141f2, but if I try
>to
>>> delete them with 'rados -p rbd rm ' it says 'No such file or
>>> directory'. This is not the behaviour I'd expect. Any ideas?
>>>
>>> Besides this rbd_data.221bf2eb141f2.00016379 is still
>causing
>>> the OSDs crashing, which leaves the cluster unusable for us at the
>>> moment. Even if it's just a proof of concept, I'd like to get this
>fixed
>>> without destroying the whole cluster.
>>>
>
> [1]
>http://cephnotes.ksperis.com/blog/2014/07/04/remove-big-rbd-image
>
> On Thu, Apr 26, 2018 at 9:25 AM Jan Marquardt  > wrote:
>
> Hi,
>
> I am currently trying to delete an rbd image which is
>seemingly
> causing
> our OSDs to crash, but it always gets stuck at 3%.
>
> root@ceph4:~# rbd rm noc_tobedeleted
> Removing image: 3% complete...
>
> Is there any way to force the deletion? Any other advices?
>
> Best Regards
>
> Jan
>>>
>>>
>>>
>>>
>>> ___
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>>
>>
>> ___
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Deleting an rbd image hangs

2018-05-08 Thread Jason Dillaman
Perhaps the image had associated snapshots? Deleting the object
doesn't delete the associated snapshots so those objects will remain
until the snapshot is removed. However, if you have removed the RBD
header, the snapshot id is now gone.

On Tue, May 8, 2018 at 12:29 AM, Eugen Block  wrote:
> Hi,
>
> I have a similar issue and would also need some advice how to get rid of the
> already deleted files.
>
> Ceph is our OpenStack backend and there was a nova clone without parent
> information. Apparently, the base image had been deleted without a warning
> or anything although there were existing clones.
> Anyway, I tried to delete the respective rbd_data and _header files as
> described in [1]. There were about 700 objects to be deleted, but 255
> objects remained according to the 'rados -p pool ls' command. The attempt to
> delete the rest (again) resulted (and still results) in "No such file or
> directory". After about half an hour later one more object vanished
> (rbd_header file), there are now still 254 objects left in the pool. First I
> thought maybe Ceph will cleanup itself, it just takes some time, but this
> was weeks ago and the number of objects has not changed since then.
>
> I would really appreciate any help.
>
> Regards,
> Eugen
>
>
> Zitat von Jan Marquardt :
>
>
>> Am 30.04.18 um 09:26 schrieb Jan Marquardt:
>>>
>>> Am 27.04.18 um 20:48 schrieb David Turner:

 This old [1] blog post about removing super large RBDs is not relevant
 if you're using object map on the RBDs, however it's method to manually
 delete an RBD is still valid.  You can see if this works for you to
 manually remove the problem RBD you're having.
>>>
>>>
>>> I followed the instructions, but it seems that 'rados -p rbd ls | grep
>>> '^rbd_data.221bf2eb141f2.' | xargs -n 200  rados -p rbd rm' gets stuck,
>>> too. It's running since Friday and still not finished. The rbd image
>>> is/was about 1 TB large.
>>>
>>> Until now the only output was:
>>> error removing rbd>rbd_data.221bf2eb141f2.51d2: (2) No such
>>> file or directory
>>> error removing rbd>rbd_data.221bf2eb141f2.e3f2: (2) No such
>>> file or directory
>>
>>
>> I am still trying to get rid of this. 'rados -p rbd ls' still shows a
>> lot of objects beginning with rbd_data.221bf2eb141f2, but if I try to
>> delete them with 'rados -p rbd rm ' it says 'No such file or
>> directory'. This is not the behaviour I'd expect. Any ideas?
>>
>> Besides this rbd_data.221bf2eb141f2.00016379 is still causing
>> the OSDs crashing, which leaves the cluster unusable for us at the
>> moment. Even if it's just a proof of concept, I'd like to get this fixed
>> without destroying the whole cluster.
>>

 [1] http://cephnotes.ksperis.com/blog/2014/07/04/remove-big-rbd-image

 On Thu, Apr 26, 2018 at 9:25 AM Jan Marquardt > wrote:

 Hi,

 I am currently trying to delete an rbd image which is seemingly
 causing
 our OSDs to crash, but it always gets stuck at 3%.

 root@ceph4:~# rbd rm noc_tobedeleted
 Removing image: 3% complete...

 Is there any way to force the deletion? Any other advices?

 Best Regards

 Jan
>>
>>
>>
>>
>> ___
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
>
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



-- 
Jason
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Deleting an rbd image hangs

2018-05-08 Thread Eugen Block

Hi,

I have a similar issue and would also need some advice how to get rid  
of the already deleted files.


Ceph is our OpenStack backend and there was a nova clone without  
parent information. Apparently, the base image had been deleted  
without a warning or anything although there were existing clones.
Anyway, I tried to delete the respective rbd_data and _header files as  
described in [1]. There were about 700 objects to be deleted, but 255  
objects remained according to the 'rados -p pool ls' command. The  
attempt to delete the rest (again) resulted (and still results) in "No  
such file or directory". After about half an hour later one more  
object vanished (rbd_header file), there are now still 254 objects  
left in the pool. First I thought maybe Ceph will cleanup itself, it  
just takes some time, but this was weeks ago and the number of objects  
has not changed since then.


I would really appreciate any help.

Regards,
Eugen


Zitat von Jan Marquardt :


Am 30.04.18 um 09:26 schrieb Jan Marquardt:

Am 27.04.18 um 20:48 schrieb David Turner:

This old [1] blog post about removing super large RBDs is not relevant
if you're using object map on the RBDs, however it's method to manually
delete an RBD is still valid.  You can see if this works for you to
manually remove the problem RBD you're having.


I followed the instructions, but it seems that 'rados -p rbd ls | grep
'^rbd_data.221bf2eb141f2.' | xargs -n 200  rados -p rbd rm' gets stuck,
too. It's running since Friday and still not finished. The rbd image
is/was about 1 TB large.

Until now the only output was:
error removing rbd>rbd_data.221bf2eb141f2.51d2: (2) No such
file or directory
error removing rbd>rbd_data.221bf2eb141f2.e3f2: (2) No such
file or directory


I am still trying to get rid of this. 'rados -p rbd ls' still shows a
lot of objects beginning with rbd_data.221bf2eb141f2, but if I try to
delete them with 'rados -p rbd rm ' it says 'No such file or
directory'. This is not the behaviour I'd expect. Any ideas?

Besides this rbd_data.221bf2eb141f2.00016379 is still causing
the OSDs crashing, which leaves the cluster unusable for us at the
moment. Even if it's just a proof of concept, I'd like to get this fixed
without destroying the whole cluster.



[1] http://cephnotes.ksperis.com/blog/2014/07/04/remove-big-rbd-image

On Thu, Apr 26, 2018 at 9:25 AM Jan Marquardt > wrote:

Hi,

I am currently trying to delete an rbd image which is seemingly causing
our OSDs to crash, but it always gets stuck at 3%.

root@ceph4:~# rbd rm noc_tobedeleted
Removing image: 3% complete...

Is there any way to force the deletion? Any other advices?

Best Regards

Jan




___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Deleting an rbd image hangs

2018-05-07 Thread Jan Marquardt

Am 30.04.18 um 09:26 schrieb Jan Marquardt:
> Am 27.04.18 um 20:48 schrieb David Turner:
>> This old [1] blog post about removing super large RBDs is not relevant
>> if you're using object map on the RBDs, however it's method to manually
>> delete an RBD is still valid.  You can see if this works for you to
>> manually remove the problem RBD you're having.
> 
> I followed the instructions, but it seems that 'rados -p rbd ls | grep
> '^rbd_data.221bf2eb141f2.' | xargs -n 200  rados -p rbd rm' gets stuck,
> too. It's running since Friday and still not finished. The rbd image
> is/was about 1 TB large.
> 
> Until now the only output was:
> error removing rbd>rbd_data.221bf2eb141f2.51d2: (2) No such
> file or directory
> error removing rbd>rbd_data.221bf2eb141f2.e3f2: (2) No such
> file or directory

I am still trying to get rid of this. 'rados -p rbd ls' still shows a
lot of objects beginning with rbd_data.221bf2eb141f2, but if I try to
delete them with 'rados -p rbd rm ' it says 'No such file or
directory'. This is not the behaviour I'd expect. Any ideas?

Besides this rbd_data.221bf2eb141f2.00016379 is still causing
the OSDs crashing, which leaves the cluster unusable for us at the
moment. Even if it's just a proof of concept, I'd like to get this fixed
without destroying the whole cluster.

>>
>> [1] http://cephnotes.ksperis.com/blog/2014/07/04/remove-big-rbd-image
>>
>> On Thu, Apr 26, 2018 at 9:25 AM Jan Marquardt > > wrote:
>>
>> Hi,
>>
>> I am currently trying to delete an rbd image which is seemingly causing
>> our OSDs to crash, but it always gets stuck at 3%.
>>
>> root@ceph4:~# rbd rm noc_tobedeleted
>> Removing image: 3% complete...
>>
>> Is there any way to force the deletion? Any other advices?
>>
>> Best Regards
>>
>> Jan



___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Deleting an rbd image hangs

2018-04-30 Thread Jan Marquardt
Am 27.04.18 um 22:33 schrieb Jason Dillaman:
> Do you have any reason for why the OSDs crash? Anything the logs? Can
> you provide an "rbd info noc_tobedeleted"?

The reason why they are crashing is this assert:
https://github.com/ceph/ceph/blob/luminous/src/osd/PrimaryLogPG.cc#L353

With debug 20 we see this right before the OSD crashes:

2018-04-24 13:59:38.047697 7f929ba0d700 20 osd.4 pg_epoch: 144994
pg[0.103( v 140091'469328 (125640'467824,140091'469328] lb
0:c0e04acc:::rbd_data.221bf2eb141f2.00016379:head (bitwise)
local-lis/les=137681/137682 n=9535 ec=115/115 lis/c 144979/49591 les/c/f
144980/49596/0 144978/144979/144979) [4,17,2]/[2,17] r=-1 lpr=144979
pi=[49591,144979)/3 luod=0'0 crt=140091'469328 lcod 0'0 active+remapped]
 snapset 0=[]:[] legacy_snaps []

2018-04-24 16:34:54.558159 7f1c40e32700 20 osd.11 pg_epoch: 145549
pg[0.103( v 140091'469328 (125640'467824,140091'469328] lb
0:c0e04acc:::rbd_data.221bf2eb141f2.00016379:head (bitwise)
local-lis/les=138310/138311 n=9535 ec=115/115 lis/c 145548/49591 les/c/f
145549/49596/0 145547/145548/145548) [11,17,2]/[2,17] r=-1 lpr=145548
pi=[49591,145548)/3 luod=0'0 crt=140091'469328 lcod 0'0 active+remapped]
 snapset 0=[]:[] legacy_snaps []

Which is caused from this code:
https://github.com/ceph/ceph/blob/luminous/src/osd/PrimaryLogPG.cc#L349-L350

Unfortunately rbd info is not available anymore for this image, because
I already followed the instructions under
http://cephnotes.ksperis.com/blog/2014/07/04/remove-big-rbd-image until
'Remove all rbd data', which seems to be hanging, too.

> On Thu, Apr 26, 2018 at 9:24 AM, Jan Marquardt  wrote:
>> Hi,
>>
>> I am currently trying to delete an rbd image which is seemingly causing
>> our OSDs to crash, but it always gets stuck at 3%.
>>
>> root@ceph4:~# rbd rm noc_tobedeleted
>> Removing image: 3% complete...
>>
>> Is there any way to force the deletion? Any other advices?
>>
>> Best Regards
>>
>> Jan
>> ___
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> 
> 
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Deleting an rbd image hangs

2018-04-30 Thread Jan Marquardt
Am 27.04.18 um 20:48 schrieb David Turner:
> This old [1] blog post about removing super large RBDs is not relevant
> if you're using object map on the RBDs, however it's method to manually
> delete an RBD is still valid.  You can see if this works for you to
> manually remove the problem RBD you're having.

I followed the instructions, but it seems that 'rados -p rbd ls | grep
'^rbd_data.221bf2eb141f2.' | xargs -n 200  rados -p rbd rm' gets stuck,
too. It's running since Friday and still not finished. The rbd image
is/was about 1 TB large.

Until now the only output was:
error removing rbd>rbd_data.221bf2eb141f2.51d2: (2) No such
file or directory
error removing rbd>rbd_data.221bf2eb141f2.e3f2: (2) No such
file or directory

> 
> [1] http://cephnotes.ksperis.com/blog/2014/07/04/remove-big-rbd-image
> 
> On Thu, Apr 26, 2018 at 9:25 AM Jan Marquardt  > wrote:
> 
> Hi,
> 
> I am currently trying to delete an rbd image which is seemingly causing
> our OSDs to crash, but it always gets stuck at 3%.
> 
> root@ceph4:~# rbd rm noc_tobedeleted
> Removing image: 3% complete...
> 
> Is there any way to force the deletion? Any other advices?
> 
> Best Regards
> 
> Jan
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com 
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Deleting an rbd image hangs

2018-04-27 Thread Jason Dillaman
Do you have any reason for why the OSDs crash? Anything the logs? Can
you provide an "rbd info noc_tobedeleted"?

On Thu, Apr 26, 2018 at 9:24 AM, Jan Marquardt  wrote:
> Hi,
>
> I am currently trying to delete an rbd image which is seemingly causing
> our OSDs to crash, but it always gets stuck at 3%.
>
> root@ceph4:~# rbd rm noc_tobedeleted
> Removing image: 3% complete...
>
> Is there any way to force the deletion? Any other advices?
>
> Best Regards
>
> Jan
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



-- 
Jason
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Deleting an rbd image hangs

2018-04-27 Thread David Turner
This old [1] blog post about removing super large RBDs is not relevant if
you're using object map on the RBDs, however it's method to manually delete
an RBD is still valid.  You can see if this works for you to manually
remove the problem RBD you're having.

[1] http://cephnotes.ksperis.com/blog/2014/07/04/remove-big-rbd-image

On Thu, Apr 26, 2018 at 9:25 AM Jan Marquardt  wrote:

> Hi,
>
> I am currently trying to delete an rbd image which is seemingly causing
> our OSDs to crash, but it always gets stuck at 3%.
>
> root@ceph4:~# rbd rm noc_tobedeleted
> Removing image: 3% complete...
>
> Is there any way to force the deletion? Any other advices?
>
> Best Regards
>
> Jan
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Deleting an rbd image hangs

2018-04-26 Thread Jan Marquardt
Hi,

I am currently trying to delete an rbd image which is seemingly causing
our OSDs to crash, but it always gets stuck at 3%.

root@ceph4:~# rbd rm noc_tobedeleted
Removing image: 3% complete...

Is there any way to force the deletion? Any other advices?

Best Regards

Jan
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com