Ceph-disk didn't remove an osd from the cluster either. That has never been
a thing for ceph-disk or ceph-volume. There are other commands for that.

On Sat, Jun 2, 2018, 4:29 PM Marc Roos <m.r...@f1-outsourcing.eu> wrote:

>
> But leaves still entries in crush map and maybe also ceph auth ls, and
> the dir in /var/lib/ceph/osd
>
>
>
> -----Original Message-----
> From: Oliver Freyermuth [mailto:freyerm...@physik.uni-bonn.de]
> Sent: zaterdag 2 juni 2018 18:29
> To: Marc Roos; ceph-users
> Subject: Re: [ceph-users] Bug? ceph-volume zap not working
>
> The command mapping from ceph-disk to ceph-volume is certainly not 1:1.
> What we are ended up using is:
> ceph-volume lvm zap /dev/sda --destroy
> This takes care of destroying Pvs and Lvs (as the documentation says).
>
> Cheers,
>         Oliver
>
> Am 02.06.2018 um 12:16 schrieb Marc Roos:
> >
> > I guess zap should be used instead of destroy? Maybe keep ceph-disk
> > backwards compatibility and keep destroy??
> >
> > [root@c03 bootstrap-osd]# ceph-volume lvm zap /dev/sdf
> > --> Zapping: /dev/sdf
> > --> Unmounting /var/lib/ceph/osd/ceph-19
> > Running command: umount -v /var/lib/ceph/osd/ceph-19
> >  stderr: umount: /var/lib/ceph/osd/ceph-19 (tmpfs) unmounted Running
> > command: wipefs --all /dev/sdf
> >  stderr: wipefs: error: /dev/sdf: probing initialization failed:
> > Device or resource busy
> > -->  RuntimeError: command returned non-zero exit status: 1
> >
> > Pvs / lvs are still there, I guess these are keeping the 'resource
> busy'
> >
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to