I am wondering what the best way is of deleting a cluster, removing all
the OSDs, and basically start over. I plan to create a few ceph test
clusters to determine what works best in our use-case. There is no real
data being stored, so I don't care about data-loss.

I have a cephfs setup on top of two pools: data and metadata. Presumably
I can remove this easily with 'ceph fs rm'

1. Do I need to delete the OSD pools?
2. How to I remove the OSDs from the cluster without ceph doing what it
does and rebalancing data between the remaining OSDs?

I read the Manually Remove OSD documentation page,
https://docs.ceph.com/docs/master/rados/operations/add-or-rm-osds/#removing-osds-manual,
but I want to remove ALL OSDs from the cluster? Is this still the right
set of steps/commands?

Thanks for any insight for a ceph newbie.

PS - If it matters the servers running ceph-mon and ceph-mgr are on
separate computers than the servers running ceph-osd.

Sincerely,
Shawn Kwang
-- 
Associate Scientist
Center for Gravitation, Cosmology, and Astrophysics
University of Wisconsin-Milwaukee
office: +1 414 229 4960

Attachment: signature.asc
Description: OpenPGP digital signature

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to