Hi all,

I use puppet to deploy and manage my clusters.

Recently, as I have been doing a removal of old hardware and adding of
new I've noticed that sometimes the "ceph osd create" is returning
repeated IDs. Usually it's on the same server, but yesterday I saw it
in different servers.

I was expecting the OSD ID's to be unique, and when they come on the
same server puppet starts spewing errors - which is desirable - but
when it's in different servers it broke those OSDs in Ceph. As they
hadn't backfill any full PGs I just wiped, removed and started anew.

As for the process itself: The OSDs are marked out and removed from
crush, when empty they are auth del and osd rm. After building the
server puppet will osd create, and use the generated ID for crush move
and mkfs.

Unfortunately I haven't been able to reproduce in isolation, and being
a production cluster logging is tuned way down.

This has happened in several different clusters, but they are all
running 10.2.7.

Any ideas?

thanks,
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to