Oh, thank you!
On Wed, Mar 20, 2013 at 7:52 PM, Greg Farnum <[email protected]> wrote: > Yeah. If you run "ceph auth list" you'll get a dump of all the users and > keys the cluster knows about; each of your daemons has that key stored > somewhere locally (generally in /var/lib/ceph/ceph-[osd|mds|mon].$id). You > can create more or copy an unused MDS one. I believe the docs include > information on how this works. > -Greg > > Software Engineer #42 @ http://inktank.com | http://ceph.com > > > On Wednesday, March 20, 2013 at 10:48 AM, Igor Laskovy wrote: > > > Well, can you please clarify what exactly key I must to use? Do I need > to get/generate it somehow from working cluster? > > > > > > On Wed, Mar 20, 2013 at 7:41 PM, Greg Farnum <[email protected] (mailto: > [email protected])> wrote: > > > The MDS doesn't have any local state. You just need start up the > daemon somewhere with a name and key that are known to the cluster (these > can be different from or the same as the one that existed on the dead node; > doesn't matter!). > > > -Greg > > > > > > Software Engineer #42 @ http://inktank.com | http://ceph.com > > > > > > > > > On Wednesday, March 20, 2013 at 10:40 AM, Igor Laskovy wrote: > > > > > > > Actually, I already have recovered OSDs and MON daemon back to the > cluster according to > http://ceph.com/docs/master/rados/operations/add-or-rm-osds/ and > http://ceph.com/docs/master/rados/operations/add-or-rm-mons/ . > > > > > > > > But doc has missed info about removing/add MDS. > > > > How I can recovery MDS daemon for failed node? > > > > > > > > > > > > > > > > On Wed, Mar 20, 2013 at 3:23 PM, Dave (Bob) > > > > <[email protected](mailto: > [email protected]) (mailto:[email protected])> wrote: > > > > > Igor, > > > > > > > > > > I am sure that I'm right in saying that you just have to create a > new > > > > > filesystem (btrfs?) on the new block device, mount it, and then > > > > > initialise the osd with: > > > > > > > > > > ceph-osd -i <the osd number> --mkfs > > > > > > > > > > Then you can start the osd with: > > > > > > > > > > ceph-osd -i <the osd number> > > > > > > > > > > Since you are replacing an osd that already existed, the cluster > knows > > > > > about it, and there is a key for it that is known. > > > > > > > > > > I don't claim any great expertise, but this is what I've been > doing, and > > > > > the cluster seems to adopt the new osd and sort everything out. > > > > > > > > > > David > > > > > _______________________________________________ > > > > > ceph-users mailing list > > > > > [email protected] (mailto:[email protected]) > (mailto:[email protected]) > > > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > Igor Laskovy > > > > facebook.com/igor.laskovy (http://facebook.com/igor.laskovy) ( > http://facebook.com/igor.laskovy) > > > > Kiev, Ukraine > > > > _______________________________________________ > > > > ceph-users mailing list > > > > [email protected] (mailto:[email protected]) > (mailto:[email protected]) > > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > > > > > > > > -- > > Igor Laskovy > > facebook.com/igor.laskovy (http://facebook.com/igor.laskovy) > > Kiev, Ukraine > > > > -- Igor Laskovy facebook.com/igor.laskovy Kiev, Ukraine
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
