Re: [ceph-users] MDS dying on cuttlefish
Hi Greg, just for your own information, ceph mds newfs has disappeared from the help screen of the ceph command and it was a nightmare to understand the syntax (that has changed)... luckily sources were there :) For the flight log: ceph mds newfs metadata pool id data pool id --yes-i-really-mean-it Cheers, Gippa ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] MDS dying on cuttlefish
... and BTW, I know it's my fault that I haven't done the mds newfs, but I think it would be better to print an error rather that going in core dump with a trace. Just my eur 0.02 :) Cheers, Giuseppe ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] MDS dying on cuttlefish
On Wed, May 29, 2013 at 11:20 PM, Giuseppe 'Gippa' Paterno' gpate...@gpaterno.com wrote: Hi Greg, Oh, not the OSD stuff, just the CephFS stuff that goes on top. Look at http://www.mail-archive.com/ceph-users@lists.ceph.com/msg00029.html Although if you were re-creating pools and things, I think that would explain the crash you're seeing. -Greg I was thinking about that the problem is that with cuttlefish (0.61.2) seems that the command is no longer there. Has that moved? Nope, still there. -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
[ceph-users] MDS dying on cuttlefish
Hi! I've got a cluster of two nodes on Ubuntu 12.04 with cuttlefish from the ceph.com repo. ceph version 0.61.2 (fea782543a844bb277ae94d3391788b76c5bee60) The MDS process is dying after a while with a stack trace, but I can't understand why. I reproduced the same problem on debian 7 with the same repository. -3 2013-05-23 23:00:42.957679 7fa39e28e700 1 -- 10.123.200.189:6800/28919 == osd.0 10.123.200.188:6802/27665 1 osd_op_reply(5 200. [read 0~0] ack = -2 (No such file or directory)) v4 111+0+0 (2261481792 0 0) 0x29afe00 con 0x29c4b00 -2 2013-05-23 23:00:42.957780 7fa39e28e700 0 mds.0.journaler(ro) error getting journal off disk -1 2013-05-23 23:00:42.960974 7fa39e28e700 1 -- 10.123.200.189:6800/28919 == osd.0 10.123.200.188:6802/27665 2 osd_op_reply(1 mds0_inotable [read 0~0] ack = -2 (No such file or directory)) v4 112+0+0 (1612134461 0 0) 0x2a1c200 con 0x29c4b00 0 2013-05-23 23:00:42.963326 7fa39e28e700 -1 mds/MDSTable.cc: In function 'void MDSTable::load_2(int, ceph::bufferlist, Context*)' thread 7fa39e28e700 time 2013-05-23 23:00:42.961076 mds/MDSTable.cc: 150: FAILED assert(0) ceph version 0.61.2 (fea782543a844bb277ae94d3391788b76c5bee60) 1: (MDSTable::load_2(int, ceph::buffer::list, Context*)+0x3bb) [0x6dd2db] 2: (Objecter::handle_osd_op_reply(MOSDOpReply*)+0xe1b) [0x7275bb] 3: (MDS::handle_core_message(Message*)+0xae7) [0x513c57] 4: (MDS::_dispatch(Message*)+0x33) [0x513d53] 5: (MDS::ms_dispatch(Message*)+0xab) [0x515b3b] 6: (DispatchQueue::entry()+0x393) [0x847ca3] 7: (DispatchQueue::DispatchThread::entry()+0xd) [0x7caeed] 8: (()+0x6b50) [0x7fa3a3376b50] 9: (clone()+0x6d) [0x7fa3a1d24a7d] Full logs here: http://pastebin.com/C81g5jFd I can't understand why and I'd really appreciate an hint. Thanks! Regards, Giuseppe ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com