Re: [ceph-users] MDS dying on cuttlefish

2013-05-31 Thread Giuseppe 'Gippa' Paterno'
Hi Greg,
just for your own information, ceph mds newfs has disappeared from the
help screen of the ceph command and it was a nightmare to understand
the syntax (that has changed)... luckily sources were there :)

For the flight log:
ceph mds newfs metadata pool id data pool id --yes-i-really-mean-it

Cheers,
Gippa
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] MDS dying on cuttlefish

2013-05-31 Thread Giuseppe 'Gippa' Paterno'
... and BTW, I know it's my fault that I haven't done the mds newfs, but
I think it would be better to print an error rather that going in core
dump with a trace.
Just my eur 0.02 :)
Cheers,
Giuseppe
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] MDS dying on cuttlefish

2013-05-30 Thread Gregory Farnum
On Wed, May 29, 2013 at 11:20 PM, Giuseppe 'Gippa' Paterno'
gpate...@gpaterno.com wrote:
 Hi Greg,
 Oh, not the OSD stuff, just the CephFS stuff that goes on top. Look at
 http://www.mail-archive.com/ceph-users@lists.ceph.com/msg00029.html
 Although if you were re-creating pools and things, I think that would
 explain the crash you're seeing.
 -Greg

 I was thinking about that  the problem is that with cuttlefish
 (0.61.2) seems that the command is no longer there.
 Has that moved?

Nope, still there.
-Greg
Software Engineer #42 @ http://inktank.com | http://ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] MDS dying on cuttlefish

2013-05-23 Thread Giuseppe 'Gippa' Paterno'
Hi!

I've got a cluster of two nodes on Ubuntu 12.04 with cuttlefish from the
ceph.com repo.
ceph version 0.61.2 (fea782543a844bb277ae94d3391788b76c5bee60)

The MDS process is dying after a while with a stack trace, but I can't
understand why.
I reproduced the same problem on debian 7 with the same repository.

-3 2013-05-23 23:00:42.957679 7fa39e28e700  1 --
10.123.200.189:6800/28919 == osd.0 10.123.200.188:6802/27665 1 
osd_op_reply(5 200. [read 0~0] ack = -2 (No such file or
directory)) v4  111+0+0 (2261481792 0 0) 0x29afe00 con 0x29c4b00
-2 2013-05-23 23:00:42.957780 7fa39e28e700  0 mds.0.journaler(ro)
error getting journal off disk
-1 2013-05-23 23:00:42.960974 7fa39e28e700  1 --
10.123.200.189:6800/28919 == osd.0 10.123.200.188:6802/27665 2 
osd_op_reply(1 mds0_inotable [read 0~0] ack = -2 (No such file or
directory)) v4  112+0+0 (1612134461 0 0) 0x2a1c200 con 0x29c4b00
 0 2013-05-23 23:00:42.963326 7fa39e28e700 -1 mds/MDSTable.cc: In
function 'void MDSTable::load_2(int, ceph::bufferlist, Context*)'
thread 7fa39e28e700 time 2013-05-23 23:00:42.961076
mds/MDSTable.cc: 150: FAILED assert(0)

 ceph version 0.61.2 (fea782543a844bb277ae94d3391788b76c5bee60)
 1: (MDSTable::load_2(int, ceph::buffer::list, Context*)+0x3bb) [0x6dd2db]
 2: (Objecter::handle_osd_op_reply(MOSDOpReply*)+0xe1b) [0x7275bb]
 3: (MDS::handle_core_message(Message*)+0xae7) [0x513c57]
 4: (MDS::_dispatch(Message*)+0x33) [0x513d53]
 5: (MDS::ms_dispatch(Message*)+0xab) [0x515b3b]
 6: (DispatchQueue::entry()+0x393) [0x847ca3]
 7: (DispatchQueue::DispatchThread::entry()+0xd) [0x7caeed]
 8: (()+0x6b50) [0x7fa3a3376b50]
 9: (clone()+0x6d) [0x7fa3a1d24a7d]

Full logs here:
http://pastebin.com/C81g5jFd

I can't understand why and I'd really appreciate an hint.
Thanks!
Regards,
  Giuseppe
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com