[ceph-users] Scrub shutdown the OSD process

2013-04-15 Thread Olivier Bonvalet
Hi, I have an OSD process which is regulary shutdown by scrub, if I well understand that trace : 0 2013-04-15 09:29:53.708141 7f5a8e3cc700 -1 *** Caught signal (Aborted) ** in thread 7f5a8e3cc700 ceph version 0.56.4-4-gd89ab0e (d89ab0ea6fa8d0961cad82f6a81eccbd3bbd3f55) 1:

Re: [ceph-users] SL4500 as a storage machine

2013-04-15 Thread Mark Nelson
On 04/15/2013 03:25 AM, Stas Oskin wrote: Hi, Like I said, it's just my instinct. For a 180TB (raw) cluster you've got some tough choices to make. Some options might include: 1) high density and low cost by just stick a bunch of 3GB drives in 5 2U nodes and make sure you

[ceph-users] RBD snapshots are not «readable», because of LVM ?

2013-04-15 Thread Olivier Bonvalet
Hi, I'm trying to map a RBD snapshot, which contains an LVM PV. I can do the «map» : rbd map hdd3copies/jason@20130415-065314 --id alg Then pvscan works : pvscan | grep rbd PV /dev/rbd58 VG vg-jason lvm2 [19,94 GiB / 1,44 GiB free] But enabling LV doesn't work

Re: [ceph-users] ceph -w question

2013-04-15 Thread Samuel Just
Can you post the output of ceph osd tree? -Sam On Mon, Apr 15, 2013 at 9:52 AM, Jeppesen, Nelson nelson.jeppe...@disney.com wrote: Thanks for the help but how do I track down this issue? If data is inaccessible, that's a very bad thing given this is production. # ceph osd dump | grep pool

Re: [ceph-users] Scrub shutdown the OSD process

2013-04-15 Thread Gregory Farnum
On Mon, Apr 15, 2013 at 2:42 AM, Olivier Bonvalet ceph.l...@daevel.fr wrote: Hi, I have an OSD process which is regulary shutdown by scrub, if I well understand that trace : 0 2013-04-15 09:29:53.708141 7f5a8e3cc700 -1 *** Caught signal (Aborted) ** in thread 7f5a8e3cc700 ceph

Re: [ceph-users] ceph -w question

2013-04-15 Thread Samuel Just
Which host did you reboot to cause the incomplete pgs? Do you happen to have the output of ceph -s or, even better, ceph pg dump from the period with the incomplete pgs? From what i can see, the pgs should not have gone incomplete (at least, not for long). -Sam On Mon, Apr 15, 2013 at 10:17 AM,

Re: [ceph-users] Scrub shutdown the OSD process

2013-04-15 Thread Olivier Bonvalet
Le lundi 15 avril 2013 à 10:57 -0700, Gregory Farnum a écrit : On Mon, Apr 15, 2013 at 10:19 AM, Olivier Bonvalet ceph.l...@daevel.fr wrote: Le lundi 15 avril 2013 à 10:16 -0700, Gregory Farnum a écrit : Are you saying you saw this problem more than once, and so you completely wiped the

Re: [ceph-users] Upgrade stale PG

2013-04-15 Thread Darryl Bond
Ping, Any ideas? A week later and it is still the same, 300 pgs stuck stale. I have seen a few references since recommending that there are no gaps in the OSD numbers. Mine has gaps. Might this the be cause of my problem. Darryl On 04/05/13 07:27, Darryl Bond wrote: I have a 3 node ceph

Re: [ceph-users] mon crash

2013-04-15 Thread Dan Mick
Two is a strange choice for number of monitors; you really want an odd number. With two, if either one fails (or you have a network fault), the cluster is dead because there's no majority. That said, we certainly don't expect monitors to die when the network fault goes away. Searching the

Re: [ceph-users] mon crash

2013-04-15 Thread Matthew Roy
I'd bet that's 3495, it looks and sounds really, really similar. A lot of the devs are at a conference, but if you see Joao on IRC he'd know for sure. On 04/15/2013 04:56 PM, Craig Lewis wrote: I'm doing a test of Ceph in two colo facilities. Since it's just a test, I only have 2 VMs

[ceph-users] Ceph error: active+clean+scrubbing+deep

2013-04-15 Thread kakito
Dear all, I use Ceph Storage, Recently, I often get an error: mon.0 [INF] pgmap v277690: 640 pgs: 639 active+clean, 1 active+clean+scrubbing+deep; 14384 GB data, 14409 GB used, 90007 GB / 107 TB avail. It seems that it is not correct. I tried to restart. But not ok. It lows my system. I