Hi,
I have an OSD process which is regulary shutdown by scrub, if I well
understand that trace :
0 2013-04-15 09:29:53.708141 7f5a8e3cc700 -1 *** Caught signal (Aborted)
**
in thread 7f5a8e3cc700
ceph version 0.56.4-4-gd89ab0e (d89ab0ea6fa8d0961cad82f6a81eccbd3bbd3f55)
1:
On 04/15/2013 03:25 AM, Stas Oskin wrote:
Hi,
Like I said, it's just my instinct. For a 180TB (raw) cluster
you've got some tough choices to make. Some options might include:
1) high density and low cost by just stick a bunch of 3GB drives in
5 2U nodes and make sure you
Hi,
I'm trying to map a RBD snapshot, which contains an LVM PV.
I can do the «map» :
rbd map hdd3copies/jason@20130415-065314 --id alg
Then pvscan works :
pvscan | grep rbd
PV /dev/rbd58 VG vg-jason lvm2 [19,94 GiB / 1,44 GiB free]
But enabling LV doesn't work
Can you post the output of ceph osd tree?
-Sam
On Mon, Apr 15, 2013 at 9:52 AM, Jeppesen, Nelson
nelson.jeppe...@disney.com wrote:
Thanks for the help but how do I track down this issue? If data is
inaccessible, that's a very bad thing given this is production.
# ceph osd dump | grep pool
On Mon, Apr 15, 2013 at 2:42 AM, Olivier Bonvalet ceph.l...@daevel.fr wrote:
Hi,
I have an OSD process which is regulary shutdown by scrub, if I well
understand that trace :
0 2013-04-15 09:29:53.708141 7f5a8e3cc700 -1 *** Caught signal
(Aborted) **
in thread 7f5a8e3cc700
ceph
Which host did you reboot to cause the incomplete pgs? Do you happen
to have the output of ceph -s or, even better, ceph pg dump from the
period with the incomplete pgs? From what i can see, the pgs should
not have gone incomplete (at least, not for long).
-Sam
On Mon, Apr 15, 2013 at 10:17 AM,
Le lundi 15 avril 2013 à 10:57 -0700, Gregory Farnum a écrit :
On Mon, Apr 15, 2013 at 10:19 AM, Olivier Bonvalet ceph.l...@daevel.fr
wrote:
Le lundi 15 avril 2013 à 10:16 -0700, Gregory Farnum a écrit :
Are you saying you saw this problem more than once, and so you
completely wiped the
Ping,
Any ideas? A week later and it is still the same, 300 pgs stuck stale.
I have seen a few references since recommending that there are no gaps
in the OSD numbers. Mine has gaps. Might this the be cause of my problem.
Darryl
On 04/05/13 07:27, Darryl Bond wrote:
I have a 3 node ceph
Two is a strange choice for number of monitors; you really want an odd
number. With two, if either one fails (or you have a network fault),
the cluster is dead because there's no majority.
That said, we certainly don't expect monitors to die when the network
fault goes away. Searching the
I'd bet that's 3495, it looks and sounds really, really similar. A lot
of the devs are at a conference, but if you see Joao on IRC he'd know
for sure.
On 04/15/2013 04:56 PM, Craig Lewis wrote:
I'm doing a test of Ceph in two colo facilities. Since it's just a
test, I only have 2 VMs
Dear all,
I use Ceph Storage,
Recently, I often get an error:
mon.0 [INF] pgmap v277690: 640 pgs: 639 active+clean, 1
active+clean+scrubbing+deep; 14384 GB data, 14409 GB used, 90007 GB / 107 TB
avail.
It seems that it is not correct.
I tried to restart. But not ok.
It lows my system.
I
11 matches
Mail list logo