Re: 0.56.6: MDS asserts on restart

2013-05-13 Thread Christopher Kunz
Am 10.05.13 15:21, schrieb Christopher Kunz:
 Hi all,
 
 after our update to 0.56.6, all MDSes assert and crash on start. This is
 not essential for our setup, but still not great. We assume this might
 be due to a bug - see the nopaste below for a trace.
 
For reference: I created http://tracker.ceph.com/issues/5037 for this
issue.

Regards,

--ck

--
To unsubscribe from this list: send the line unsubscribe ceph-devel in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


0.56.6: MDS asserts on restart

2013-05-10 Thread Christopher Kunz
Hi all,

after our update to 0.56.6, all MDSes assert and crash on start. This is
not essential for our setup, but still not great. We assume this might
be due to a bug - see the nopaste below for a trace.

http://nopaste.info/aa06c50a2f.html

Regards,

--ck
--
To unsubscribe from this list: send the line unsubscribe ceph-devel in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


0.56 scrub OSD memleaks, WAS Re: [0.48.3] OSD memory leak when scrubbing

2013-02-18 Thread Christopher Kunz
Am 16.02.13 10:09, schrieb Wido den Hollander:
 On 02/16/2013 08:09 AM, Andrey Korolyov wrote:
 Can anyone who hit this bug please confirm that your system contains
 libc 2.15+?

 
Hello,

when we started a deep scrub on our 0.56.2 cluster today, we saw a
massive memleak about 1 hour into the scrub. One OSD claimed over
53GByte within 10 minutes. We had to restart the OSD to keep the cluster
stable.

Another OSD is currently claiming about 27GByte and will be restarted
soon. All circumstantial evidence points to the deep scrub as the source
of the leak.

One affected node is running libc 2.15 (Ubuntu 12.04 LTS), the other one
is using libc 2.11.3 (Debian Squeeze). So it seems this is not a
libc-dependant issue.

We have disabled scrub completely.

Regards,

--ck

PS: Do we have any idea when this will be fixed?
--
To unsubscribe from this list: send the line unsubscribe ceph-devel in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: the autobuild gpg key expired!

2013-02-08 Thread Christopher Kunz
Am 08.02.13 07:46, schrieb Sage Weil:
  wget -q -O- https://raw.github.com/ceph/ceph/master/keys/autobuild.asc | 
 sudo apt-key add -

This doesn't work because raw.github.com uses a certificate issued to CN
= *.a.ssl.fastly.net
for whatever reason, and wget will complain.

It complains rightly, because there is no way to verifiy that this is
*actually* github.com we're speaking to. If you want the key regardless,
you can use this line:

wget --no-check-certificate -q -O-
https://raw.github.com/ceph/ceph/master/keys/autobuild.asc | sudo
apt-key add -

Regards,

--ck
--
To unsubscribe from this list: send the line unsubscribe ceph-devel in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Current OSD weight vs. target weight

2013-01-17 Thread Christopher Kunz
Hi,

if run during a user-issued reweight (i.e. ceph osd crush reweight x
y), ceph osd tree shows the target weight of an OSD. Is there a way
to see the *current* weight of the OSD?

We would like to be able to approximate the amount of rollback necessary
per OSD if we need to cancel a larger reweight at some point. This seems
impossible, though, without knowing the current weight.

Any hints how to do this?

Regards,

--ck

-- 
filoo GmbH
Dr. Christopher Kunz

E-Mail: ch...@filoo.de
Tel.: (+49) 0 52 41 8 67 30 -18
Fax: (+49) 0 52 41 / 8 67 30 -20

Please sign  encrypt mail wherever possible, my key:
C882 8ED1 7DD1 9011 C088 EA50 5CFA 2EEB 397A CAC1

Moltkestraße 25a
0 Gütersloh

HRB4355, AG Gütersloh
Geschäftsführer: S.Grewing, J.Rehpöhler, Dr. C.Kunz

Filoo im Web: http://www.filoo.de/
Folgen Sie uns auf Twitter: http://twitter.com/filoogmbh
Werden Sie unser Fan auf Facebook: http://facebook.com/filoogmbh
--
To unsubscribe from this list: send the line unsubscribe ceph-devel in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Are there significant performance enhancements in 0.56.x to be expected soon or planned in the near future?

2013-01-09 Thread Christopher Kunz
Hi,
 
 Yes, 0.56(.1) has a significant performance increase compared to 0.48
 
That is not exactly the OP's question, though. If I understand
correctly, she is concerned about ongoing performance improvements
within the bobtail branch, i.e. between 0.56.1 and 0.56.X (with X1).

Jutta, what kind of use case do you have in mind, i.e. how complex are
your benchmarking scenarios?

Regards,

--ck
--
To unsubscribe from this list: send the line unsubscribe ceph-devel in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


btrfs vulnerable to hash-DoS attack

2012-12-13 Thread Christopher Kunz
Hi all,

since a couple people here are using btrfs (we, fortunately, are not), I
thought you might be interested in this blog entry:
http://crypto.junod.info/2012/12/13/hash-dos-and-btrfs/

Basically, it seems that if you create files with filenames that have
colliding CRC32 hashes, all kind of hell breaks loose.

Definitely an interesting read.

Regards,

--ck
--
To unsubscribe from this list: send the line unsubscribe ceph-devel in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Best practice with 0.48.2 to take a node into maintenance

2012-12-04 Thread Christopher Kunz
Hey,
 That has to be old information - we're using it in anger on 64 bit x86
 machines, from both Intel
 and AMD.  I have no idea about ARM etc, however.
just to make absolutely sure: Are you using kexec in anger to switch
Ceph nodes to new kernels? I'm very doubtful that the approach will work
seamlessly with Ceph, as it is rather picky regarding kernel state...

Regards,

--ck
--
To unsubscribe from this list: send the line unsubscribe ceph-devel in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Best practice with 0.48.2 to take a node into maintenance

2012-12-03 Thread Christopher Kunz
Am 03.12.12 20:14, schrieb Josh Durgin:
 On 12/03/2012 11:05 AM, Oliver Francke wrote:
 Hi *,

 well, even if 0.48.2 is really stable and reliable, it is not
 everytime the case with linux kernel. We have a couple of nodes,
 where an update would make life better.
 So, as our OSD-nodes have to care for VM's too, it's not the problem
 to let them drain so migrate all of them to other nodes.
 Just reboot? Perhaps not, cause all OSD's will begin to
 remap/backfill, they are instructed to do so. Well, declare them as
 osd lost?
 Dangerous. Is there another way I miss in doing node-maintenance?
 Will we have to wait for bobtail for far less hassle with all
 remapping and resources?

 By default the monitors won't mark an OSD out in the time it takes to
 reboot, but if maintenance takes longer, you can drain data from the
 node.
Hi,

what time is that (in seconds) and how can we reliably test this?

Regards,

--ck
--
To unsubscribe from this list: send the line unsubscribe ceph-devel in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html