Ilya, 

I see. My server is has 24GB of ram + 3GB of swap. While running the tests, 
I've noticed that the server had 14GB of ram shown as cached and only 2MB were 
used from the swap. Not sure if this is helpful to your debugging. 

Andrei 

-- 
Andrei Mikhailovsky 
Director 
Arhont Information Security 

Web: http://www.arhont.com 
http://www.wi-foo.com 
Tel: +44 (0)870 4431337 
Fax: +44 (0)208 429 3111 
PGP: Key ID - 0x2B3438DE 
PGP: Server - keyserver.pgp.com 

DISCLAIMER 

The information contained in this email is intended only for the use of the 
person(s) to whom it is addressed and may be confidential or contain legally 
privileged information. If you are not the intended recipient you are hereby 
notified that any perusal, use, distribution, copying or disclosure is strictly 
prohibited. If you have received this email in error please immediately advise 
us by return email at and...@arhont.com and delete and purge the email and any 
attachments without making a copy. 

----- Original Message -----

> From: "Ilya Dryomov" <ilya.dryo...@inktank.com>
> To: "Andrei Mikhailovsky" <and...@arhont.com>
> Cc: "ceph-users" <ceph-users@lists.ceph.com>, "Gregory Farnum"
> <g...@gregs42.com>
> Sent: Monday, 1 December, 2014 11:06:37 AM
> Subject: Re: [ceph-users] Giant + nfs over cephfs hang tasks

> On Mon, Dec 1, 2014 at 1:39 PM, Andrei Mikhailovsky
> <and...@arhont.com> wrote:
> > Ilya,
> >
> > I will try doing that once again tonight as this is a production
> > cluster and
> > when dds trigger that dmesg error the cluster's io becomes very bad
> > and I
> > have to reboot the server to get things on track. Most of my vms
> > start
> > having 70-90% iowait until that server is rebooted.

> That's easily explained - those splats in dmesg indicate a case of a
> severe memory pressure.

> >
> > I've actually checked what you've asked last time i've ran the
> > test.
> >
> > When I do 4 dds concurrently nothing aprears in the dmesg output.
> > No
> > messages at all.
> >
> > The kern.log file that i've sent last time is what I got about a
> > minute
> > after i've started 8 dds. I've pasted the full output. The 8 dds
> > did
> > actually complete, but it took a rather long time. I was getting
> > about 6MB/s
> > per dd process compared to around 70MB/s per dd process when 4 dds
> > were
> > running. Do you still want me to run this or is the information
> > i've
> > provided enough?

> No, no need if it's a production cluster.

> Thanks,

> Ilya
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to