Ilya, I see. My server is has 24GB of ram + 3GB of swap. While running the tests, I've noticed that the server had 14GB of ram shown as cached and only 2MB were used from the swap. Not sure if this is helpful to your debugging.
Andrei -- Andrei Mikhailovsky Director Arhont Information Security Web: http://www.arhont.com http://www.wi-foo.com Tel: +44 (0)870 4431337 Fax: +44 (0)208 429 3111 PGP: Key ID - 0x2B3438DE PGP: Server - keyserver.pgp.com DISCLAIMER The information contained in this email is intended only for the use of the person(s) to whom it is addressed and may be confidential or contain legally privileged information. If you are not the intended recipient you are hereby notified that any perusal, use, distribution, copying or disclosure is strictly prohibited. If you have received this email in error please immediately advise us by return email at and...@arhont.com and delete and purge the email and any attachments without making a copy. ----- Original Message ----- > From: "Ilya Dryomov" <ilya.dryo...@inktank.com> > To: "Andrei Mikhailovsky" <and...@arhont.com> > Cc: "ceph-users" <ceph-users@lists.ceph.com>, "Gregory Farnum" > <g...@gregs42.com> > Sent: Monday, 1 December, 2014 11:06:37 AM > Subject: Re: [ceph-users] Giant + nfs over cephfs hang tasks > On Mon, Dec 1, 2014 at 1:39 PM, Andrei Mikhailovsky > <and...@arhont.com> wrote: > > Ilya, > > > > I will try doing that once again tonight as this is a production > > cluster and > > when dds trigger that dmesg error the cluster's io becomes very bad > > and I > > have to reboot the server to get things on track. Most of my vms > > start > > having 70-90% iowait until that server is rebooted. > That's easily explained - those splats in dmesg indicate a case of a > severe memory pressure. > > > > I've actually checked what you've asked last time i've ran the > > test. > > > > When I do 4 dds concurrently nothing aprears in the dmesg output. > > No > > messages at all. > > > > The kern.log file that i've sent last time is what I got about a > > minute > > after i've started 8 dds. I've pasted the full output. The 8 dds > > did > > actually complete, but it took a rather long time. I was getting > > about 6MB/s > > per dd process compared to around 70MB/s per dd process when 4 dds > > were > > running. Do you still want me to run this or is the information > > i've > > provided enough? > No, no need if it's a production cluster. > Thanks, > Ilya
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com