On 07/17/2015 02:50 PM, Gregory Farnum wrote:
On Fri, Jul 17, 2015 at 1:13 PM, Kenneth Waegeman
<kenneth.waege...@ugent.be> wrote:
Hi all,

I've read in the documentation that OSDs use around 512MB on a healthy
cluster.(http://ceph.com/docs/master/start/hardware-recommendations/#ram)
Now, our OSD's are all using around 2GB of RAM memory while the cluster is
healthy.


   PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+ COMMAND
29784 root      20   0 6081276 2.535g   4740 S   0.7  8.1   1346:55 ceph-osd
32818 root      20   0 5417212 2.164g  24780 S  16.2  6.9   1238:55 ceph-osd
25053 root      20   0 5386604 2.159g  27864 S   0.7  6.9   1192:08 ceph-osd
33875 root      20   0 5345288 2.092g   3544 S   0.7  6.7   1188:53 ceph-osd
30779 root      20   0 5474832 2.090g  28892 S   1.0  6.7   1142:29 ceph-osd
22068 root      20   0 5191516 2.000g  28664 S   0.7  6.4  31:56.72 ceph-osd
34932 root      20   0 5242656 1.994g   4536 S   0.3  6.4   1144:48 ceph-osd
26883 root      20   0 5178164 1.938g   6164 S   0.3  6.2   1173:01 ceph-osd
31796 root      20   0 5193308 1.916g  27000 S  16.2  6.1 923:14.87 ceph-osd
25958 root      20   0 5193436 1.901g   2900 S   0.7  6.1   1039:53 ceph-osd
27826 root      20   0 5225764 1.845g   5576 S   1.0  5.9   1031:15 ceph-osd
36011 root      20   0 5111660 1.823g  20512 S  15.9  5.8   1093:01 ceph-osd
19736 root      20   0 2134680 0.994g      0 S   0.3  3.2  46:13.47 ceph-osd



[root@osd003 ~]# ceph status
2015-07-17 14:03:13.865063 7f1fde5f0700 -1 WARNING: the following dangerous
and experimental features are enabled: keyvaluestore
2015-07-17 14:03:13.887087 7f1fde5f0700 -1 WARNING: the following dangerous
and experimental features are enabled: keyvaluestore
     cluster 92bfcf0a-1d39-43b3-b60f-44f01b630e47
      health HEALTH_OK
      monmap e1: 3 mons at
{mds01=10.141.16.1:6789/0,mds02=10.141.16.2:6789/0,mds03=10.141.16.3:6789/0}
             election epoch 58, quorum 0,1,2 mds01,mds02,mds03
      mdsmap e17218: 1/1/1 up {0=mds03=up:active}, 1 up:standby
      osdmap e25542: 258 osds: 258 up, 258 in
       pgmap v2460163: 4160 pgs, 4 pools, 228 TB data, 154 Mobjects
             270 TB used, 549 TB / 819 TB avail
                 4152 active+clean
                    8 active+clean+scrubbing+deep


We are using erasure code on most of our OSDs, so maybe that is a reason.
But also the cache-pool filestore OSDS on 200GB SSDs are using 2GB of RAM.
Our erasure code pool (16*14 osds) have a pg_num of 2048; our cache pool
(2*14 OSDS) has a pg_num of 1024.

Are these normal values for this configuration, and is the documentation a
bit outdated, or should we look into something else?

2GB of RSS is larger than I would have expected, but not unreasonable.
In particular I don't think we've gathered numbers on either EC pools
or on the effects of the caching processes.

Which data is actually in memory of the OSDS?
Is this mostly cached data?
We are short on memory on these servers, can we have influence on this?

Thanks again!
Kenneth

-Greg

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to