Keep in mind that the mds server is cpu-bound, so during heavy workloads it
will eat up CPU usage, so the OSD daemons can affect or be affected by the
MDS daemon.
But it does work well. We've been running a few clusters with MON, MDS and
OSDs sharing the same hosts for a couple of years now.

Regards,

Webert Lima
DevOps Engineer at MAV Tecnologia
*Belo Horizonte - Brasil*
*IRC NICK - WebertRLZ*


On Tue, Jun 19, 2018 at 11:03 AM Paul Emmerich <paul.emmer...@croit.io>
wrote:

> Just co-locate them with your OSDs. You can can control how much RAM the
> MDSs use with the "mds cache memory limit" option. (default 1 GB)
> Note that the cache should be large enough RAM to keep the active working
> set in the mds cache but 1 million files is not really a lot.
> As a rule of thumb: ~1GB of MDS cache per ~100k files.
>
> 64GB of RAM for 12 OSDs and an MDS is enough in most cases.
>
> Paul
>
> 2018-06-19 15:34 GMT+02:00 Denny Fuchs <linuxm...@4lin.net>:
>
>> Hi,
>>
>> Am 19.06.2018 15:14, schrieb Stefan Kooman:
>>
>> Storage doesn't matter for MDS, as they won't use it to store ceph data
>>> (but instead use the (meta)data pool to store meta data).
>>> I would not colocate the MDS daemons with the OSDS, but instead create a
>>> couple of VMs (active / standby) and give them as much RAM as you
>>> possibly can.
>>>
>>
>> thanks a lot. I think, we would start with round about 8GB and see, what
>> happens.
>>
>> cu denny
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
>
> --
> Paul Emmerich
>
> Looking for help with your Ceph cluster? Contact us at https://croit.io
>
> croit GmbH
> Freseniusstr. 31h
> 81247 München
> www.croit.io
> Tel: +49 89 1896585 90
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to