> On Jun 6, 2017, at 10:35 AM, Webert de Souza Lima <webert.b...@gmail.com> > wrote: > > I'd like to add that, from all tests I did, the writing of new files only go > directly to the cache tier if you set hit set count = 0. > >
Is there any concern or disadvantage once I set hit set count = 0? Looks like it is used to record the read/write recency only? Sorry for the stupid question but i’m trying to understand the cache-tier behavior :) Thanks, Ting Yi Lin > Em Seg, 5 de jun de 2017 23:26, TYLin <wooer...@gmail.com > <mailto:wooer...@gmail.com>> escreveu: >> On Jun 5, 2017, at 6:47 PM, Christian Balzer <ch...@gol.com >> <mailto:ch...@gol.com>> wrote: >> >> Personally I avoid odd numbered releases, but my needs for stability >> and low update frequency seem to be far off the scale for "normal" Ceph >> users. >> >> W/o precise numbers of files and the size of your SSDs (which type?) it is >> hard to say, but you're likely to be better off just having all metadata >> on an SSD pool instead of cache-tiering. >> 800MB/s sounds about right for your network and cluster in general (no >> telling for sure w/o SSD/HDD details of course). >> >> As I pointed out before and will try to explain again below, that speed >> difference, while pretty daunting, isn't all that surprising. >> > > SSD: Intel S3520 240GB > HDD: WDC WD5003ABYZ-011FA0 500GB > fio: bs=4m iodepth=32 > dd: bs=4m > The test file is 20GB. > >> No, not quite. Re-read what I wrote, there's a difference between RADOS >> object creation and actual data (contents). >> >> The devs or other people with more code familiarity will correct me, but >> essentially as I understand it this happens when a new RADOS object gets >> created in conjunction with a cache-tier: >> >> 1. Client (cephfs, rbd, whatever) talks to the cache-tier and the >> transaction causes a new object to be created. >> Since the tier is an overlay of the actual backing storage, the object >> (but not necessarily the curent data in it) needs to exist on both. >> 2. Object gets created on backing storage which involves creating the >> file (at zero length), any needed directories above and the entry in the >> OMAP leveldb. All on HDDs, all slow. >> I'm pretty sure this needs to be done and finished before the object is >> usable, no journals to speed this up. >> 3. Cache-tier pseudo-promotes the new object (it is empty after all) and >> starts accepting writes. >> >> This is leaving out any metadata stuff CephFS needs to do for new "blocks" >> and files, which may also be more involved than overwrites. >> >> Christian > > You make it clear to me! thanks! Really appreciate your kind explanation. > > Thanks, > Ting Yi Lin > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com