thanks David, that's confirming what I was assuming. To bad that there is no estimate/method to calculate the db partition size.
Dietmar On 09/25/2017 05:10 PM, David Turner wrote: > db/wal partitions are per OSD. DB partitions need to be made as big as > you need them. If they run out of space, they will fall back to the > block device. If the DB and block are on the same device, then there's > no reason to partition them and figure out the best size. If they are > on separate devices, then you need to make it as big as you need to to > ensure that it won't spill over (or if it does that you're ok with the > degraded performance while the db partition is full). I haven't come > across an equation to judge what size should be used for either > partition yet. > > On Mon, Sep 25, 2017 at 10:53 AM Dietmar Rieder > <dietmar.rie...@i-med.ac.at <mailto:dietmar.rie...@i-med.ac.at>> wrote: > > On 09/25/2017 02:59 PM, Mark Nelson wrote: > > On 09/25/2017 03:31 AM, TYLin wrote: > >> Hi, > >> > >> To my understand, the bluestore write workflow is > >> > >> For normal big write > >> 1. Write data to block > >> 2. Update metadata to rocksdb > >> 3. Rocksdb write to memory and block.wal > >> 4. Once reach threshold, flush entries in block.wal to block.db > >> > >> For overwrite and small write > >> 1. Write data and metadata to rocksdb > >> 2. Apply the data to block > >> > >> Seems we don’t have a formula or suggestion to the size of block.db. > >> It depends on the object size and number of objects in your pool. You > >> can just give big partition to block.db to ensure all the database > >> files are on that fast partition. If block.db full, it will use block > >> to put db files, however, this will slow down the db performance. So > >> give db size as much as you can. > > > > This is basically correct. What's more, it's not just the object > size, > > but the number of extents, checksums, RGW bucket indices, and > > potentially other random stuff. I'm skeptical how well we can > estimate > > all of this in the long run. I wonder if we would be better served by > > just focusing on making it easy to understand how the DB device is > being > > used, how much is spilling over to the block device, and make it > easy to > > upgrade to a new device once it gets full. > > > >> > >> If you want to put wal and db on same ssd, you don’t need to create > >> block.wal. It will implicitly use block.db to put wal. The only case > >> you need block.wal is that you want to separate wal to another disk. > > > > I always make explicit partitions, but only because I (potentially > > illogically) like it that way. There may actually be some benefits to > > using a single partition for both if sharing a single device. > > is this "Single db/wal partition" then to be used for all OSDs on a node > or do you need to create a seperate "Single db/wal partition" for each > OSD on the node? > > > > >> > >> I’m also studying bluestore, this is what I know so far. Any > >> correction is welcomed. > >> > >> Thanks > >> > >> > >>> On Sep 22, 2017, at 5:27 PM, Richard Hesketh > >>> <richard.hesk...@rd.bbc.co.uk > <mailto:richard.hesk...@rd.bbc.co.uk>> wrote: > >>> > >>> I asked the same question a couple of weeks ago. No response I got > >>> contradicted the documentation but nobody actively confirmed the > >>> documentation was correct on this subject, either; my end state was > >>> that I was relatively confident I wasn't making some horrible > mistake > >>> by simply specifying a big DB partition and letting bluestore work > >>> itself out (in my case, I've just got HDDs and SSDs that were > >>> journals under filestore), but I could not be sure there wasn't some > >>> sort of performance tuning I was missing out on by not specifying > >>> them separately. > >>> > >>> Rich > >>> > >>> On 21/09/17 20:37, Benjeman Meekhof wrote: > >>>> Some of this thread seems to contradict the documentation and > confuses > >>>> me. Is the statement below correct? > >>>> > >>>> "The BlueStore journal will always be placed on the fastest device > >>>> available, so using a DB device will provide the same benefit > that the > >>>> WAL device would while also allowing additional metadata to be > stored > >>>> there (if it will fix)." > >>>> > >>>> > > http://docs.ceph.com/docs/master/rados/configuration/bluestore-config-ref/#devices > >>>> > >>>> > >>>> it seems to be saying that there's no reason to create > separate WAL > >>>> and DB partitions if they are on the same device. Specifying one > >>>> large DB partition per OSD will cover both uses. > >>>> > >>>> thanks, > >>>> Ben > >>>> > >>>> On Thu, Sep 21, 2017 at 12:15 PM, Dietmar Rieder > >>>> <dietmar.rie...@i-med.ac.at > <mailto:dietmar.rie...@i-med.ac.at>> wrote: > >>>>> On 09/21/2017 05:03 PM, Mark Nelson wrote: > >>>>>> > >>>>>> On 09/21/2017 03:17 AM, Dietmar Rieder wrote: > >>>>>>> On 09/21/2017 09:45 AM, Maged Mokhtar wrote: > >>>>>>>> On 2017-09-21 07:56, Lazuardi Nasution wrote: > >>>>>>>> > >>>>>>>>> Hi, > >>>>>>>>> > >>>>>>>>> I'm still looking for the answer of these questions. Maybe > >>>>>>>>> someone can > >>>>>>>>> share their thought on these. Any comment will be helpful too. > >>>>>>>>> > >>>>>>>>> Best regards, > >>>>>>>>> > >>>>>>>>> On Sat, Sep 16, 2017 at 1:39 AM, Lazuardi Nasution > >>>>>>>>> <mrxlazuar...@gmail.com <mailto:mrxlazuar...@gmail.com> > <mailto:mrxlazuar...@gmail.com <mailto:mrxlazuar...@gmail.com>>> wrote: > >>>>>>>>> > >>>>>>>>> Hi, > >>>>>>>>> > >>>>>>>>> 1. Is it possible configure use osd_data not as small > >>>>>>>>> partition on > >>>>>>>>> OSD but a folder (ex. on root disk)? If yes, how to do > that > >>>>>>>>> with > >>>>>>>>> ceph-disk and any pros/cons of doing that? > >>>>>>>>> 2. Is WAL & DB size calculated based on OSD size or > expected > >>>>>>>>> throughput like on journal device of filestore? If no, > what > >>>>>>>>> is the > >>>>>>>>> default value and pro/cons of adjusting that? > >>>>>>>>> 3. Is partition alignment matter on Bluestore, including > >>>>>>>>> WAL & DB > >>>>>>>>> if using separate device for them? > >>>>>>>>> > >>>>>>>>> Best regards, > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> _______________________________________________ > >>>>>>>>> ceph-users mailing list > >>>>>>>>> ceph-users@lists.ceph.com > <mailto:ceph-users@lists.ceph.com> <mailto:ceph-users@lists.ceph.com > <mailto:ceph-users@lists.ceph.com>> > >>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >>>>>>>> > >>>>>>>> > >>>>>>>> I am also looking for recommendations on wal/db partition > sizes. > >>>>>>>> Some > >>>>>>>> hints: > >>>>>>>> > >>>>>>>> ceph-disk defaults used in case it does not find > >>>>>>>> bluestore_block_wal_size or bluestore_block_db_size in > config file: > >>>>>>>> > >>>>>>>> wal = 512MB > >>>>>>>> > >>>>>>>> db = if bluestore_block_size (data size) is in config file it > >>>>>>>> uses 1/100 > >>>>>>>> of it else it uses 1G. > >>>>>>>> > >>>>>>>> There is also a presentation by Sage back in March, see > page 16: > >>>>>>>> > >>>>>>>> > > https://www.slideshare.net/sageweil1/bluestore-a-new-storage-backend-for-ceph-one-year-in > >>>>>>>> > >>>>>>>> > >>>>>>>> > >>>>>>>> wal: 512 MB > >>>>>>>> > >>>>>>>> db: "a few" GB > >>>>>>>> > >>>>>>>> the wal size is probably not debatable, it will be like a > >>>>>>>> journal for > >>>>>>>> small block sizes which are constrained by iops hence 512 MB is > >>>>>>>> more > >>>>>>>> than enough. Probably we will see more on the db size in the > >>>>>>>> future. > >>>>>>> This is what I understood so far. > >>>>>>> I wonder if it makes sense to set the db size as big as > possible and > >>>>>>> divide entire db device is by the number of OSDs it will serve. > >>>>>>> > >>>>>>> E.g. 10 OSDs / 1 NVME (800GB) > >>>>>>> > >>>>>>> (800GB - 10x1GB wal ) / 10 = ~79Gb db size per OSD > >>>>>>> > >>>>>>> Is this smart/stupid? > >>>>>> Personally I'd use 512MB-2GB for the WAL (larger buffers > reduce write > >>>>>> amp but mean larger memtables and potentially higher overhead > >>>>>> scanning > >>>>>> through memtables). 4x256MB buffers works pretty well, but > it means > >>>>>> memory overhead too. Beyond that, I'd devote the entire rest > of the > >>>>>> device to DB partitions. > >>>>>> > >>>>> thanks for your suggestion Mark! > >>>>> > >>>>> So, just to make sure I understood this right: > >>>>> > >>>>> You'd use a separeate 512MB-2GB WAL partition for each OSD > and the > >>>>> entire rest for DB partitions. > >>>>> > >>>>> In the example case with 10xHDD OSD and 1 NVME it would then > be 10 WAL > >>>>> partitions with each 512MB-2GB and 10 equal sized DB partitions > >>>>> consuming the rest of the NVME. > >>>>> > >>>>> > >>>>> Thanks > >>>>> Dietmar > >>>>> -- > >>>>> _________________________________________ > >>>>> D i e t m a r R i e d e r, Mag.Dr. > >>>>> Innsbruck Medical University > >>>>> Biocenter - Division for Bioinformatics > >>>>> > >>>>> > >>>>> > >>>>> _______________________________________________ > >>>>> ceph-users mailing list > >>>>> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> > >>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >>>>> > >>>> _______________________________________________ > >>>> ceph-users mailing list > >>>> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> > >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >>> > >>> _______________________________________________ > >>> ceph-users mailing list > >>> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> > >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> > >> _______________________________________________ > >> ceph-users mailing list > >> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > -- > _________________________________________ > D i e t m a r R i e d e r, Mag.Dr. > Innsbruck Medical University > Biocenter - Division for Bioinformatics > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > -- _________________________________________ D i e t m a r R i e d e r, Mag.Dr. Innsbruck Medical University Biocenter - Division for Bioinformatics
signature.asc
Description: OpenPGP digital signature
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com