On Mon, Aug 5, 2019 at 6:35 PM <vita...@yourcmc.ru> wrote: > > Hi Team, > > @vita...@yourcmc.ru , thank you for information and could you please > > clarify on the below quires as well, > > > > 1. Average object size we use will be 256KB to 512KB , will there be > > deferred write queue ? > > With the default settings, no (bluestore_prefer_deferred_size_hdd = > 32KB) > > Are you sure that 256-512KB operations aren't counted as multiple
> operations in your disk stats? > I think it is not taking multiple operations. > > > 2. Share the link of existing rocksdb ticket which does 2 write + > > syncs. > > My PR is here https://github.com/ceph/ceph/pull/26909, you can find the > issue tracker links inside it. > > > 3. Any configuration by which we can reduce/optimize the iops ? > > As already said part of your I/O may be caused by the metadata (rocksdb) > reads if it doesn't fit into RAM. You can try to add more RAM in that > case... :) > I can add RAM ans is there a way to increase rocksdb caching , can I increase bluestore_cache_size_hdd to higher value to cache rocksdb? > > You can also try to add SSDs for metadata (block.db/block.wal). > This we have planned to add some SSDs and how many OSD's rocks db we can add per SSDs and i guess if one SSD is down then all related OSDs has to be re-installed. > > Is there something else?... I don't think so. > > -- > Vitaliy Filippov >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com