My first question is about this metric: ceph_bluefs_read_prefetch_bytes and
I want to know what operation is related to this metric?

On Thu, Dec 3, 2020 at 7:49 PM Seena Fallah <seenafal...@gmail.com> wrote:

> Hi all,
>
> When my cluster gets into a recovery state (adding new node) I see a huge
> read throughput on its disks and it affects the latency! Disks are SSD and
> they don't have a separate WAL/DB.
> I'm using nautilus 14.2.14 and bluefs_buffered_io is false by default.
> When this throughput came on my disk it will get too much high latency.
> After I turned on bluefs_buffered_io another huge throughput around 1.2GB/s
> came in and it against affect my latency but much less than the previous
> one! (Graphs are attached and bluefs_buffered_io was turned on with ceph
> tell injectargs at 13:41 also I have restarted the OSD at 13:16 because it
> doesn't get better at the moment)
>
> I have four questions:
> 1. What are they? I see the recovery speed is 20MB/s and client io on that
> OSD is 10MB/s so what is this high throughput for?!
> 2. How can I control this throughput? Because my disks don't support this
> much throughput!
> 3. I see a common issue here https://tracker.ceph.com/issues/36482 that I
> think it's similar to my case. It was discussed about read_ahead, well
> should I change the read_ahead_kb config of my disk to support this type of
> request? I'm using the default value in ubuntu (128)
> 4. Is there any tuning required that can help to turn off the
> bluefs_buffered_io again?
>
> Configs I used for recovery:
> osd max backfills = 1
> osd recovery max active = 1
> osd recovery op priority = 1
> osd recovery priority = 1
> osd recovery sleep ssd = 0.2
>
> My OSD memory target is around 6GB.
>
> Thanks.
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to