On 09/01, Chao Yu wrote:
> This patch enables f2fs_balance_fs_bg() to check all metadatas' dirty
> threshold rather than just checking node block's, so that checkpoint()
> from background can be triggered more frequently to avoid heaping up
> too much dirty metadatas.
> 
> Threshold value by default:
> race with foreground ops      single type     global
> No                            16MB            24MB
> Yes                           24MB            36MB
> 
> In addtion, let f2fs_balance_fs_bg() be aware of roll-forward sapce
> as well as fsync().
> 
> Signed-off-by: Chao Yu <c...@kernel.org>
> ---
>  fs/f2fs/f2fs.h    |  3 +++
>  fs/f2fs/node.h    |  5 -----
>  fs/f2fs/segment.c | 23 +++++++++++++++++++++--
>  3 files changed, 24 insertions(+), 7 deletions(-)
> 
> diff --git a/fs/f2fs/f2fs.h b/fs/f2fs/f2fs.h
> index 6c5e75f86da4..5ae2ca6dba96 100644
> --- a/fs/f2fs/f2fs.h
> +++ b/fs/f2fs/f2fs.h
> @@ -562,6 +562,9 @@ enum {
>  
>  #define MAX_DIR_RA_PAGES     4       /* maximum ra pages of dir */
>  
> +/* dirty segments threshold for triggering CP */
> +#define DEFAULT_DIRTY_THRESHOLD              4
> +
>  /* for in-memory extent cache entry */
>  #define F2FS_MIN_EXTENT_LEN  64      /* minimum extent length */
>  
> diff --git a/fs/f2fs/node.h b/fs/f2fs/node.h
> index ff14a6e5ac1c..18b98cf0465b 100644
> --- a/fs/f2fs/node.h
> +++ b/fs/f2fs/node.h
> @@ -138,11 +138,6 @@ static inline bool excess_cached_nats(struct 
> f2fs_sb_info *sbi)
>       return NM_I(sbi)->nat_cnt[TOTAL_NAT] >= DEF_NAT_CACHE_THRESHOLD;
>  }
>  
> -static inline bool excess_dirty_nodes(struct f2fs_sb_info *sbi)
> -{
> -     return get_pages(sbi, F2FS_DIRTY_NODES) >= sbi->blocks_per_seg * 8;
> -}
> -
>  enum mem_type {
>       FREE_NIDS,      /* indicates the free nid list */
>       NAT_ENTRIES,    /* indicates the cached nat entry */
> diff --git a/fs/f2fs/segment.c b/fs/f2fs/segment.c
> index 7358342652ec..ffd148429a9d 100644
> --- a/fs/f2fs/segment.c
> +++ b/fs/f2fs/segment.c
> @@ -535,6 +535,25 @@ void f2fs_balance_fs(struct f2fs_sb_info *sbi, bool need)
>       }
>  }
>  
> +static inline bool excess_dirty_threshold(struct f2fs_sb_info *sbi)
> +{
> +     int factor = rwsem_is_locked(&sbi->cp_rwsem) ? 3 : 2;
> +     unsigned int dents = get_pages(sbi, F2FS_DIRTY_DENTS);
> +     unsigned int qdata = get_pages(sbi, F2FS_DIRTY_QDATA);
> +     unsigned int nodes = get_pages(sbi, F2FS_DIRTY_NODES);
> +     unsigned int meta = get_pages(sbi, F2FS_DIRTY_META);
> +     unsigned int imeta = get_pages(sbi, F2FS_DIRTY_IMETA);
> +     unsigned int threshold = sbi->blocks_per_seg * factor *
> +                                     DEFAULT_DIRTY_THRESHOLD;
> +     unsigned int global_threshold = threshold * 3 / 2;
> +
> +     if (dents >= threshold || qdata >= threshold ||
> +             nodes >= threshold || meta >= threshold ||
> +             imeta >= threshold)
> +             return true;
> +     return dents + qdata + nodes + meta + imeta >  global_threshold;
> +}
> +
>  void f2fs_balance_fs_bg(struct f2fs_sb_info *sbi, bool from_bg)
>  {
>       if (unlikely(is_sbi_flag_set(sbi, SBI_POR_DOING)))
> @@ -553,8 +572,8 @@ void f2fs_balance_fs_bg(struct f2fs_sb_info *sbi, bool 
> from_bg)
>       else
>               f2fs_build_free_nids(sbi, false, false);
>  
> -     if (excess_dirty_nats(sbi) || excess_dirty_nodes(sbi) ||
> -             excess_prefree_segs(sbi))
> +     if (excess_dirty_nats(sbi) || excess_dirty_threshold(sbi) ||
> +             excess_prefree_segs(sbi) || f2fs_space_for_roll_forward(sbi))

f2fs_space_for_roll_forward() == 0?


>               goto do_sync;
>  
>       /* there is background inflight IO or foreground operation recently */
> -- 
> 2.32.0


_______________________________________________
Linux-f2fs-devel mailing list
Linux-f2fs-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/linux-f2fs-devel

Reply via email to