Hi Jaegeuk,

> -----Original Message-----
> From: Jaegeuk Kim [mailto:jaeg...@kernel.org]
> Sent: Tuesday, September 23, 2014 12:47 PM
> To: Chao Yu
> Cc: linux-kernel@vger.kernel.org; linux-fsde...@vger.kernel.org;
> linux-f2fs-de...@lists.sourceforge.net
> Subject: Re: [f2fs-dev] [PATCH 07/10] f2fs: use meta_inode cache to improve 
> roll-forward speed
> 
> Hi Chao,
> 
> On Mon, Sep 22, 2014 at 10:36:25AM +0800, Chao Yu wrote:
> > Hi Jaegeuk,
> >
> > > -----Original Message-----
> > > From: Jaegeuk Kim [mailto:jaeg...@kernel.org]
> > > Sent: Monday, September 15, 2014 6:14 AM
> > > To: linux-kernel@vger.kernel.org; linux-fsde...@vger.kernel.org;
> > > linux-f2fs-de...@lists.sourceforge.net
> > > Cc: Jaegeuk Kim
> > > Subject: [f2fs-dev] [PATCH 07/10] f2fs: use meta_inode cache to improve 
> > > roll-forward speed
> > >
> > > Previously, all the dnode pages should be read during the roll-forward 
> > > recovery.
> > > Even worsely, whole the chain was traversed twice.
> > > This patch removes that redundant and costly read operations by using 
> > > page cache
> > > of meta_inode and readahead function as well.
> > >
> > > Signed-off-by: Jaegeuk Kim <jaeg...@kernel.org>
> > > ---
> > >  fs/f2fs/checkpoint.c | 11 ++++++++--
> > >  fs/f2fs/f2fs.h       |  5 +++--
> > >  fs/f2fs/recovery.c   | 59 
> > > +++++++++++++++++++++++++---------------------------
> > >  fs/f2fs/segment.h    |  5 +++--
> > >  4 files changed, 43 insertions(+), 37 deletions(-)
> > >
> > > diff --git a/fs/f2fs/checkpoint.c b/fs/f2fs/checkpoint.c
> > > index 7262d99..d1ed889 100644
> > > --- a/fs/f2fs/checkpoint.c
> > > +++ b/fs/f2fs/checkpoint.c
> > > @@ -82,6 +82,8 @@ static inline int get_max_meta_blks(struct f2fs_sb_info 
> > > *sbi, int type)
> > >   case META_SSA:
> > >   case META_CP:
> > >           return 0;
> > > + case META_POR:
> > > +         return SM_I(sbi)->main_blkaddr + sbi->user_block_count;
> >
> > Here we will skip virtual over-provision segments, so better to use 
> > TOTAL_BLKS(sbi).
> >
> > >   default:
> > >           BUG();
> > >   }
> > > @@ -90,11 +92,11 @@ static inline int get_max_meta_blks(struct 
> > > f2fs_sb_info *sbi, int type)
> > >  /*
> > >   * Readahead CP/NAT/SIT/SSA pages
> > >   */
> > > -int ra_meta_pages(struct f2fs_sb_info *sbi, int start, int nrpages, int 
> > > type)
> > > +int ra_meta_pages(struct f2fs_sb_info *sbi, block_t start, int nrpages, 
> > > int type)
> > >  {
> > >   block_t prev_blk_addr = 0;
> > >   struct page *page;
> > > - int blkno = start;
> > > + block_t blkno = start;
> > >   int max_blks = get_max_meta_blks(sbi, type);
> > >
> > >   struct f2fs_io_info fio = {
> > > @@ -128,6 +130,11 @@ int ra_meta_pages(struct f2fs_sb_info *sbi, int 
> > > start, int nrpages,
> int
> > > type)
> > >                   /* get ssa/cp block addr */
> > >                   blk_addr = blkno;
> > >                   break;
> > > +         case META_POR:
> > > +                 if (unlikely(blkno >= max_blks))
> > > +                         goto out;
> > > +                 blk_addr = blkno;
> > > +                 break;
> >
> > The real modification in patch which is merged to dev of f2fs is as 
> > following:
> >
> > - /* get ssa/cp block addr */
> > + case META_POR:
> > + if (blkno >= max_blks || blkno < min_blks)
> > + goto out;
> >
> > IMHO, it's better to verify boundary separately for META_{SSA,CP,POR} with 
> > unlikely.
> > How do you think?
> 
> Not bad.
> Could you check the v2 below?
> 
> >
> > >           default:
> > >                   BUG();
> > >           }
> > > diff --git a/fs/f2fs/f2fs.h b/fs/f2fs/f2fs.h
> > > index 4f84d2a..48d7d46 100644
> > > --- a/fs/f2fs/f2fs.h
> > > +++ b/fs/f2fs/f2fs.h
> > > @@ -103,7 +103,8 @@ enum {
> > >   META_CP,
> > >   META_NAT,
> > >   META_SIT,
> > > - META_SSA
> > > + META_SSA,
> > > + META_POR,
> > >  };
> > >
> > >  /* for the list of ino */
> > > @@ -1291,7 +1292,7 @@ void destroy_segment_manager_caches(void);
> > >   */
> > >  struct page *grab_meta_page(struct f2fs_sb_info *, pgoff_t);
> > >  struct page *get_meta_page(struct f2fs_sb_info *, pgoff_t);
> > > -int ra_meta_pages(struct f2fs_sb_info *, int, int, int);
> > > +int ra_meta_pages(struct f2fs_sb_info *, block_t, int, int);
> > >  long sync_meta_pages(struct f2fs_sb_info *, enum page_type, long);
> > >  void add_dirty_inode(struct f2fs_sb_info *, nid_t, int type);
> > >  void remove_dirty_inode(struct f2fs_sb_info *, nid_t, int type);
> > > diff --git a/fs/f2fs/recovery.c b/fs/f2fs/recovery.c
> > > index 3736728..6f7fbfa 100644
> > > --- a/fs/f2fs/recovery.c
> > > +++ b/fs/f2fs/recovery.c
> > > @@ -173,7 +173,7 @@ static int find_fsync_dnodes(struct f2fs_sb_info 
> > > *sbi, struct list_head
> > > *head)
> > >  {
> > >   unsigned long long cp_ver = cur_cp_version(F2FS_CKPT(sbi));
> > >   struct curseg_info *curseg;
> > > - struct page *page;
> > > + struct page *page = NULL;
> > >   block_t blkaddr;
> > >   int err = 0;
> > >
> > > @@ -181,20 +181,19 @@ static int find_fsync_dnodes(struct f2fs_sb_info 
> > > *sbi, struct list_head
> > > *head)
> > >   curseg = CURSEG_I(sbi, CURSEG_WARM_NODE);
> > >   blkaddr = NEXT_FREE_BLKADDR(sbi, curseg);
> > >
> > > - /* read node page */
> > > - page = alloc_page(GFP_F2FS_ZERO);
> > > - if (!page)
> > > -         return -ENOMEM;
> > > - lock_page(page);
> > > -
> > >   while (1) {
> > >           struct fsync_inode_entry *entry;
> > >
> > > -         err = f2fs_submit_page_bio(sbi, page, blkaddr, READ_SYNC);
> > > -         if (err)
> > > -                 return err;
> > > +         if (blkaddr < SM_I(sbi)->main_blkaddr ||
> > > +                                 blkaddr > TOTAL_BLKS(sbi))
> >
> > blkaddr >= TOTAL_BLKS(sbi) ?
> 
> Right.
> 
> >
> > > +                 break;
> > >
> > > -         lock_page(page);
> > > +         page = get_meta_page(sbi, blkaddr);
> > > +         if (IS_ERR(page))
> > > +                 return PTR_ERR(page);
> > > +
> > > +         ra_meta_pages(sbi, next_blkaddr_of_node(page),
> > > +                         MAX_BIO_BLOCKS(max_hw_blocks(sbi)), META_POR);
> >
> > In first round we readahead [0,n] pages, then in second round we readahead 
> > [1,n+1] pages,
> > but as [1,n] pages is already uptodate, so we will readahead only one page, 
> > the same for the
> > following rounds. It's low efficiency.
> >
> > So how about modify in find_fsync_dnodes/recover_data like this:
> >
> >     block_t blkaddr, ra_start_blk;
> >     int err = 0;
> >
> >     /* get node pages in the current segment */
> >     curseg = CURSEG_I(sbi, CURSEG_WARM_NODE);
> >     blkaddr = NEXT_FREE_BLKADDR(sbi, curseg);
> >     ra_start_blk = blkaddr;
> >
> >     while (1) {
> >             struct fsync_inode_entry *entry;
> >
> >             if (blkaddr < SM_I(sbi)->main_blkaddr ||
> >                                     blkaddr > TOTAL_BLKS(sbi))
> >                     return 0;
> >
> >             if (ra_start_blk == blkaddr) {
> >                     ra_meta_pages(sbi, ra_start_blk, MAX_BIO_BLOCKS(sbi),
> >                                                             META_POR);
> >                     ra_start_blk += MAX_BIO_BLOCKS(sbi);
> >             }
> >
> >             page = get_meta_page(sbi, blkaddr);
> 
> Well. the blkaddr would not be alway able to be matched.
> How about this?
> 
> commit 1cdbd53437f32046af8d94f551505754446c9720
> Author: Jaegeuk Kim <jaeg...@kernel.org>
> Date:   Thu Sep 11 13:49:55 2014 -0700
> 
>     f2fs: use meta_inode cache to improve roll-forward speed
> 
>     Previously, all the dnode pages should be read during the roll-forward 
> recovery.
>     Even worsely, whole the chain was traversed twice.
>     This patch removes that redundant and costly read operations by using 
> page cache
>     of meta_inode and readahead function as well.
> 
>     Signed-off-by: Jaegeuk Kim <jaeg...@kernel.org>

Reviewed-by: Chao Yu <chao2...@samsung.com>

> 
> diff --git a/fs/f2fs/checkpoint.c b/fs/f2fs/checkpoint.c
> index d44d287..700869c 100644
> --- a/fs/f2fs/checkpoint.c
> +++ b/fs/f2fs/checkpoint.c
> @@ -72,7 +72,23 @@ out:
>       return page;
>  }
> 
> -static inline int get_max_meta_blks(struct f2fs_sb_info *sbi, int type)
> +struct page *get_meta_page_ra(struct f2fs_sb_info *sbi, pgoff_t index)
> +{
> +     bool readahead = false;
> +     struct page *page;
> +
> +     page = find_get_page(META_MAPPING(sbi), index);
> +     if (!page || (page && !PageUptodate(page)))
> +             readahead = true;
> +     f2fs_put_page(page, 0);
> +
> +     if (readahead)
> +             ra_meta_pages(sbi, index,
> +                             MAX_BIO_BLOCKS(max_hw_blocks(sbi)), META_POR);
> +     return get_meta_page(sbi, index);
> +}
> +
> +static inline block_t get_max_meta_blks(struct f2fs_sb_info *sbi, int type)
>  {
>       switch (type) {
>       case META_NAT:
> @@ -82,6 +98,8 @@ static inline int get_max_meta_blks(struct f2fs_sb_info 
> *sbi, int type)
>       case META_SSA:
>       case META_CP:
>               return 0;
> +     case META_POR:
> +             return SM_I(sbi)->main_blkaddr + TOTAL_BLKS(sbi);

return SM_I(sbi)->seg0_blkaddr + TOTAL_BLKS(sbi);

>       default:
>               BUG();
>       }
> @@ -90,12 +108,13 @@ static inline int get_max_meta_blks(struct f2fs_sb_info 
> *sbi, int type)
>  /*
>   * Readahead CP/NAT/SIT/SSA pages
>   */
> -int ra_meta_pages(struct f2fs_sb_info *sbi, int start, int nrpages, int type)
> +int ra_meta_pages(struct f2fs_sb_info *sbi, block_t start, int nrpages, int 
> type)
>  {
>       block_t prev_blk_addr = 0;
>       struct page *page;
> -     int blkno = start;
> -     int max_blks = get_max_meta_blks(sbi, type);
> +     block_t blkno = start;
> +     block_t max_blks = get_max_meta_blks(sbi, type);
> +     block_t min_blks = SM_I(sbi)->seg0_blkaddr;
> 
>       struct f2fs_io_info fio = {
>               .type = META,
> @@ -125,7 +144,11 @@ int ra_meta_pages(struct f2fs_sb_info *sbi, int start, 
> int nrpages, int
> type)
>                       break;
>               case META_SSA:
>               case META_CP:
> -                     /* get ssa/cp block addr */
> +             case META_POR:
> +                     if (unlikely(blkno >= max_blks))
> +                             goto out;
> +                     if (unlikely(blkno < min_blks))
> +                             goto out;
>                       blk_addr = blkno;
>                       break;
>               default:
> diff --git a/fs/f2fs/f2fs.h b/fs/f2fs/f2fs.h
> index 8efa170..b6439c3 100644
> --- a/fs/f2fs/f2fs.h
> +++ b/fs/f2fs/f2fs.h
> @@ -103,7 +103,8 @@ enum {
>       META_CP,
>       META_NAT,
>       META_SIT,
> -     META_SSA
> +     META_SSA,
> +     META_POR,
>  };
> 
>  /* for the list of ino */
> @@ -1294,7 +1295,8 @@ void destroy_segment_manager_caches(void);
>   */
>  struct page *grab_meta_page(struct f2fs_sb_info *, pgoff_t);
>  struct page *get_meta_page(struct f2fs_sb_info *, pgoff_t);
> -int ra_meta_pages(struct f2fs_sb_info *, int, int, int);
> +struct page *get_meta_page_ra(struct f2fs_sb_info *, pgoff_t);
> +int ra_meta_pages(struct f2fs_sb_info *, block_t, int, int);
>  long sync_meta_pages(struct f2fs_sb_info *, enum page_type, long);
>  void add_dirty_inode(struct f2fs_sb_info *, nid_t, int type);
>  void remove_dirty_inode(struct f2fs_sb_info *, nid_t, int type);
> diff --git a/fs/f2fs/recovery.c b/fs/f2fs/recovery.c
> index e587ee1..3a259a9 100644
> --- a/fs/f2fs/recovery.c
> +++ b/fs/f2fs/recovery.c
> @@ -138,7 +138,7 @@ static int find_fsync_dnodes(struct f2fs_sb_info *sbi, 
> struct list_head
> *head)
>  {
>       unsigned long long cp_ver = cur_cp_version(F2FS_CKPT(sbi));
>       struct curseg_info *curseg;
> -     struct page *page;
> +     struct page *page = NULL;
>       block_t blkaddr;
>       int err = 0;
> 
> @@ -146,20 +146,14 @@ static int find_fsync_dnodes(struct f2fs_sb_info *sbi, 
> struct list_head
> *head)
>       curseg = CURSEG_I(sbi, CURSEG_WARM_NODE);
>       blkaddr = NEXT_FREE_BLKADDR(sbi, curseg);
> 
> -     /* read node page */
> -     page = alloc_page(GFP_F2FS_ZERO);
> -     if (!page)
> -             return -ENOMEM;
> -     lock_page(page);
> -
>       while (1) {
>               struct fsync_inode_entry *entry;
> 
> -             err = f2fs_submit_page_bio(sbi, page, blkaddr, READ_SYNC);
> -             if (err)
> -                     return err;
> +             if (blkaddr < SM_I(sbi)->main_blkaddr ||
> +                                     blkaddr >= TOTAL_BLKS(sbi))

Here, upper boundary should be the same as value in get_max_meta_blks().

blkaddr >= SM_I(sbi)->seg0_blkaddr + TOTAL_BLKS(sbi)

> +                     return 0;
> 
> -             lock_page(page);
> +             page = get_meta_page_ra(sbi, blkaddr);
> 
>               if (cp_ver != cpver_of_node(page))
>                       break;
> @@ -202,11 +196,9 @@ static int find_fsync_dnodes(struct f2fs_sb_info *sbi, 
> struct list_head
> *head)
>  next:
>               /* check next segment */
>               blkaddr = next_blkaddr_of_node(page);
> +             f2fs_put_page(page, 1);
>       }
> -
> -     unlock_page(page);
> -     __free_pages(page, 0);
> -
> +     f2fs_put_page(page, 1);
>       return err;
>  }
> 
> @@ -400,7 +392,7 @@ static int recover_data(struct f2fs_sb_info *sbi,
>  {
>       unsigned long long cp_ver = cur_cp_version(F2FS_CKPT(sbi));
>       struct curseg_info *curseg;
> -     struct page *page;
> +     struct page *page = NULL;
>       int err = 0;
>       block_t blkaddr;
> 
> @@ -408,32 +400,29 @@ static int recover_data(struct f2fs_sb_info *sbi,
>       curseg = CURSEG_I(sbi, type);
>       blkaddr = NEXT_FREE_BLKADDR(sbi, curseg);
> 
> -     /* read node page */
> -     page = alloc_page(GFP_F2FS_ZERO);
> -     if (!page)
> -             return -ENOMEM;
> -
> -     lock_page(page);
> -
>       while (1) {
>               struct fsync_inode_entry *entry;
> 
> -             err = f2fs_submit_page_bio(sbi, page, blkaddr, READ_SYNC);
> -             if (err)
> -                     return err;
> +             if (blkaddr < SM_I(sbi)->main_blkaddr ||
> +                                     blkaddr > TOTAL_BLKS(sbi))

blkaddr >= SM_I(sbi)->seg0_blkaddr + TOTAL_BLKS(sbi)

It's my mistaken that I do not reminder you to fix this in recover_data.

Thanks,
Yu

> +                     break;
> 
> -             lock_page(page);
> +             page = get_meta_page_ra(sbi, blkaddr);
> 
> -             if (cp_ver != cpver_of_node(page))
> +             if (cp_ver != cpver_of_node(page)) {
> +                     f2fs_put_page(page, 1);
>                       break;
> +             }
> 
>               entry = get_fsync_inode(head, ino_of_node(page));
>               if (!entry)
>                       goto next;
> 
>               err = do_recover_data(sbi, entry->inode, page, blkaddr);
> -             if (err)
> +             if (err) {
> +                     f2fs_put_page(page, 1);
>                       break;
> +             }
> 
>               if (entry->blkaddr == blkaddr) {
>                       iput(entry->inode);
> @@ -443,11 +432,8 @@ static int recover_data(struct f2fs_sb_info *sbi,
>  next:
>               /* check next segment */
>               blkaddr = next_blkaddr_of_node(page);
> +             f2fs_put_page(page, 1);
>       }
> -
> -     unlock_page(page);
> -     __free_pages(page, 0);
> -
>       if (!err)
>               allocate_new_segments(sbi);
>       return err;
> @@ -493,6 +479,10 @@ out:
>       destroy_fsync_dnodes(&inode_list);
>       kmem_cache_destroy(fsync_entry_slab);
> 
> +     /* truncate meta pages to be used by the recovery */
> +     truncate_inode_pages_range(META_MAPPING(sbi),
> +             SM_I(sbi)->main_blkaddr << PAGE_CACHE_SHIFT, -1);
> +
>       if (err) {
>               truncate_inode_pages_final(NODE_MAPPING(sbi));
>               truncate_inode_pages_final(META_MAPPING(sbi));
> diff --git a/fs/f2fs/segment.h b/fs/f2fs/segment.h
> index 013aed2..c6f37f2 100644
> --- a/fs/f2fs/segment.h
> +++ b/fs/f2fs/segment.h
> @@ -87,6 +87,7 @@
>       (BITS_TO_LONGS(nr) * sizeof(unsigned long))
>  #define TOTAL_SEGS(sbi)      (SM_I(sbi)->main_segments)
>  #define TOTAL_SECS(sbi)      (sbi->total_sections)
> +#define TOTAL_BLKS(sbi)      (SM_I(sbi)->segment_count << 
> sbi->log_blocks_per_seg)
> 
>  #define SECTOR_FROM_BLOCK(sbi, blk_addr)                             \
>       (((sector_t)blk_addr) << (sbi)->log_sectors_per_block)
> @@ -553,7 +554,7 @@ static inline void check_seg_range(struct f2fs_sb_info 
> *sbi, unsigned int
> segno)
>  static inline void verify_block_addr(struct f2fs_sb_info *sbi, block_t 
> blk_addr)
>  {
>       struct f2fs_sm_info *sm_info = SM_I(sbi);
> -     block_t total_blks = sm_info->segment_count << sbi->log_blocks_per_seg;
> +     block_t total_blks = TOTAL_BLKS(sbi);
>       block_t start_addr = sm_info->seg0_blkaddr;
>       block_t end_addr = start_addr + total_blks - 1;
>       BUG_ON(blk_addr < start_addr);
> @@ -606,7 +607,7 @@ static inline void check_seg_range(struct f2fs_sb_info 
> *sbi, unsigned int
> segno)
>  static inline void verify_block_addr(struct f2fs_sb_info *sbi, block_t 
> blk_addr)
>  {
>       struct f2fs_sm_info *sm_info = SM_I(sbi);
> -     block_t total_blks = sm_info->segment_count << sbi->log_blocks_per_seg;
> +     block_t total_blks = TOTAL_BLKS(sbi);
>       block_t start_addr = sm_info->seg0_blkaddr;
>       block_t end_addr = start_addr + total_blks - 1;
> 

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to