On Tue, Nov 27, 2018 at 10:29:57AM +0200, Nikolay Borisov wrote:
> 
> 
> On 21.11.18 г. 21:09 ч., Josef Bacik wrote:
> > The throttle path doesn't take cleaner_delayed_iput_mutex, which means
> 
> Which one is the throttle path? btrfs_end_transaction_throttle is only
> called during snapshot drop and relocation? What scenario triggered your
> observation and prompted this fix?
> 

One of my enospc tests runs snapshot creation/deletion in the background.

> > we could think we're done flushing iputs in the data space reservation
> > path when we could have a throttler doing an iput.  There's no real
> > reason to serialize the delayed iput flushing, so instead of taking the
> > cleaner_delayed_iput_mutex whenever we flush the delayed iputs just
> > replace it with an atomic counter and a waitqueue.  This removes the
> > short (or long depending on how big the inode is) window where we think
> > there are no more pending iputs when there really are some.
> > 
> > Signed-off-by: Josef Bacik <jo...@toxicpanda.com>
> > ---
> >  fs/btrfs/ctree.h       |  4 +++-
> >  fs/btrfs/disk-io.c     |  5 ++---
> >  fs/btrfs/extent-tree.c |  9 +++++----
> >  fs/btrfs/inode.c       | 21 +++++++++++++++++++++
> >  4 files changed, 31 insertions(+), 8 deletions(-)
> > 
> > diff --git a/fs/btrfs/ctree.h b/fs/btrfs/ctree.h
> > index 709de7471d86..a835fe7076eb 100644
> > --- a/fs/btrfs/ctree.h
> > +++ b/fs/btrfs/ctree.h
> > @@ -912,7 +912,8 @@ struct btrfs_fs_info {
> >  
> >     spinlock_t delayed_iput_lock;
> >     struct list_head delayed_iputs;
> > -   struct mutex cleaner_delayed_iput_mutex;
> > +   atomic_t nr_delayed_iputs;
> > +   wait_queue_head_t delayed_iputs_wait;
> >  
> >     /* this protects tree_mod_seq_list */
> >     spinlock_t tree_mod_seq_lock;
> > @@ -3237,6 +3238,7 @@ int btrfs_orphan_cleanup(struct btrfs_root *root);
> >  int btrfs_cont_expand(struct inode *inode, loff_t oldsize, loff_t size);
> >  void btrfs_add_delayed_iput(struct inode *inode);
> >  void btrfs_run_delayed_iputs(struct btrfs_fs_info *fs_info);
> > +int btrfs_wait_on_delayed_iputs(struct btrfs_fs_info *fs_info);
> >  int btrfs_prealloc_file_range(struct inode *inode, int mode,
> >                           u64 start, u64 num_bytes, u64 min_size,
> >                           loff_t actual_len, u64 *alloc_hint);
> > diff --git a/fs/btrfs/disk-io.c b/fs/btrfs/disk-io.c
> > index c5918ff8241b..3f81dfaefa32 100644
> > --- a/fs/btrfs/disk-io.c
> > +++ b/fs/btrfs/disk-io.c
> > @@ -1692,9 +1692,7 @@ static int cleaner_kthread(void *arg)
> >                     goto sleep;
> >             }
> >  
> > -           mutex_lock(&fs_info->cleaner_delayed_iput_mutex);
> >             btrfs_run_delayed_iputs(fs_info);
> > -           mutex_unlock(&fs_info->cleaner_delayed_iput_mutex);
> >  
> >             again = btrfs_clean_one_deleted_snapshot(root);
> >             mutex_unlock(&fs_info->cleaner_mutex);
> > @@ -2651,7 +2649,6 @@ int open_ctree(struct super_block *sb,
> >     mutex_init(&fs_info->delete_unused_bgs_mutex);
> >     mutex_init(&fs_info->reloc_mutex);
> >     mutex_init(&fs_info->delalloc_root_mutex);
> > -   mutex_init(&fs_info->cleaner_delayed_iput_mutex);
> >     seqlock_init(&fs_info->profiles_lock);
> >  
> >     INIT_LIST_HEAD(&fs_info->dirty_cowonly_roots);
> > @@ -2673,6 +2670,7 @@ int open_ctree(struct super_block *sb,
> >     atomic_set(&fs_info->defrag_running, 0);
> >     atomic_set(&fs_info->qgroup_op_seq, 0);
> >     atomic_set(&fs_info->reada_works_cnt, 0);
> > +   atomic_set(&fs_info->nr_delayed_iputs, 0);
> >     atomic64_set(&fs_info->tree_mod_seq, 0);
> >     fs_info->sb = sb;
> >     fs_info->max_inline = BTRFS_DEFAULT_MAX_INLINE;
> > @@ -2750,6 +2748,7 @@ int open_ctree(struct super_block *sb,
> >     init_waitqueue_head(&fs_info->transaction_wait);
> >     init_waitqueue_head(&fs_info->transaction_blocked_wait);
> >     init_waitqueue_head(&fs_info->async_submit_wait);
> > +   init_waitqueue_head(&fs_info->delayed_iputs_wait);
> >  
> >     INIT_LIST_HEAD(&fs_info->pinned_chunks);
> >  
> > diff --git a/fs/btrfs/extent-tree.c b/fs/btrfs/extent-tree.c
> > index 3a90dc1d6b31..36f43876be22 100644
> > --- a/fs/btrfs/extent-tree.c
> > +++ b/fs/btrfs/extent-tree.c
> > @@ -4272,8 +4272,9 @@ int btrfs_alloc_data_chunk_ondemand(struct 
> > btrfs_inode *inode, u64 bytes)
> >                              * operations. Wait for it to finish so that
> >                              * more space is released.
> >                              */
> > -                           
> > mutex_lock(&fs_info->cleaner_delayed_iput_mutex);
> > -                           
> > mutex_unlock(&fs_info->cleaner_delayed_iput_mutex);
> > +                           ret = btrfs_wait_on_delayed_iputs(fs_info);
> > +                           if (ret)
> > +                                   return ret;
> >                             goto again;
> >                     } else {
> >                             btrfs_end_transaction(trans);
> > @@ -4838,9 +4839,9 @@ static int may_commit_transaction(struct 
> > btrfs_fs_info *fs_info,
> >      * pinned space, so make sure we run the iputs before we do our pinned
> >      * bytes check below.
> >      */
> > -   mutex_lock(&fs_info->cleaner_delayed_iput_mutex);
> >     btrfs_run_delayed_iputs(fs_info);
> > -   mutex_unlock(&fs_info->cleaner_delayed_iput_mutex);
> > +   wait_event(fs_info->delayed_iputs_wait,
> > +              atomic_read(&fs_info->nr_delayed_iputs) == 0);
> 
> Why open code btrfs_wait_on_delayed_iputs(fs_info)? Just make it use
> wait_event instead of the killable version and use the function
> uniformally across the code base.

I don't want the flusher thread to be killable, but now that I've said that out
loud I realize that's not possible, so I'll just use the helper.  Thanks,

Josef

Reply via email to