For btrfs tree locking, there are only 2 functions can sleep:
- btrfs_tree_read_lock()
  It will wait for any blocking writers
- btrfs_tree_lock()
  It will wait for any blocking readers or writers

Other functions only depends on rwlock which won't sleep.
We doesn't really care about the spinning lock version.

The overheads introduced are:
- two ktime_get() calls
- several if branches
- percpu_counter_add()

Which should be smaller than to ftrace function_graph.

Signed-off-by: Qu Wenruo <w...@suse.com>
---
 fs/btrfs/locking.c | 11 +++++++++++
 fs/btrfs/perf.c    | 20 ++++++++++++++++++++
 fs/btrfs/perf.h    |  8 ++++++++
 3 files changed, 39 insertions(+)

diff --git a/fs/btrfs/locking.c b/fs/btrfs/locking.c
index 1da768e5ef75..c23c8f7450e8 100644
--- a/fs/btrfs/locking.c
+++ b/fs/btrfs/locking.c
@@ -11,6 +11,7 @@
 #include "ctree.h"
 #include "extent_io.h"
 #include "locking.h"
+#include "perf.h"
 
 static void btrfs_assert_tree_read_locked(struct extent_buffer *eb);
 
@@ -85,10 +86,14 @@ void btrfs_clear_lock_blocking_rw(struct extent_buffer *eb, 
int rw)
  */
 void btrfs_tree_read_lock(struct extent_buffer *eb)
 {
+       u64 start_ns;
+
+       start_ns = btrfs_perf_start();
 again:
        BUG_ON(!atomic_read(&eb->blocking_writers) &&
               current->pid == eb->lock_owner);
 
+
        read_lock(&eb->lock);
        if (atomic_read(&eb->blocking_writers) &&
            current->pid == eb->lock_owner) {
@@ -101,16 +106,19 @@ void btrfs_tree_read_lock(struct extent_buffer *eb)
                BUG_ON(eb->lock_nested);
                eb->lock_nested = 1;
                read_unlock(&eb->lock);
+               btrfs_perf_end(eb->fs_info, btrfs_header_owner(eb), start_ns);
                return;
        }
        if (atomic_read(&eb->blocking_writers)) {
                read_unlock(&eb->lock);
+
                wait_event(eb->write_lock_wq,
                           atomic_read(&eb->blocking_writers) == 0);
                goto again;
        }
        atomic_inc(&eb->read_locks);
        atomic_inc(&eb->spinning_readers);
+       btrfs_perf_end(eb->fs_info, btrfs_header_owner(eb), start_ns);
 }
 
 /*
@@ -227,7 +235,9 @@ void btrfs_tree_read_unlock_blocking(struct extent_buffer 
*eb)
  */
 void btrfs_tree_lock(struct extent_buffer *eb)
 {
+       u64 start_ns;
        WARN_ON(eb->lock_owner == current->pid);
+       start_ns = btrfs_perf_start();
 again:
        wait_event(eb->read_lock_wq, atomic_read(&eb->blocking_readers) == 0);
        wait_event(eb->write_lock_wq, atomic_read(&eb->blocking_writers) == 0);
@@ -248,6 +258,7 @@ void btrfs_tree_lock(struct extent_buffer *eb)
        atomic_inc(&eb->spinning_writers);
        atomic_inc(&eb->write_locks);
        eb->lock_owner = current->pid;
+       btrfs_perf_end(eb->fs_info, btrfs_header_owner(eb), start_ns);
 }
 
 /*
diff --git a/fs/btrfs/perf.c b/fs/btrfs/perf.c
index 2880111f3b0c..893bfad8e6d3 100644
--- a/fs/btrfs/perf.c
+++ b/fs/btrfs/perf.c
@@ -50,3 +50,23 @@ void btrfs_perf_free_profiler(struct btrfs_fs_info *fs_info)
        kfree(profiler);
 }
 
+void btrfs_perf_end(struct btrfs_fs_info *fs_info, u64 eb_owner, u64 start_ns)
+{
+       struct btrfs_perf_profiler *profiler = fs_info->profiler;
+       u64 end_ns;
+       int i;
+
+       if (!profiler)
+               return;
+
+       end_ns = ktime_get_ns();
+       if (eb_owner == BTRFS_ROOT_TREE_OBJECTID)
+               i = BTRFS_PERF_TREE_LOCK_ROOT;
+       else if (is_fstree(eb_owner))
+               i = BTRFS_PERF_TREE_LOCK_FS;
+       else if (eb_owner == BTRFS_EXTENT_TREE_OBJECTID)
+               i = BTRFS_PERF_TREE_LOCK_EXTENT;
+       else
+               i = BTRFS_PERF_TREE_LOCK_OTHER;
+       percpu_counter_add(&profiler->perf_counters[i], end_ns - start_ns);
+}
diff --git a/fs/btrfs/perf.h b/fs/btrfs/perf.h
index 9dbea6458d86..7cf4b8c9a0ad 100644
--- a/fs/btrfs/perf.h
+++ b/fs/btrfs/perf.h
@@ -24,4 +24,12 @@ struct btrfs_perf_profiler {
 
 struct btrfs_perf_profiler *btrfs_perf_alloc_profiler(void);
 void btrfs_perf_free_profiler(struct btrfs_fs_info *fs_info);
+void btrfs_perf_update_lock(struct btrfs_fs_info *fs_info,
+                           u64 eb_owner, u64 ns_diff);
+static inline u64 btrfs_perf_start(void)
+{
+       return ktime_get_ns();
+}
+
+void btrfs_perf_end(struct btrfs_fs_info *fs_info, u64 eb_owner, u64 start_ns);
 #endif
-- 
2.21.0

Reply via email to