@@ -11,6 +11,7 @@
#include "ctree.h"
#include "extent_io.h"
#include "locking.h"
+#include "perf.h"
static void btrfs_assert_tree_read_locked(struct extent_buffer *eb);
@@ -85,10 +86,14 @@ void btrfs_clear_lock_blocking_rw(struct extent_buffer *eb, int rw)
*/
void btrfs_tree_read_lock(struct extent_buffer *eb)
{
+ u64 start_ns;
+
+ start_ns = btrfs_perf_start();
again:
BUG_ON(!atomic_read(&eb->blocking_writers) &&
current->pid == eb->lock_owner);
+
read_lock(&eb->lock);
if (atomic_read(&eb->blocking_writers) &&
current->pid == eb->lock_owner) {
@@ -101,16 +106,19 @@ void btrfs_tree_read_lock(struct extent_buffer *eb)
BUG_ON(eb->lock_nested);
eb->lock_nested = 1;
read_unlock(&eb->lock);
+ btrfs_perf_end(eb->fs_info, btrfs_header_owner(eb), start_ns);
return;
}
if (atomic_read(&eb->blocking_writers)) {
read_unlock(&eb->lock);
+
wait_event(eb->write_lock_wq,
atomic_read(&eb->blocking_writers) == 0);
goto again;
}
atomic_inc(&eb->read_locks);
atomic_inc(&eb->spinning_readers);
+ btrfs_perf_end(eb->fs_info, btrfs_header_owner(eb), start_ns);
}
/*
@@ -227,7 +235,9 @@ void btrfs_tree_read_unlock_blocking(struct extent_buffer *eb)
*/
void btrfs_tree_lock(struct extent_buffer *eb)
{
+ u64 start_ns;
WARN_ON(eb->lock_owner == current->pid);
+ start_ns = btrfs_perf_start();
again:
wait_event(eb->read_lock_wq, atomic_read(&eb->blocking_readers) == 0);
wait_event(eb->write_lock_wq, atomic_read(&eb->blocking_writers) == 0);
@@ -248,6 +258,7 @@ void btrfs_tree_lock(struct extent_buffer *eb)
atomic_inc(&eb->spinning_writers);
atomic_inc(&eb->write_locks);
eb->lock_owner = current->pid;
+ btrfs_perf_end(eb->fs_info, btrfs_header_owner(eb), start_ns);
}
/*
@@ -50,3 +50,23 @@ void btrfs_perf_free_profiler(struct btrfs_fs_info *fs_info)
kfree(profiler);
}
+void btrfs_perf_end(struct btrfs_fs_info *fs_info, u64 eb_owner, u64 start_ns)
+{
+ struct btrfs_perf_profiler *profiler = fs_info->profiler;
+ u64 end_ns;
+ int i;
+
+ if (!profiler)
+ return;
+
+ end_ns = ktime_get_ns();
+ if (eb_owner == BTRFS_ROOT_TREE_OBJECTID)
+ i = BTRFS_PERF_TREE_LOCK_ROOT;
+ else if (is_fstree(eb_owner))
+ i = BTRFS_PERF_TREE_LOCK_FS;
+ else if (eb_owner == BTRFS_EXTENT_TREE_OBJECTID)
+ i = BTRFS_PERF_TREE_LOCK_EXTENT;
+ else
+ i = BTRFS_PERF_TREE_LOCK_OTHER;
+ percpu_counter_add(&profiler->perf_counters[i], end_ns - start_ns);
+}
@@ -24,4 +24,12 @@ struct btrfs_perf_profiler {
struct btrfs_perf_profiler *btrfs_perf_alloc_profiler(void);
void btrfs_perf_free_profiler(struct btrfs_fs_info *fs_info);
+void btrfs_perf_update_lock(struct btrfs_fs_info *fs_info,
+ u64 eb_owner, u64 ns_diff);
+static inline u64 btrfs_perf_start(void)
+{
+ return ktime_get_ns();
+}
+
+void btrfs_perf_end(struct btrfs_fs_info *fs_info, u64 eb_owner, u64 start_ns);
#endif
For btrfs tree locking, there are only 2 functions can sleep: - btrfs_tree_read_lock() It will wait for any blocking writers - btrfs_tree_lock() It will wait for any blocking readers or writers Other functions only depends on rwlock which won't sleep. We doesn't really care about the spinning lock version. The overheads introduced are: - two ktime_get() calls - several if branches - percpu_counter_add() Which should be smaller than to ftrace function_graph. Signed-off-by: Qu Wenruo <wqu@suse.com> --- fs/btrfs/locking.c | 11 +++++++++++ fs/btrfs/perf.c | 20 ++++++++++++++++++++ fs/btrfs/perf.h | 8 ++++++++ 3 files changed, 39 insertions(+)