diff mbox series

[RFC] btrfs: zoned: auto reclaim low mostly full block-groups first

Message ID a5b7e730eeeeebd701d807f7aa950dc1f52caade.1632150570.git.johannes.thumshirn@wdc.com (mailing list archive)
State New, archived
Headers show
Series [RFC] btrfs: zoned: auto reclaim low mostly full block-groups first | expand

Commit Message

Johannes Thumshirn Sept. 20, 2021, 3:11 p.m. UTC
Currently auto reclaim of unusable zones reclaims the block-groups in the
order they have been added to the reclaim list.

Sort the list so  we have the block-groups with the least amount of bytes
to preserve at the beginning before starting the garbage collection loop.

Signed-off-by: Johannes Thumshirn <johannes.thumshirn@wdc.com>
---
 fs/btrfs/block-group.c | 18 ++++++++++++++++++
 1 file changed, 18 insertions(+)

Comments

David Sterba Sept. 20, 2021, 3:50 p.m. UTC | #1
On Tue, Sep 21, 2021 at 12:11:01AM +0900, Johannes Thumshirn wrote:
> Currently auto reclaim of unusable zones reclaims the block-groups in the
> order they have been added to the reclaim list.
> 
> Sort the list so  we have the block-groups with the least amount of bytes
> to preserve at the beginning before starting the garbage collection loop.

Makes sense as an optimization.

> Signed-off-by: Johannes Thumshirn <johannes.thumshirn@wdc.com>
> ---
>  fs/btrfs/block-group.c | 18 ++++++++++++++++++
>  1 file changed, 18 insertions(+)
> 
> diff --git a/fs/btrfs/block-group.c b/fs/btrfs/block-group.c
> index 46fdef7bbe20..d90297fb99e1 100644
> --- a/fs/btrfs/block-group.c
> +++ b/fs/btrfs/block-group.c
> @@ -1,5 +1,7 @@
>  // SPDX-License-Identifier: GPL-2.0
>  
> +#include <linux/list_sort.h>
> +
>  #include "misc.h"
>  #include "ctree.h"
>  #include "block-group.h"
> @@ -1486,6 +1488,21 @@ void btrfs_mark_bg_unused(struct btrfs_block_group *bg)
>  	spin_unlock(&fs_info->unused_bgs_lock);
>  }
>  
> +/*
> + * We want block groups with a low number of used bytes to be in the beginning
> + * of the list, so they will get reclaimed first.
> + */
> +static int reclaim_bgs_cmp(void *unused, const struct list_head *a,
> +			   const struct list_head *b)
> +{
> +	const struct btrfs_block_group *bg1, *bg2;
> +
> +	bg1 = list_entry(a, struct btrfs_block_group, bg_list);
> +	bg2 = list_entry(b, struct btrfs_block_group, bg_list);
> +
> +	return bg1->used - bg2->used;
> +}
> +
>  void btrfs_reclaim_bgs_work(struct work_struct *work)
>  {
>  	struct btrfs_fs_info *fs_info =
> @@ -1510,6 +1527,7 @@ void btrfs_reclaim_bgs_work(struct work_struct *work)
>  	}
>  
>  	spin_lock(&fs_info->unused_bgs_lock);
> +	list_sort(NULL, &fs_info->reclaim_bgs, reclaim_bgs_cmp);

The sort is under a spinlock, though it's probably not a highly
contended lock, I think we should try to move it outside. Something like

  lock()
  list_splice_init(&splice, &reclaim_bgs)
  unlock()

  list_sort(&splice);

  while (!list_empty(splice)) {
  }

We already use splice in the again_list so it could build on top of it.

OTOH, it may not be absolutelly necessary to do the sort outside of the
lock but rather because as a matter of good programming hygiene to not
introduce unnecessary delays due to contended lock here and there that
could potentially cascade further.
Johannes Thumshirn Sept. 21, 2021, 7:31 a.m. UTC | #2
On 20/09/2021 17:50, David Sterba wrote:
> On Tue, Sep 21, 2021 at 12:11:01AM +0900, Johannes Thumshirn wrote:
>> Currently auto reclaim of unusable zones reclaims the block-groups in the
>> order they have been added to the reclaim list.
>>
>> Sort the list so  we have the block-groups with the least amount of bytes
>> to preserve at the beginning before starting the garbage collection loop.
> 
> Makes sense as an optimization.
> 
>> Signed-off-by: Johannes Thumshirn <johannes.thumshirn@wdc.com>
>> ---
>>  fs/btrfs/block-group.c | 18 ++++++++++++++++++
>>  1 file changed, 18 insertions(+)
>>
>> diff --git a/fs/btrfs/block-group.c b/fs/btrfs/block-group.c
>> index 46fdef7bbe20..d90297fb99e1 100644
>> --- a/fs/btrfs/block-group.c
>> +++ b/fs/btrfs/block-group.c
>> @@ -1,5 +1,7 @@
>>  // SPDX-License-Identifier: GPL-2.0
>>  
>> +#include <linux/list_sort.h>
>> +
>>  #include "misc.h"
>>  #include "ctree.h"
>>  #include "block-group.h"
>> @@ -1486,6 +1488,21 @@ void btrfs_mark_bg_unused(struct btrfs_block_group *bg)
>>  	spin_unlock(&fs_info->unused_bgs_lock);
>>  }
>>  
>> +/*
>> + * We want block groups with a low number of used bytes to be in the beginning
>> + * of the list, so they will get reclaimed first.
>> + */
>> +static int reclaim_bgs_cmp(void *unused, const struct list_head *a,
>> +			   const struct list_head *b)
>> +{
>> +	const struct btrfs_block_group *bg1, *bg2;
>> +
>> +	bg1 = list_entry(a, struct btrfs_block_group, bg_list);
>> +	bg2 = list_entry(b, struct btrfs_block_group, bg_list);
>> +
>> +	return bg1->used - bg2->used;
>> +}
>> +
>>  void btrfs_reclaim_bgs_work(struct work_struct *work)
>>  {
>>  	struct btrfs_fs_info *fs_info =
>> @@ -1510,6 +1527,7 @@ void btrfs_reclaim_bgs_work(struct work_struct *work)
>>  	}
>>  
>>  	spin_lock(&fs_info->unused_bgs_lock);
>> +	list_sort(NULL, &fs_info->reclaim_bgs, reclaim_bgs_cmp);
> 
> The sort is under a spinlock, though it's probably not a highly
> contended lock, I think we should try to move it outside. Something like
> 
>   lock()
>   list_splice_init(&splice, &reclaim_bgs)
>   unlock()
> 
>   list_sort(&splice);
> 
>   while (!list_empty(splice)) {
>   }
> 
> We already use splice in the again_list so it could build on top of it.
> 
> OTOH, it may not be absolutelly necessary to do the sort outside of the
> lock but rather because as a matter of good programming hygiene to not
> introduce unnecessary delays due to contended lock here and there that
> could potentially cascade further.
> 

I'm expecting the number of entries in the list to be in the single or two
digit range, so the sort should be rather quick. 

But I agree using a spliced list looks more future proof.
diff mbox series

Patch

diff --git a/fs/btrfs/block-group.c b/fs/btrfs/block-group.c
index 46fdef7bbe20..d90297fb99e1 100644
--- a/fs/btrfs/block-group.c
+++ b/fs/btrfs/block-group.c
@@ -1,5 +1,7 @@ 
 // SPDX-License-Identifier: GPL-2.0
 
+#include <linux/list_sort.h>
+
 #include "misc.h"
 #include "ctree.h"
 #include "block-group.h"
@@ -1486,6 +1488,21 @@  void btrfs_mark_bg_unused(struct btrfs_block_group *bg)
 	spin_unlock(&fs_info->unused_bgs_lock);
 }
 
+/*
+ * We want block groups with a low number of used bytes to be in the beginning
+ * of the list, so they will get reclaimed first.
+ */
+static int reclaim_bgs_cmp(void *unused, const struct list_head *a,
+			   const struct list_head *b)
+{
+	const struct btrfs_block_group *bg1, *bg2;
+
+	bg1 = list_entry(a, struct btrfs_block_group, bg_list);
+	bg2 = list_entry(b, struct btrfs_block_group, bg_list);
+
+	return bg1->used - bg2->used;
+}
+
 void btrfs_reclaim_bgs_work(struct work_struct *work)
 {
 	struct btrfs_fs_info *fs_info =
@@ -1510,6 +1527,7 @@  void btrfs_reclaim_bgs_work(struct work_struct *work)
 	}
 
 	spin_lock(&fs_info->unused_bgs_lock);
+	list_sort(NULL, &fs_info->reclaim_bgs, reclaim_bgs_cmp);
 	while (!list_empty(&fs_info->reclaim_bgs)) {
 		u64 zone_unusable;
 		int ret = 0;