diff mbox series

[2/3] dm zoned: metadata: use refcount_t for dm zoned reference counters

Message ID 20180823173557.19665-3-jpittman@redhat.com (mailing list archive)
State Changes Requested, archived
Delegated to: Mike Snitzer
Headers show
Series dm: replace atomic_t reference counters with refcount_t | expand

Commit Message

John Pittman Aug. 23, 2018, 5:35 p.m. UTC
The API surrounding refcount_t should be used in place of atomic_t
when variables are being used as reference counters.  This API can
prevent issues such as counter overflows and use-after-free
conditions.  Within the dm zoned metadata stack, the atomic_t API
is used for mblk->ref and zone->refcount.  Change these to use
refcount_t, avoiding the issues mentioned.

Signed-off-by: John Pittman <jpittman@redhat.com>
---
 drivers/md/dm-zoned-metadata.c | 25 +++++++++++++------------
 drivers/md/dm-zoned.h          |  2 +-
 2 files changed, 14 insertions(+), 13 deletions(-)

Comments

Damien Le Moal Aug. 23, 2018, 10:12 p.m. UTC | #1
John,

On 2018/08/23 10:37, John Pittman wrote:
> The API surrounding refcount_t should be used in place of atomic_t
> when variables are being used as reference counters.  This API can
> prevent issues such as counter overflows and use-after-free
> conditions.  Within the dm zoned metadata stack, the atomic_t API
> is used for mblk->ref and zone->refcount.  Change these to use
> refcount_t, avoiding the issues mentioned.
> 
> Signed-off-by: John Pittman <jpittman@redhat.com>
> ---
>  drivers/md/dm-zoned-metadata.c | 25 +++++++++++++------------
>  drivers/md/dm-zoned.h          |  2 +-
>  2 files changed, 14 insertions(+), 13 deletions(-)
> 
> diff --git a/drivers/md/dm-zoned-metadata.c b/drivers/md/dm-zoned-metadata.c
> index 969954915566..92e635749414 100644
> --- a/drivers/md/dm-zoned-metadata.c
> +++ b/drivers/md/dm-zoned-metadata.c
> @@ -99,7 +99,7 @@ struct dmz_mblock {
>  	struct rb_node		node;
>  	struct list_head	link;
>  	sector_t		no;
> -	atomic_t		ref;
> +	refcount_t		ref;

While reviewing your patch, I realized that this ref is always manipulated under
the zmd->mblk_lock spinlock. So there is no need for it to be an atomic or a
refcount. An unsigned int would do as well and be faster. My fault.

I will send a patch to go on top of yours to fix that.

Otherwise:

Reviewed-by: Damien Le Moal <damien.lemoal@wdc.com>
Tested-by: Damien Le Moal <damien.lemoal@wdc.com>

Thanks !


>  	unsigned long		state;
>  	struct page		*page;
>  	void			*data;
> @@ -296,7 +296,7 @@ static struct dmz_mblock *dmz_alloc_mblock(struct dmz_metadata *zmd,
>  
>  	RB_CLEAR_NODE(&mblk->node);
>  	INIT_LIST_HEAD(&mblk->link);
> -	atomic_set(&mblk->ref, 0);
> +	refcount_set(&mblk->ref, 0);
>  	mblk->state = 0;
>  	mblk->no = mblk_no;
>  	mblk->data = page_address(mblk->page);
> @@ -397,7 +397,7 @@ static struct dmz_mblock *dmz_fetch_mblock(struct dmz_metadata *zmd,
>  		return NULL;
>  
>  	spin_lock(&zmd->mblk_lock);
> -	atomic_inc(&mblk->ref);
> +	refcount_inc(&mblk->ref);
>  	set_bit(DMZ_META_READING, &mblk->state);
>  	dmz_insert_mblock(zmd, mblk);
>  	spin_unlock(&zmd->mblk_lock);
> @@ -484,7 +484,7 @@ static void dmz_release_mblock(struct dmz_metadata *zmd,
>  
>  	spin_lock(&zmd->mblk_lock);
>  
> -	if (atomic_dec_and_test(&mblk->ref)) {
> +	if (refcount_dec_and_test(&mblk->ref)) {
>  		if (test_bit(DMZ_META_ERROR, &mblk->state)) {
>  			rb_erase(&mblk->node, &zmd->mblk_rbtree);
>  			dmz_free_mblock(zmd, mblk);
> @@ -511,7 +511,8 @@ static struct dmz_mblock *dmz_get_mblock(struct dmz_metadata *zmd,
>  	mblk = dmz_lookup_mblock(zmd, mblk_no);
>  	if (mblk) {
>  		/* Cache hit: remove block from LRU list */
> -		if (atomic_inc_return(&mblk->ref) == 1 &&
> +		refcount_inc(&mblk->ref);
> +		if (refcount_read(&mblk->ref) == 1 &&
>  		    !test_bit(DMZ_META_DIRTY, &mblk->state))
>  			list_del_init(&mblk->link);
>  	}
> @@ -753,7 +754,7 @@ int dmz_flush_metadata(struct dmz_metadata *zmd)
>  
>  		spin_lock(&zmd->mblk_lock);
>  		clear_bit(DMZ_META_DIRTY, &mblk->state);
> -		if (atomic_read(&mblk->ref) == 0)
> +		if (refcount_read(&mblk->ref) == 0)
>  			list_add_tail(&mblk->link, &zmd->mblk_lru_list);
>  		spin_unlock(&zmd->mblk_lock);
>  	}
> @@ -1048,7 +1049,7 @@ static int dmz_init_zone(struct dmz_metadata *zmd, struct dm_zone *zone,
>  	}
>  
>  	INIT_LIST_HEAD(&zone->link);
> -	atomic_set(&zone->refcount, 0);
> +	refcount_set(&zone->refcount, 0);
>  	zone->chunk = DMZ_MAP_UNMAPPED;
>  
>  	if (blkz->type == BLK_ZONE_TYPE_CONVENTIONAL) {
> @@ -1574,7 +1575,7 @@ struct dm_zone *dmz_get_zone_for_reclaim(struct dmz_metadata *zmd)
>  void dmz_activate_zone(struct dm_zone *zone)
>  {
>  	set_bit(DMZ_ACTIVE, &zone->flags);
> -	atomic_inc(&zone->refcount);
> +	refcount_inc(&zone->refcount);
>  }
>  
>  /*
> @@ -1585,7 +1586,7 @@ void dmz_activate_zone(struct dm_zone *zone)
>   */
>  void dmz_deactivate_zone(struct dm_zone *zone)
>  {
> -	if (atomic_dec_and_test(&zone->refcount)) {
> +	if (refcount_dec_and_test(&zone->refcount)) {
>  		WARN_ON(!test_bit(DMZ_ACTIVE, &zone->flags));
>  		clear_bit_unlock(DMZ_ACTIVE, &zone->flags);
>  		smp_mb__after_atomic();
> @@ -2308,7 +2309,7 @@ static void dmz_cleanup_metadata(struct dmz_metadata *zmd)
>  		mblk = list_first_entry(&zmd->mblk_dirty_list,
>  					struct dmz_mblock, link);
>  		dmz_dev_warn(zmd->dev, "mblock %llu still in dirty list (ref %u)",
> -			     (u64)mblk->no, atomic_read(&mblk->ref));
> +			     (u64)mblk->no, refcount_read(&mblk->ref));
>  		list_del_init(&mblk->link);
>  		rb_erase(&mblk->node, &zmd->mblk_rbtree);
>  		dmz_free_mblock(zmd, mblk);
> @@ -2326,8 +2327,8 @@ static void dmz_cleanup_metadata(struct dmz_metadata *zmd)
>  	root = &zmd->mblk_rbtree;
>  	rbtree_postorder_for_each_entry_safe(mblk, next, root, node) {
>  		dmz_dev_warn(zmd->dev, "mblock %llu ref %u still in rbtree",
> -			     (u64)mblk->no, atomic_read(&mblk->ref));
> -		atomic_set(&mblk->ref, 0);
> +			     (u64)mblk->no, refcount_read(&mblk->ref));
> +		refcount_set(&mblk->ref, 0);
>  		dmz_free_mblock(zmd, mblk);
>  	}
>  
> diff --git a/drivers/md/dm-zoned.h b/drivers/md/dm-zoned.h
> index 12419f0bfe78..b7829a615d26 100644
> --- a/drivers/md/dm-zoned.h
> +++ b/drivers/md/dm-zoned.h
> @@ -78,7 +78,7 @@ struct dm_zone {
>  	unsigned long		flags;
>  
>  	/* Zone activation reference count */
> -	atomic_t		refcount;
> +	refcount_t		refcount;
>  
>  	/* Zone write pointer block (relative to the zone start block) */
>  	unsigned int		wp_block;
>
John Pittman Aug. 23, 2018, 10:54 p.m. UTC | #2
Sounds good Damien.  Thanks for reviewing!

On Thu, Aug 23, 2018 at 6:12 PM, Damien Le Moal <Damien.LeMoal@wdc.com> wrote:
> John,
>
> On 2018/08/23 10:37, John Pittman wrote:
>> The API surrounding refcount_t should be used in place of atomic_t
>> when variables are being used as reference counters.  This API can
>> prevent issues such as counter overflows and use-after-free
>> conditions.  Within the dm zoned metadata stack, the atomic_t API
>> is used for mblk->ref and zone->refcount.  Change these to use
>> refcount_t, avoiding the issues mentioned.
>>
>> Signed-off-by: John Pittman <jpittman@redhat.com>
>> ---
>>  drivers/md/dm-zoned-metadata.c | 25 +++++++++++++------------
>>  drivers/md/dm-zoned.h          |  2 +-
>>  2 files changed, 14 insertions(+), 13 deletions(-)
>>
>> diff --git a/drivers/md/dm-zoned-metadata.c b/drivers/md/dm-zoned-metadata.c
>> index 969954915566..92e635749414 100644
>> --- a/drivers/md/dm-zoned-metadata.c
>> +++ b/drivers/md/dm-zoned-metadata.c
>> @@ -99,7 +99,7 @@ struct dmz_mblock {
>>       struct rb_node          node;
>>       struct list_head        link;
>>       sector_t                no;
>> -     atomic_t                ref;
>> +     refcount_t              ref;
>
> While reviewing your patch, I realized that this ref is always manipulated under
> the zmd->mblk_lock spinlock. So there is no need for it to be an atomic or a
> refcount. An unsigned int would do as well and be faster. My fault.
>
> I will send a patch to go on top of yours to fix that.
>
> Otherwise:
>
> Reviewed-by: Damien Le Moal <damien.lemoal@wdc.com>
> Tested-by: Damien Le Moal <damien.lemoal@wdc.com>
>
> Thanks !
>
>
>>       unsigned long           state;
>>       struct page             *page;
>>       void                    *data;
>> @@ -296,7 +296,7 @@ static struct dmz_mblock *dmz_alloc_mblock(struct dmz_metadata *zmd,
>>
>>       RB_CLEAR_NODE(&mblk->node);
>>       INIT_LIST_HEAD(&mblk->link);
>> -     atomic_set(&mblk->ref, 0);
>> +     refcount_set(&mblk->ref, 0);
>>       mblk->state = 0;
>>       mblk->no = mblk_no;
>>       mblk->data = page_address(mblk->page);
>> @@ -397,7 +397,7 @@ static struct dmz_mblock *dmz_fetch_mblock(struct dmz_metadata *zmd,
>>               return NULL;
>>
>>       spin_lock(&zmd->mblk_lock);
>> -     atomic_inc(&mblk->ref);
>> +     refcount_inc(&mblk->ref);
>>       set_bit(DMZ_META_READING, &mblk->state);
>>       dmz_insert_mblock(zmd, mblk);
>>       spin_unlock(&zmd->mblk_lock);
>> @@ -484,7 +484,7 @@ static void dmz_release_mblock(struct dmz_metadata *zmd,
>>
>>       spin_lock(&zmd->mblk_lock);
>>
>> -     if (atomic_dec_and_test(&mblk->ref)) {
>> +     if (refcount_dec_and_test(&mblk->ref)) {
>>               if (test_bit(DMZ_META_ERROR, &mblk->state)) {
>>                       rb_erase(&mblk->node, &zmd->mblk_rbtree);
>>                       dmz_free_mblock(zmd, mblk);
>> @@ -511,7 +511,8 @@ static struct dmz_mblock *dmz_get_mblock(struct dmz_metadata *zmd,
>>       mblk = dmz_lookup_mblock(zmd, mblk_no);
>>       if (mblk) {
>>               /* Cache hit: remove block from LRU list */
>> -             if (atomic_inc_return(&mblk->ref) == 1 &&
>> +             refcount_inc(&mblk->ref);
>> +             if (refcount_read(&mblk->ref) == 1 &&
>>                   !test_bit(DMZ_META_DIRTY, &mblk->state))
>>                       list_del_init(&mblk->link);
>>       }
>> @@ -753,7 +754,7 @@ int dmz_flush_metadata(struct dmz_metadata *zmd)
>>
>>               spin_lock(&zmd->mblk_lock);
>>               clear_bit(DMZ_META_DIRTY, &mblk->state);
>> -             if (atomic_read(&mblk->ref) == 0)
>> +             if (refcount_read(&mblk->ref) == 0)
>>                       list_add_tail(&mblk->link, &zmd->mblk_lru_list);
>>               spin_unlock(&zmd->mblk_lock);
>>       }
>> @@ -1048,7 +1049,7 @@ static int dmz_init_zone(struct dmz_metadata *zmd, struct dm_zone *zone,
>>       }
>>
>>       INIT_LIST_HEAD(&zone->link);
>> -     atomic_set(&zone->refcount, 0);
>> +     refcount_set(&zone->refcount, 0);
>>       zone->chunk = DMZ_MAP_UNMAPPED;
>>
>>       if (blkz->type == BLK_ZONE_TYPE_CONVENTIONAL) {
>> @@ -1574,7 +1575,7 @@ struct dm_zone *dmz_get_zone_for_reclaim(struct dmz_metadata *zmd)
>>  void dmz_activate_zone(struct dm_zone *zone)
>>  {
>>       set_bit(DMZ_ACTIVE, &zone->flags);
>> -     atomic_inc(&zone->refcount);
>> +     refcount_inc(&zone->refcount);
>>  }
>>
>>  /*
>> @@ -1585,7 +1586,7 @@ void dmz_activate_zone(struct dm_zone *zone)
>>   */
>>  void dmz_deactivate_zone(struct dm_zone *zone)
>>  {
>> -     if (atomic_dec_and_test(&zone->refcount)) {
>> +     if (refcount_dec_and_test(&zone->refcount)) {
>>               WARN_ON(!test_bit(DMZ_ACTIVE, &zone->flags));
>>               clear_bit_unlock(DMZ_ACTIVE, &zone->flags);
>>               smp_mb__after_atomic();
>> @@ -2308,7 +2309,7 @@ static void dmz_cleanup_metadata(struct dmz_metadata *zmd)
>>               mblk = list_first_entry(&zmd->mblk_dirty_list,
>>                                       struct dmz_mblock, link);
>>               dmz_dev_warn(zmd->dev, "mblock %llu still in dirty list (ref %u)",
>> -                          (u64)mblk->no, atomic_read(&mblk->ref));
>> +                          (u64)mblk->no, refcount_read(&mblk->ref));
>>               list_del_init(&mblk->link);
>>               rb_erase(&mblk->node, &zmd->mblk_rbtree);
>>               dmz_free_mblock(zmd, mblk);
>> @@ -2326,8 +2327,8 @@ static void dmz_cleanup_metadata(struct dmz_metadata *zmd)
>>       root = &zmd->mblk_rbtree;
>>       rbtree_postorder_for_each_entry_safe(mblk, next, root, node) {
>>               dmz_dev_warn(zmd->dev, "mblock %llu ref %u still in rbtree",
>> -                          (u64)mblk->no, atomic_read(&mblk->ref));
>> -             atomic_set(&mblk->ref, 0);
>> +                          (u64)mblk->no, refcount_read(&mblk->ref));
>> +             refcount_set(&mblk->ref, 0);
>>               dmz_free_mblock(zmd, mblk);
>>       }
>>
>> diff --git a/drivers/md/dm-zoned.h b/drivers/md/dm-zoned.h
>> index 12419f0bfe78..b7829a615d26 100644
>> --- a/drivers/md/dm-zoned.h
>> +++ b/drivers/md/dm-zoned.h
>> @@ -78,7 +78,7 @@ struct dm_zone {
>>       unsigned long           flags;
>>
>>       /* Zone activation reference count */
>> -     atomic_t                refcount;
>> +     refcount_t              refcount;
>>
>>       /* Zone write pointer block (relative to the zone start block) */
>>       unsigned int            wp_block;
>>
>
>
> --
> Damien Le Moal
> Western Digital Research

--
dm-devel mailing list
dm-devel@redhat.com
https://www.redhat.com/mailman/listinfo/dm-devel
Mike Snitzer Oct. 16, 2018, 6:33 p.m. UTC | #3
On Thu, Aug 23 2018 at  6:12pm -0400,
Damien Le Moal <Damien.LeMoal@wdc.com> wrote:

> John,
> 
> On 2018/08/23 10:37, John Pittman wrote:
> > The API surrounding refcount_t should be used in place of atomic_t
> > when variables are being used as reference counters.  This API can
> > prevent issues such as counter overflows and use-after-free
> > conditions.  Within the dm zoned metadata stack, the atomic_t API
> > is used for mblk->ref and zone->refcount.  Change these to use
> > refcount_t, avoiding the issues mentioned.
> > 
> > Signed-off-by: John Pittman <jpittman@redhat.com>
> > ---
> >  drivers/md/dm-zoned-metadata.c | 25 +++++++++++++------------
> >  drivers/md/dm-zoned.h          |  2 +-
> >  2 files changed, 14 insertions(+), 13 deletions(-)
> > 
> > diff --git a/drivers/md/dm-zoned-metadata.c b/drivers/md/dm-zoned-metadata.c
> > index 969954915566..92e635749414 100644
> > --- a/drivers/md/dm-zoned-metadata.c
> > +++ b/drivers/md/dm-zoned-metadata.c
> > @@ -99,7 +99,7 @@ struct dmz_mblock {
> >  	struct rb_node		node;
> >  	struct list_head	link;
> >  	sector_t		no;
> > -	atomic_t		ref;
> > +	refcount_t		ref;
> 
> While reviewing your patch, I realized that this ref is always manipulated under
> the zmd->mblk_lock spinlock. So there is no need for it to be an atomic or a
> refcount. An unsigned int would do as well and be faster. My fault.
> 
> I will send a patch to go on top of yours to fix that.

Hi Damien,

Given what you've said I'm not seeing the point in the intermediate
refcount_t conversion. 

I'd rather you just send a patch that switches atomic_t to int.

Thanks,
Mike

--
dm-devel mailing list
dm-devel@redhat.com
https://www.redhat.com/mailman/listinfo/dm-devel
Damien Le Moal Oct. 17, 2018, 2:59 a.m. UTC | #4
Mike,

On 2018/10/17 3:34, Mike Snitzer wrote:
> On Thu, Aug 23 2018 at  6:12pm -0400,
> Damien Le Moal <Damien.LeMoal@wdc.com> wrote:
> 
>> John,
>>
>> On 2018/08/23 10:37, John Pittman wrote:
>>> The API surrounding refcount_t should be used in place of atomic_t
>>> when variables are being used as reference counters.  This API can
>>> prevent issues such as counter overflows and use-after-free
>>> conditions.  Within the dm zoned metadata stack, the atomic_t API
>>> is used for mblk->ref and zone->refcount.  Change these to use
>>> refcount_t, avoiding the issues mentioned.
>>>
>>> Signed-off-by: John Pittman <jpittman@redhat.com>
>>> ---
>>>  drivers/md/dm-zoned-metadata.c | 25 +++++++++++++------------
>>>  drivers/md/dm-zoned.h          |  2 +-
>>>  2 files changed, 14 insertions(+), 13 deletions(-)
>>>
>>> diff --git a/drivers/md/dm-zoned-metadata.c b/drivers/md/dm-zoned-metadata.c
>>> index 969954915566..92e635749414 100644
>>> --- a/drivers/md/dm-zoned-metadata.c
>>> +++ b/drivers/md/dm-zoned-metadata.c
>>> @@ -99,7 +99,7 @@ struct dmz_mblock {
>>>  	struct rb_node		node;
>>>  	struct list_head	link;
>>>  	sector_t		no;
>>> -	atomic_t		ref;
>>> +	refcount_t		ref;
>>
>> While reviewing your patch, I realized that this ref is always manipulated under
>> the zmd->mblk_lock spinlock. So there is no need for it to be an atomic or a
>> refcount. An unsigned int would do as well and be faster. My fault.
>>
>> I will send a patch to go on top of yours to fix that.
> 
> Hi Damien,
> 
> Given what you've said I'm not seeing the point in the intermediate
> refcount_t conversion. 
> 
> I'd rather you just send a patch that switches atomic_t to int.

OK. Will send that shortly (and thanks for reminding me, I completely forgot
about this !).

Best regards.
diff mbox series

Patch

diff --git a/drivers/md/dm-zoned-metadata.c b/drivers/md/dm-zoned-metadata.c
index 969954915566..92e635749414 100644
--- a/drivers/md/dm-zoned-metadata.c
+++ b/drivers/md/dm-zoned-metadata.c
@@ -99,7 +99,7 @@  struct dmz_mblock {
 	struct rb_node		node;
 	struct list_head	link;
 	sector_t		no;
-	atomic_t		ref;
+	refcount_t		ref;
 	unsigned long		state;
 	struct page		*page;
 	void			*data;
@@ -296,7 +296,7 @@  static struct dmz_mblock *dmz_alloc_mblock(struct dmz_metadata *zmd,
 
 	RB_CLEAR_NODE(&mblk->node);
 	INIT_LIST_HEAD(&mblk->link);
-	atomic_set(&mblk->ref, 0);
+	refcount_set(&mblk->ref, 0);
 	mblk->state = 0;
 	mblk->no = mblk_no;
 	mblk->data = page_address(mblk->page);
@@ -397,7 +397,7 @@  static struct dmz_mblock *dmz_fetch_mblock(struct dmz_metadata *zmd,
 		return NULL;
 
 	spin_lock(&zmd->mblk_lock);
-	atomic_inc(&mblk->ref);
+	refcount_inc(&mblk->ref);
 	set_bit(DMZ_META_READING, &mblk->state);
 	dmz_insert_mblock(zmd, mblk);
 	spin_unlock(&zmd->mblk_lock);
@@ -484,7 +484,7 @@  static void dmz_release_mblock(struct dmz_metadata *zmd,
 
 	spin_lock(&zmd->mblk_lock);
 
-	if (atomic_dec_and_test(&mblk->ref)) {
+	if (refcount_dec_and_test(&mblk->ref)) {
 		if (test_bit(DMZ_META_ERROR, &mblk->state)) {
 			rb_erase(&mblk->node, &zmd->mblk_rbtree);
 			dmz_free_mblock(zmd, mblk);
@@ -511,7 +511,8 @@  static struct dmz_mblock *dmz_get_mblock(struct dmz_metadata *zmd,
 	mblk = dmz_lookup_mblock(zmd, mblk_no);
 	if (mblk) {
 		/* Cache hit: remove block from LRU list */
-		if (atomic_inc_return(&mblk->ref) == 1 &&
+		refcount_inc(&mblk->ref);
+		if (refcount_read(&mblk->ref) == 1 &&
 		    !test_bit(DMZ_META_DIRTY, &mblk->state))
 			list_del_init(&mblk->link);
 	}
@@ -753,7 +754,7 @@  int dmz_flush_metadata(struct dmz_metadata *zmd)
 
 		spin_lock(&zmd->mblk_lock);
 		clear_bit(DMZ_META_DIRTY, &mblk->state);
-		if (atomic_read(&mblk->ref) == 0)
+		if (refcount_read(&mblk->ref) == 0)
 			list_add_tail(&mblk->link, &zmd->mblk_lru_list);
 		spin_unlock(&zmd->mblk_lock);
 	}
@@ -1048,7 +1049,7 @@  static int dmz_init_zone(struct dmz_metadata *zmd, struct dm_zone *zone,
 	}
 
 	INIT_LIST_HEAD(&zone->link);
-	atomic_set(&zone->refcount, 0);
+	refcount_set(&zone->refcount, 0);
 	zone->chunk = DMZ_MAP_UNMAPPED;
 
 	if (blkz->type == BLK_ZONE_TYPE_CONVENTIONAL) {
@@ -1574,7 +1575,7 @@  struct dm_zone *dmz_get_zone_for_reclaim(struct dmz_metadata *zmd)
 void dmz_activate_zone(struct dm_zone *zone)
 {
 	set_bit(DMZ_ACTIVE, &zone->flags);
-	atomic_inc(&zone->refcount);
+	refcount_inc(&zone->refcount);
 }
 
 /*
@@ -1585,7 +1586,7 @@  void dmz_activate_zone(struct dm_zone *zone)
  */
 void dmz_deactivate_zone(struct dm_zone *zone)
 {
-	if (atomic_dec_and_test(&zone->refcount)) {
+	if (refcount_dec_and_test(&zone->refcount)) {
 		WARN_ON(!test_bit(DMZ_ACTIVE, &zone->flags));
 		clear_bit_unlock(DMZ_ACTIVE, &zone->flags);
 		smp_mb__after_atomic();
@@ -2308,7 +2309,7 @@  static void dmz_cleanup_metadata(struct dmz_metadata *zmd)
 		mblk = list_first_entry(&zmd->mblk_dirty_list,
 					struct dmz_mblock, link);
 		dmz_dev_warn(zmd->dev, "mblock %llu still in dirty list (ref %u)",
-			     (u64)mblk->no, atomic_read(&mblk->ref));
+			     (u64)mblk->no, refcount_read(&mblk->ref));
 		list_del_init(&mblk->link);
 		rb_erase(&mblk->node, &zmd->mblk_rbtree);
 		dmz_free_mblock(zmd, mblk);
@@ -2326,8 +2327,8 @@  static void dmz_cleanup_metadata(struct dmz_metadata *zmd)
 	root = &zmd->mblk_rbtree;
 	rbtree_postorder_for_each_entry_safe(mblk, next, root, node) {
 		dmz_dev_warn(zmd->dev, "mblock %llu ref %u still in rbtree",
-			     (u64)mblk->no, atomic_read(&mblk->ref));
-		atomic_set(&mblk->ref, 0);
+			     (u64)mblk->no, refcount_read(&mblk->ref));
+		refcount_set(&mblk->ref, 0);
 		dmz_free_mblock(zmd, mblk);
 	}
 
diff --git a/drivers/md/dm-zoned.h b/drivers/md/dm-zoned.h
index 12419f0bfe78..b7829a615d26 100644
--- a/drivers/md/dm-zoned.h
+++ b/drivers/md/dm-zoned.h
@@ -78,7 +78,7 @@  struct dm_zone {
 	unsigned long		flags;
 
 	/* Zone activation reference count */
-	atomic_t		refcount;
+	refcount_t		refcount;
 
 	/* Zone write pointer block (relative to the zone start block) */
 	unsigned int		wp_block;