diff mbox series

mm/mremap_pages: Save a few cycles in 'get_dev_pagemap()'

Message ID b4a47154877853cc64be3a35dcfd594d40cc2bce.1635975283.git.christophe.jaillet@wanadoo.fr (mailing list archive)
State New
Headers show
Series mm/mremap_pages: Save a few cycles in 'get_dev_pagemap()' | expand

Commit Message

Christophe JAILLET Nov. 3, 2021, 9:35 p.m. UTC
Use 'percpu_ref_tryget_live_rcu()' instead of 'percpu_ref_tryget_live()' to
save a few cycles when it is known that the rcu lock is already
taken/released.

Signed-off-by: Christophe JAILLET <christophe.jaillet@wanadoo.fr>
---
 mm/memremap.c | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

Comments

Matthew Wilcox Nov. 3, 2021, 9:41 p.m. UTC | #1
On Wed, Nov 03, 2021 at 10:35:34PM +0100, Christophe JAILLET wrote:
> Use 'percpu_ref_tryget_live_rcu()' instead of 'percpu_ref_tryget_live()' to
> save a few cycles when it is known that the rcu lock is already
> taken/released.

If this is really important, we can add an __xa_load() which doesn't
take the RCU read lock.

I honestly think that the xarray is the wrong data structure here,
and we'd be better off with a simple array of (start, pointer)
tuples.

> Signed-off-by: Christophe JAILLET <christophe.jaillet@wanadoo.fr>
> ---
>  mm/memremap.c | 2 +-
>  1 file changed, 1 insertion(+), 1 deletion(-)
> 
> diff --git a/mm/memremap.c b/mm/memremap.c
> index 84de22c14567..012e8d23d365 100644
> --- a/mm/memremap.c
> +++ b/mm/memremap.c
> @@ -506,7 +506,7 @@ struct dev_pagemap *get_dev_pagemap(unsigned long pfn,
>  	/* fall back to slow path lookup */
>  	rcu_read_lock();
>  	pgmap = xa_load(&pgmap_array, PHYS_PFN(phys));
> -	if (pgmap && !percpu_ref_tryget_live(pgmap->ref))
> +	if (pgmap && !percpu_ref_tryget_live_rcu(pgmap->ref))
>  		pgmap = NULL;
>  	rcu_read_unlock();
>  
> -- 
> 2.30.2
> 
>
Christophe JAILLET Nov. 3, 2021, 9:54 p.m. UTC | #2
Le 03/11/2021 à 22:41, Matthew Wilcox a écrit :
> On Wed, Nov 03, 2021 at 10:35:34PM +0100, Christophe JAILLET wrote:
>> Use 'percpu_ref_tryget_live_rcu()' instead of 'percpu_ref_tryget_live()' to
>> save a few cycles when it is known that the rcu lock is already
>> taken/released.
> 
> If this is really important, we can add an __xa_load() which doesn't
> take the RCU read lock.

There are a few:
    rcu_read_lock();
    mem = xa_load(...);
    rcu_read_unlock();
patterns here and there.

I don't have any numbers of if saving some rcu_read_lock/rcu_read_unlock 
would be useful in these cases.

The only numbers I have are in [1].

[1]: 
https://lore.kernel.org/linux-kernel/cover.1634822969.git.asml.silence@gmail.com/

CJ


> 
> I honestly think that the xarray is the wrong data structure here,
> and we'd be better off with a simple array of (start, pointer)
> tuples.
> 
>> Signed-off-by: Christophe JAILLET <christophe.jaillet@wanadoo.fr>
>> ---
>>   mm/memremap.c | 2 +-
>>   1 file changed, 1 insertion(+), 1 deletion(-)
>>
>> diff --git a/mm/memremap.c b/mm/memremap.c
>> index 84de22c14567..012e8d23d365 100644
>> --- a/mm/memremap.c
>> +++ b/mm/memremap.c
>> @@ -506,7 +506,7 @@ struct dev_pagemap *get_dev_pagemap(unsigned long pfn,
>>   	/* fall back to slow path lookup */
>>   	rcu_read_lock();
>>   	pgmap = xa_load(&pgmap_array, PHYS_PFN(phys));
>> -	if (pgmap && !percpu_ref_tryget_live(pgmap->ref))
>> +	if (pgmap && !percpu_ref_tryget_live_rcu(pgmap->ref))
>>   		pgmap = NULL;
>>   	rcu_read_unlock();
>>   
>> -- 
>> 2.30.2
>>
>>
>
Matthew Wilcox Nov. 3, 2021, 10:01 p.m. UTC | #3
On Wed, Nov 03, 2021 at 10:54:09PM +0100, Christophe JAILLET wrote:
> Le 03/11/2021 à 22:41, Matthew Wilcox a écrit :
> > On Wed, Nov 03, 2021 at 10:35:34PM +0100, Christophe JAILLET wrote:
> > > Use 'percpu_ref_tryget_live_rcu()' instead of 'percpu_ref_tryget_live()' to
> > > save a few cycles when it is known that the rcu lock is already
> > > taken/released.
> > 
> > If this is really important, we can add an __xa_load() which doesn't
> > take the RCU read lock.
> 
> There are a few:
>    rcu_read_lock();
>    mem = xa_load(...);
>    rcu_read_unlock();
> patterns here and there.

If that's all they are, then the rcu_read_lock() and unlock can be
deleted.

if they're actually

	rcu_read_lock()
	mem = xa_load(...);
	try_get_ref(mem);
	rcu_read_unlock();

then of course they can't be.

> I don't have any numbers of if saving some rcu_read_lock/rcu_read_unlock
> would be useful in these cases.
> 
> The only numbers I have are in [1].
> 
> [1]: https://lore.kernel.org/linux-kernel/cover.1634822969.git.asml.silence@gmail.com/

It may not be worth hyperoptimising the slow path like this patch ...
Christophe JAILLET Aug. 12, 2022, 4:58 a.m. UTC | #4
Le 03/11/2021 à 22:35, Christophe JAILLET a écrit :
> Use 'percpu_ref_tryget_live_rcu()' instead of 'percpu_ref_tryget_live()' to
> save a few cycles when it is known that the rcu lock is already
> taken/released.
> 
> Signed-off-by: Christophe JAILLET <christophe.jaillet@wanadoo.fr>
> ---
>   mm/memremap.c | 2 +-
>   1 file changed, 1 insertion(+), 1 deletion(-)
> 
> diff --git a/mm/memremap.c b/mm/memremap.c
> index 84de22c14567..012e8d23d365 100644
> --- a/mm/memremap.c
> +++ b/mm/memremap.c
> @@ -506,7 +506,7 @@ struct dev_pagemap *get_dev_pagemap(unsigned long pfn,
>   	/* fall back to slow path lookup */
>   	rcu_read_lock();
>   	pgmap = xa_load(&pgmap_array, PHYS_PFN(phys));
> -	if (pgmap && !percpu_ref_tryget_live(pgmap->ref))
> +	if (pgmap && !percpu_ref_tryget_live_rcu(pgmap->ref))
>   		pgmap = NULL;
>   	rcu_read_unlock();
>   

Hi,
gentle reminder.

Is this patch useful?
When I first posted it, percpu_ref_tryget_live_rcu() was really new.
Now it is part of linux since 5.16.

CJ
diff mbox series

Patch

diff --git a/mm/memremap.c b/mm/memremap.c
index 84de22c14567..012e8d23d365 100644
--- a/mm/memremap.c
+++ b/mm/memremap.c
@@ -506,7 +506,7 @@  struct dev_pagemap *get_dev_pagemap(unsigned long pfn,
 	/* fall back to slow path lookup */
 	rcu_read_lock();
 	pgmap = xa_load(&pgmap_array, PHYS_PFN(phys));
-	if (pgmap && !percpu_ref_tryget_live(pgmap->ref))
+	if (pgmap && !percpu_ref_tryget_live_rcu(pgmap->ref))
 		pgmap = NULL;
 	rcu_read_unlock();