diff mbox series

[rdma-next,02/12] RDMA/odp: Iterate over the whole rbtree directly

Message ID 20190819111710.18440-3-leon@kernel.org (mailing list archive)
State Accepted
Delegated to: Jason Gunthorpe
Headers show
Series Improvements for ODP | expand

Commit Message

Leon Romanovsky Aug. 19, 2019, 11:17 a.m. UTC
From: Jason Gunthorpe <jgg@mellanox.com>

Instead of intersecting a full interval, just iterate over every element
directly. This is faster and clearer.

Signed-off-by: Jason Gunthorpe <jgg@mellanox.com>
Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
---
 drivers/infiniband/core/umem_odp.c | 51 ++++++++++++++++--------------
 drivers/infiniband/hw/mlx5/odp.c   | 41 +++++++++++-------------
 2 files changed, 47 insertions(+), 45 deletions(-)

Comments

Jason Gunthorpe Aug. 21, 2019, 5:15 p.m. UTC | #1
On Mon, Aug 19, 2019 at 02:17:00PM +0300, Leon Romanovsky wrote:
> From: Jason Gunthorpe <jgg@mellanox.com>
> 
> Instead of intersecting a full interval, just iterate over every element
> directly. This is faster and clearer.
> 
> Signed-off-by: Jason Gunthorpe <jgg@mellanox.com>
> Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
>  drivers/infiniband/core/umem_odp.c | 51 ++++++++++++++++--------------
>  drivers/infiniband/hw/mlx5/odp.c   | 41 +++++++++++-------------
>  2 files changed, 47 insertions(+), 45 deletions(-)
> 
> diff --git a/drivers/infiniband/core/umem_odp.c b/drivers/infiniband/core/umem_odp.c
> index 8358eb8e3a26..b9bebef00a33 100644
> +++ b/drivers/infiniband/core/umem_odp.c
> @@ -72,35 +72,41 @@ static void ib_umem_notifier_end_account(struct ib_umem_odp *umem_odp)
>  	mutex_unlock(&umem_odp->umem_mutex);
>  }
>  
> -static int ib_umem_notifier_release_trampoline(struct ib_umem_odp *umem_odp,
> -					       u64 start, u64 end, void *cookie)
> -{
> -	/*
> -	 * Increase the number of notifiers running, to
> -	 * prevent any further fault handling on this MR.
> -	 */
> -	ib_umem_notifier_start_account(umem_odp);
> -	umem_odp->dying = 1;

This patch was not applied on top of the commit noted in the cover
letter

> -	/* Make sure that the fact the umem is dying is out before we release
> -	 * all pending page faults. */
> -	smp_wmb();
> -	complete_all(&umem_odp->notifier_completion);
> -	umem_odp->umem.context->invalidate_range(
> -		umem_odp, ib_umem_start(umem_odp), ib_umem_end(umem_odp));
> -	return 0;
> -}
> -
>  static void ib_umem_notifier_release(struct mmu_notifier *mn,
>  				     struct mm_struct *mm)
>  {
>  	struct ib_ucontext_per_mm *per_mm =
>  		container_of(mn, struct ib_ucontext_per_mm, mn);
> +	struct rb_node *node;
>  
>  	down_read(&per_mm->umem_rwsem);
> -	if (per_mm->active)
> -		rbt_ib_umem_for_each_in_range(
> -			&per_mm->umem_tree, 0, ULLONG_MAX,
> -			ib_umem_notifier_release_trampoline, true, NULL);
> +	if (!per_mm->active)
> +		goto out;
> +
> +	for (node = rb_first_cached(&per_mm->umem_tree); node;
> +	     node = rb_next(node)) {
> +		struct ib_umem_odp *umem_odp =
> +			rb_entry(node, struct ib_umem_odp, interval_tree.rb);
> +
> +		/*
> +		 * Increase the number of notifiers running, to prevent any
> +		 * further fault handling on this MR.
> +		 */
> +		ib_umem_notifier_start_account(umem_odp);
> +
> +		umem_odp->dying = 1;

So this ends up as a 'rebasing error'

I fixed it

Jason
Leon Romanovsky Aug. 21, 2019, 5:27 p.m. UTC | #2
On Wed, Aug 21, 2019 at 02:15:02PM -0300, Jason Gunthorpe wrote:
> On Mon, Aug 19, 2019 at 02:17:00PM +0300, Leon Romanovsky wrote:
> > From: Jason Gunthorpe <jgg@mellanox.com>
> >
> > Instead of intersecting a full interval, just iterate over every element
> > directly. This is faster and clearer.
> >
> > Signed-off-by: Jason Gunthorpe <jgg@mellanox.com>
> > Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
> >  drivers/infiniband/core/umem_odp.c | 51 ++++++++++++++++--------------
> >  drivers/infiniband/hw/mlx5/odp.c   | 41 +++++++++++-------------
> >  2 files changed, 47 insertions(+), 45 deletions(-)
> >
> > diff --git a/drivers/infiniband/core/umem_odp.c b/drivers/infiniband/core/umem_odp.c
> > index 8358eb8e3a26..b9bebef00a33 100644
> > +++ b/drivers/infiniband/core/umem_odp.c
> > @@ -72,35 +72,41 @@ static void ib_umem_notifier_end_account(struct ib_umem_odp *umem_odp)
> >  	mutex_unlock(&umem_odp->umem_mutex);
> >  }
> >
> > -static int ib_umem_notifier_release_trampoline(struct ib_umem_odp *umem_odp,
> > -					       u64 start, u64 end, void *cookie)
> > -{
> > -	/*
> > -	 * Increase the number of notifiers running, to
> > -	 * prevent any further fault handling on this MR.
> > -	 */
> > -	ib_umem_notifier_start_account(umem_odp);
> > -	umem_odp->dying = 1;
>
> This patch was not applied on top of the commit noted in the cover
> letter

Strange: git log --oneline on my submission queue.
....
39c10977a728 RDMA/odp: Iterate over the whole rbtree directly
779c1205d0e0 RDMA/odp: Use the common interval tree library instead of generic
25705cc22617 RDMA/mlx5: Fix MR npages calculation for IB_ACCESS_HUGETLB
---


>
> > -	/* Make sure that the fact the umem is dying is out before we release
> > -	 * all pending page faults. */
> > -	smp_wmb();
> > -	complete_all(&umem_odp->notifier_completion);
> > -	umem_odp->umem.context->invalidate_range(
> > -		umem_odp, ib_umem_start(umem_odp), ib_umem_end(umem_odp));
> > -	return 0;
> > -}
> > -
> >  static void ib_umem_notifier_release(struct mmu_notifier *mn,
> >  				     struct mm_struct *mm)
> >  {
> >  	struct ib_ucontext_per_mm *per_mm =
> >  		container_of(mn, struct ib_ucontext_per_mm, mn);
> > +	struct rb_node *node;
> >
> >  	down_read(&per_mm->umem_rwsem);
> > -	if (per_mm->active)
> > -		rbt_ib_umem_for_each_in_range(
> > -			&per_mm->umem_tree, 0, ULLONG_MAX,
> > -			ib_umem_notifier_release_trampoline, true, NULL);
> > +	if (!per_mm->active)
> > +		goto out;
> > +
> > +	for (node = rb_first_cached(&per_mm->umem_tree); node;
> > +	     node = rb_next(node)) {
> > +		struct ib_umem_odp *umem_odp =
> > +			rb_entry(node, struct ib_umem_odp, interval_tree.rb);
> > +
> > +		/*
> > +		 * Increase the number of notifiers running, to prevent any
> > +		 * further fault handling on this MR.
> > +		 */
> > +		ib_umem_notifier_start_account(umem_odp);
> > +
> > +		umem_odp->dying = 1;
>
> So this ends up as a 'rebasing error'
>
> I fixed it
>
> Jason
Jason Gunthorpe Aug. 21, 2019, 5:35 p.m. UTC | #3
On Wed, Aug 21, 2019 at 08:27:35PM +0300, Leon Romanovsky wrote:
> On Wed, Aug 21, 2019 at 02:15:02PM -0300, Jason Gunthorpe wrote:
> > On Mon, Aug 19, 2019 at 02:17:00PM +0300, Leon Romanovsky wrote:
> > > From: Jason Gunthorpe <jgg@mellanox.com>
> > >
> > > Instead of intersecting a full interval, just iterate over every element
> > > directly. This is faster and clearer.
> > >
> > > Signed-off-by: Jason Gunthorpe <jgg@mellanox.com>
> > > Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
> > >  drivers/infiniband/core/umem_odp.c | 51 ++++++++++++++++--------------
> > >  drivers/infiniband/hw/mlx5/odp.c   | 41 +++++++++++-------------
> > >  2 files changed, 47 insertions(+), 45 deletions(-)
> > >
> > > diff --git a/drivers/infiniband/core/umem_odp.c b/drivers/infiniband/core/umem_odp.c
> > > index 8358eb8e3a26..b9bebef00a33 100644
> > > +++ b/drivers/infiniband/core/umem_odp.c
> > > @@ -72,35 +72,41 @@ static void ib_umem_notifier_end_account(struct ib_umem_odp *umem_odp)
> > >  	mutex_unlock(&umem_odp->umem_mutex);
> > >  }
> > >
> > > -static int ib_umem_notifier_release_trampoline(struct ib_umem_odp *umem_odp,
> > > -					       u64 start, u64 end, void *cookie)
> > > -{
> > > -	/*
> > > -	 * Increase the number of notifiers running, to
> > > -	 * prevent any further fault handling on this MR.
> > > -	 */
> > > -	ib_umem_notifier_start_account(umem_odp);
> > > -	umem_odp->dying = 1;
> >
> > This patch was not applied on top of the commit noted in the cover
> > letter
> 
> Strange: git log --oneline on my submission queue.
> ....
> 39c10977a728 RDMA/odp: Iterate over the whole rbtree directly
> 779c1205d0e0 RDMA/odp: Use the common interval tree library instead of generic
> 25705cc22617 RDMA/mlx5: Fix MR npages calculation for IB_ACCESS_HUGETLB

But that patch has to apply on top of rc, which has the other commit
that deleted dying

Jason
Leon Romanovsky Aug. 21, 2019, 5:47 p.m. UTC | #4
On Wed, Aug 21, 2019 at 02:35:52PM -0300, Jason Gunthorpe wrote:
> On Wed, Aug 21, 2019 at 08:27:35PM +0300, Leon Romanovsky wrote:
> > On Wed, Aug 21, 2019 at 02:15:02PM -0300, Jason Gunthorpe wrote:
> > > On Mon, Aug 19, 2019 at 02:17:00PM +0300, Leon Romanovsky wrote:
> > > > From: Jason Gunthorpe <jgg@mellanox.com>
> > > >
> > > > Instead of intersecting a full interval, just iterate over every element
> > > > directly. This is faster and clearer.
> > > >
> > > > Signed-off-by: Jason Gunthorpe <jgg@mellanox.com>
> > > > Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
> > > >  drivers/infiniband/core/umem_odp.c | 51 ++++++++++++++++--------------
> > > >  drivers/infiniband/hw/mlx5/odp.c   | 41 +++++++++++-------------
> > > >  2 files changed, 47 insertions(+), 45 deletions(-)
> > > >
> > > > diff --git a/drivers/infiniband/core/umem_odp.c b/drivers/infiniband/core/umem_odp.c
> > > > index 8358eb8e3a26..b9bebef00a33 100644
> > > > +++ b/drivers/infiniband/core/umem_odp.c
> > > > @@ -72,35 +72,41 @@ static void ib_umem_notifier_end_account(struct ib_umem_odp *umem_odp)
> > > >  	mutex_unlock(&umem_odp->umem_mutex);
> > > >  }
> > > >
> > > > -static int ib_umem_notifier_release_trampoline(struct ib_umem_odp *umem_odp,
> > > > -					       u64 start, u64 end, void *cookie)
> > > > -{
> > > > -	/*
> > > > -	 * Increase the number of notifiers running, to
> > > > -	 * prevent any further fault handling on this MR.
> > > > -	 */
> > > > -	ib_umem_notifier_start_account(umem_odp);
> > > > -	umem_odp->dying = 1;
> > >
> > > This patch was not applied on top of the commit noted in the cover
> > > letter
> >
> > Strange: git log --oneline on my submission queue.
> > ....
> > 39c10977a728 RDMA/odp: Iterate over the whole rbtree directly
> > 779c1205d0e0 RDMA/odp: Use the common interval tree library instead of generic
> > 25705cc22617 RDMA/mlx5: Fix MR npages calculation for IB_ACCESS_HUGETLB
>
> But that patch has to apply on top of rc, which has the other commit
> that deleted dying

Interesting

>
> Jason
diff mbox series

Patch

diff --git a/drivers/infiniband/core/umem_odp.c b/drivers/infiniband/core/umem_odp.c
index 8358eb8e3a26..b9bebef00a33 100644
--- a/drivers/infiniband/core/umem_odp.c
+++ b/drivers/infiniband/core/umem_odp.c
@@ -72,35 +72,41 @@  static void ib_umem_notifier_end_account(struct ib_umem_odp *umem_odp)
 	mutex_unlock(&umem_odp->umem_mutex);
 }
 
-static int ib_umem_notifier_release_trampoline(struct ib_umem_odp *umem_odp,
-					       u64 start, u64 end, void *cookie)
-{
-	/*
-	 * Increase the number of notifiers running, to
-	 * prevent any further fault handling on this MR.
-	 */
-	ib_umem_notifier_start_account(umem_odp);
-	umem_odp->dying = 1;
-	/* Make sure that the fact the umem is dying is out before we release
-	 * all pending page faults. */
-	smp_wmb();
-	complete_all(&umem_odp->notifier_completion);
-	umem_odp->umem.context->invalidate_range(
-		umem_odp, ib_umem_start(umem_odp), ib_umem_end(umem_odp));
-	return 0;
-}
-
 static void ib_umem_notifier_release(struct mmu_notifier *mn,
 				     struct mm_struct *mm)
 {
 	struct ib_ucontext_per_mm *per_mm =
 		container_of(mn, struct ib_ucontext_per_mm, mn);
+	struct rb_node *node;
 
 	down_read(&per_mm->umem_rwsem);
-	if (per_mm->active)
-		rbt_ib_umem_for_each_in_range(
-			&per_mm->umem_tree, 0, ULLONG_MAX,
-			ib_umem_notifier_release_trampoline, true, NULL);
+	if (!per_mm->active)
+		goto out;
+
+	for (node = rb_first_cached(&per_mm->umem_tree); node;
+	     node = rb_next(node)) {
+		struct ib_umem_odp *umem_odp =
+			rb_entry(node, struct ib_umem_odp, interval_tree.rb);
+
+		/*
+		 * Increase the number of notifiers running, to prevent any
+		 * further fault handling on this MR.
+		 */
+		ib_umem_notifier_start_account(umem_odp);
+
+		umem_odp->dying = 1;
+		/*
+		 * Make sure that the fact the umem is dying is out before we
+		 * release all pending page faults.
+		 */
+		smp_wmb();
+		complete_all(&umem_odp->notifier_completion);
+		umem_odp->umem.context->invalidate_range(
+			umem_odp, ib_umem_start(umem_odp),
+			ib_umem_end(umem_odp));
+	}
+
+out:
 	up_read(&per_mm->umem_rwsem);
 }
 
@@ -760,4 +766,3 @@  int rbt_ib_umem_for_each_in_range(struct rb_root_cached *root,
 
 	return ret_val;
 }
-EXPORT_SYMBOL(rbt_ib_umem_for_each_in_range);
diff --git a/drivers/infiniband/hw/mlx5/odp.c b/drivers/infiniband/hw/mlx5/odp.c
index b0c5de39d186..3922fced41ec 100644
--- a/drivers/infiniband/hw/mlx5/odp.c
+++ b/drivers/infiniband/hw/mlx5/odp.c
@@ -539,34 +539,31 @@  struct mlx5_ib_mr *mlx5_ib_alloc_implicit_mr(struct mlx5_ib_pd *pd,
 	return imr;
 }
 
-static int mr_leaf_free(struct ib_umem_odp *umem_odp, u64 start, u64 end,
-			void *cookie)
+void mlx5_ib_free_implicit_mr(struct mlx5_ib_mr *imr)
 {
-	struct mlx5_ib_mr *mr = umem_odp->private, *imr = cookie;
-
-	if (mr->parent != imr)
-		return 0;
-
-	ib_umem_odp_unmap_dma_pages(umem_odp, ib_umem_start(umem_odp),
-				    ib_umem_end(umem_odp));
+	struct ib_ucontext_per_mm *per_mm = mr_to_per_mm(imr);
+	struct rb_node *node;
 
-	if (umem_odp->dying)
-		return 0;
+	down_read(&per_mm->umem_rwsem);
+	for (node = rb_first_cached(&per_mm->umem_tree); node;
+	     node = rb_next(node)) {
+		struct ib_umem_odp *umem_odp =
+			rb_entry(node, struct ib_umem_odp, interval_tree.rb);
+		struct mlx5_ib_mr *mr = umem_odp->private;
 
-	WRITE_ONCE(umem_odp->dying, 1);
-	atomic_inc(&imr->num_leaf_free);
-	schedule_work(&umem_odp->work);
+		if (mr->parent != imr)
+			continue;
 
-	return 0;
-}
+		ib_umem_odp_unmap_dma_pages(umem_odp, ib_umem_start(umem_odp),
+					    ib_umem_end(umem_odp));
 
-void mlx5_ib_free_implicit_mr(struct mlx5_ib_mr *imr)
-{
-	struct ib_ucontext_per_mm *per_mm = mr_to_per_mm(imr);
+		if (umem_odp->dying)
+			continue;
 
-	down_read(&per_mm->umem_rwsem);
-	rbt_ib_umem_for_each_in_range(&per_mm->umem_tree, 0, ULLONG_MAX,
-				      mr_leaf_free, true, imr);
+		WRITE_ONCE(umem_odp->dying, 1);
+		atomic_inc(&imr->num_leaf_free);
+		schedule_work(&umem_odp->work);
+	}
 	up_read(&per_mm->umem_rwsem);
 
 	wait_event(imr->q_leaf_free, !atomic_read(&imr->num_leaf_free));