diff mbox series

[v4,1/2] asm-generic: Add memory barrier dma_mb()

Message ID 20220523113126.171714-2-wangkefeng.wang@huawei.com (mailing list archive)
State New, archived
Headers show
Series arm64: Fix kcsan test_barrier fail and panic | expand

Commit Message

Kefeng Wang May 23, 2022, 11:31 a.m. UTC
The memory barrier dma_mb() is introduced by commit a76a37777f2c
("iommu/arm-smmu-v3: Ensure queue is read after updating prod pointer"),
which is used to ensure that prior (both reads and writes) accesses
to memory by a CPU are ordered w.r.t. a subsequent MMIO write.

Reviewed-by: Arnd Bergmann <arnd@arndb.de> # for asm-generic
Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
---
 Documentation/memory-barriers.txt | 11 ++++++-----
 include/asm-generic/barrier.h     |  8 ++++++++
 2 files changed, 14 insertions(+), 5 deletions(-)

Comments

Marco Elver May 23, 2022, 11:35 a.m. UTC | #1
On Mon, 23 May 2022 at 13:21, Kefeng Wang <wangkefeng.wang@huawei.com> wrote:
>
> The memory barrier dma_mb() is introduced by commit a76a37777f2c
> ("iommu/arm-smmu-v3: Ensure queue is read after updating prod pointer"),
> which is used to ensure that prior (both reads and writes) accesses
> to memory by a CPU are ordered w.r.t. a subsequent MMIO write.
>
> Reviewed-by: Arnd Bergmann <arnd@arndb.de> # for asm-generic
> Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>

Reviewed-by: Marco Elver <elver@google.com>

> ---
>  Documentation/memory-barriers.txt | 11 ++++++-----
>  include/asm-generic/barrier.h     |  8 ++++++++
>  2 files changed, 14 insertions(+), 5 deletions(-)
>
> diff --git a/Documentation/memory-barriers.txt b/Documentation/memory-barriers.txt
> index b12df9137e1c..832b5d36e279 100644
> --- a/Documentation/memory-barriers.txt
> +++ b/Documentation/memory-barriers.txt
> @@ -1894,6 +1894,7 @@ There are some more advanced barrier functions:
>
>   (*) dma_wmb();
>   (*) dma_rmb();
> + (*) dma_mb();
>
>       These are for use with consistent memory to guarantee the ordering
>       of writes or reads of shared memory accessible to both the CPU and a
> @@ -1925,11 +1926,11 @@ There are some more advanced barrier functions:
>       The dma_rmb() allows us guarantee the device has released ownership
>       before we read the data from the descriptor, and the dma_wmb() allows
>       us to guarantee the data is written to the descriptor before the device
> -     can see it now has ownership.  Note that, when using writel(), a prior
> -     wmb() is not needed to guarantee that the cache coherent memory writes
> -     have completed before writing to the MMIO region.  The cheaper
> -     writel_relaxed() does not provide this guarantee and must not be used
> -     here.
> +     can see it now has ownership.  The dma_mb() implies both a dma_rmb() and
> +     a dma_wmb().  Note that, when using writel(), a prior wmb() is not needed
> +     to guarantee that the cache coherent memory writes have completed before
> +     writing to the MMIO region.  The cheaper writel_relaxed() does not provide
> +     this guarantee and must not be used here.
>
>       See the subsection "Kernel I/O barrier effects" for more information on
>       relaxed I/O accessors and the Documentation/core-api/dma-api.rst file for
> diff --git a/include/asm-generic/barrier.h b/include/asm-generic/barrier.h
> index fd7e8fbaeef1..961f4d88f9ef 100644
> --- a/include/asm-generic/barrier.h
> +++ b/include/asm-generic/barrier.h
> @@ -38,6 +38,10 @@
>  #define wmb()  do { kcsan_wmb(); __wmb(); } while (0)
>  #endif
>
> +#ifdef __dma_mb
> +#define dma_mb()       do { kcsan_mb(); __dma_mb(); } while (0)
> +#endif
> +
>  #ifdef __dma_rmb
>  #define dma_rmb()      do { kcsan_rmb(); __dma_rmb(); } while (0)
>  #endif
> @@ -65,6 +69,10 @@
>  #define wmb()  mb()
>  #endif
>
> +#ifndef dma_mb
> +#define dma_mb()       mb()
> +#endif
> +
>  #ifndef dma_rmb
>  #define dma_rmb()      rmb()
>  #endif
> --
> 2.35.3
>
Mark Rutland May 23, 2022, 11:38 a.m. UTC | #2
On Mon, May 23, 2022 at 07:31:25PM +0800, Kefeng Wang wrote:
> The memory barrier dma_mb() is introduced by commit a76a37777f2c
> ("iommu/arm-smmu-v3: Ensure queue is read after updating prod pointer"),
> which is used to ensure that prior (both reads and writes) accesses
> to memory by a CPU are ordered w.r.t. a subsequent MMIO write.
> 
> Reviewed-by: Arnd Bergmann <arnd@arndb.de> # for asm-generic
> Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>

FWIW, this looks sane to me so:

  Acked-by: Mark Rutland <mark.rutland@arm.com>

I'll leave the final say to Will, as I assume this'll go via the arm64 tree and
he'll be the one picking this.

Mark.

> ---
>  Documentation/memory-barriers.txt | 11 ++++++-----
>  include/asm-generic/barrier.h     |  8 ++++++++
>  2 files changed, 14 insertions(+), 5 deletions(-)
> 
> diff --git a/Documentation/memory-barriers.txt b/Documentation/memory-barriers.txt
> index b12df9137e1c..832b5d36e279 100644
> --- a/Documentation/memory-barriers.txt
> +++ b/Documentation/memory-barriers.txt
> @@ -1894,6 +1894,7 @@ There are some more advanced barrier functions:
>  
>   (*) dma_wmb();
>   (*) dma_rmb();
> + (*) dma_mb();
>  
>       These are for use with consistent memory to guarantee the ordering
>       of writes or reads of shared memory accessible to both the CPU and a
> @@ -1925,11 +1926,11 @@ There are some more advanced barrier functions:
>       The dma_rmb() allows us guarantee the device has released ownership
>       before we read the data from the descriptor, and the dma_wmb() allows
>       us to guarantee the data is written to the descriptor before the device
> -     can see it now has ownership.  Note that, when using writel(), a prior
> -     wmb() is not needed to guarantee that the cache coherent memory writes
> -     have completed before writing to the MMIO region.  The cheaper
> -     writel_relaxed() does not provide this guarantee and must not be used
> -     here.
> +     can see it now has ownership.  The dma_mb() implies both a dma_rmb() and
> +     a dma_wmb().  Note that, when using writel(), a prior wmb() is not needed
> +     to guarantee that the cache coherent memory writes have completed before
> +     writing to the MMIO region.  The cheaper writel_relaxed() does not provide
> +     this guarantee and must not be used here.
>  
>       See the subsection "Kernel I/O barrier effects" for more information on
>       relaxed I/O accessors and the Documentation/core-api/dma-api.rst file for
> diff --git a/include/asm-generic/barrier.h b/include/asm-generic/barrier.h
> index fd7e8fbaeef1..961f4d88f9ef 100644
> --- a/include/asm-generic/barrier.h
> +++ b/include/asm-generic/barrier.h
> @@ -38,6 +38,10 @@
>  #define wmb()	do { kcsan_wmb(); __wmb(); } while (0)
>  #endif
>  
> +#ifdef __dma_mb
> +#define dma_mb()	do { kcsan_mb(); __dma_mb(); } while (0)
> +#endif
> +
>  #ifdef __dma_rmb
>  #define dma_rmb()	do { kcsan_rmb(); __dma_rmb(); } while (0)
>  #endif
> @@ -65,6 +69,10 @@
>  #define wmb()	mb()
>  #endif
>  
> +#ifndef dma_mb
> +#define dma_mb()	mb()
> +#endif
> +
>  #ifndef dma_rmb
>  #define dma_rmb()	rmb()
>  #endif
> -- 
> 2.35.3
>
Paul E. McKenney June 16, 2022, 11:13 p.m. UTC | #3
On Mon, May 23, 2022 at 01:35:27PM +0200, Marco Elver wrote:
> On Mon, 23 May 2022 at 13:21, Kefeng Wang <wangkefeng.wang@huawei.com> wrote:
> >
> > The memory barrier dma_mb() is introduced by commit a76a37777f2c
> > ("iommu/arm-smmu-v3: Ensure queue is read after updating prod pointer"),
> > which is used to ensure that prior (both reads and writes) accesses
> > to memory by a CPU are ordered w.r.t. a subsequent MMIO write.
> >
> > Reviewed-by: Arnd Bergmann <arnd@arndb.de> # for asm-generic
> > Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
> 
> Reviewed-by: Marco Elver <elver@google.com>

Just checking...  Did these ever get picked up?  It was suggested
that they go up via the arm64 tree, if I remember correctly.

							Thanx, Paul

> > ---
> >  Documentation/memory-barriers.txt | 11 ++++++-----
> >  include/asm-generic/barrier.h     |  8 ++++++++
> >  2 files changed, 14 insertions(+), 5 deletions(-)
> >
> > diff --git a/Documentation/memory-barriers.txt b/Documentation/memory-barriers.txt
> > index b12df9137e1c..832b5d36e279 100644
> > --- a/Documentation/memory-barriers.txt
> > +++ b/Documentation/memory-barriers.txt
> > @@ -1894,6 +1894,7 @@ There are some more advanced barrier functions:
> >
> >   (*) dma_wmb();
> >   (*) dma_rmb();
> > + (*) dma_mb();
> >
> >       These are for use with consistent memory to guarantee the ordering
> >       of writes or reads of shared memory accessible to both the CPU and a
> > @@ -1925,11 +1926,11 @@ There are some more advanced barrier functions:
> >       The dma_rmb() allows us guarantee the device has released ownership
> >       before we read the data from the descriptor, and the dma_wmb() allows
> >       us to guarantee the data is written to the descriptor before the device
> > -     can see it now has ownership.  Note that, when using writel(), a prior
> > -     wmb() is not needed to guarantee that the cache coherent memory writes
> > -     have completed before writing to the MMIO region.  The cheaper
> > -     writel_relaxed() does not provide this guarantee and must not be used
> > -     here.
> > +     can see it now has ownership.  The dma_mb() implies both a dma_rmb() and
> > +     a dma_wmb().  Note that, when using writel(), a prior wmb() is not needed
> > +     to guarantee that the cache coherent memory writes have completed before
> > +     writing to the MMIO region.  The cheaper writel_relaxed() does not provide
> > +     this guarantee and must not be used here.
> >
> >       See the subsection "Kernel I/O barrier effects" for more information on
> >       relaxed I/O accessors and the Documentation/core-api/dma-api.rst file for
> > diff --git a/include/asm-generic/barrier.h b/include/asm-generic/barrier.h
> > index fd7e8fbaeef1..961f4d88f9ef 100644
> > --- a/include/asm-generic/barrier.h
> > +++ b/include/asm-generic/barrier.h
> > @@ -38,6 +38,10 @@
> >  #define wmb()  do { kcsan_wmb(); __wmb(); } while (0)
> >  #endif
> >
> > +#ifdef __dma_mb
> > +#define dma_mb()       do { kcsan_mb(); __dma_mb(); } while (0)
> > +#endif
> > +
> >  #ifdef __dma_rmb
> >  #define dma_rmb()      do { kcsan_rmb(); __dma_rmb(); } while (0)
> >  #endif
> > @@ -65,6 +69,10 @@
> >  #define wmb()  mb()
> >  #endif
> >
> > +#ifndef dma_mb
> > +#define dma_mb()       mb()
> > +#endif
> > +
> >  #ifndef dma_rmb
> >  #define dma_rmb()      rmb()
> >  #endif
> > --
> > 2.35.3
> >
Marco Elver June 17, 2022, 10:18 a.m. UTC | #4
On Fri, 17 Jun 2022 at 01:13, Paul E. McKenney <paulmck@kernel.org> wrote:
>
> On Mon, May 23, 2022 at 01:35:27PM +0200, Marco Elver wrote:
> > On Mon, 23 May 2022 at 13:21, Kefeng Wang <wangkefeng.wang@huawei.com> wrote:
> > >
> > > The memory barrier dma_mb() is introduced by commit a76a37777f2c
> > > ("iommu/arm-smmu-v3: Ensure queue is read after updating prod pointer"),
> > > which is used to ensure that prior (both reads and writes) accesses
> > > to memory by a CPU are ordered w.r.t. a subsequent MMIO write.
> > >
> > > Reviewed-by: Arnd Bergmann <arnd@arndb.de> # for asm-generic
> > > Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
> >
> > Reviewed-by: Marco Elver <elver@google.com>
>
> Just checking...  Did these ever get picked up?  It was suggested
> that they go up via the arm64 tree, if I remember correctly.

I don't see them in -next, and as far as I can tell, they're not in
the arm64 tree.

Thanks,
-- Marco
Catalin Marinas June 19, 2022, 9:45 a.m. UTC | #5
On Fri, Jun 17, 2022 at 12:18:41PM +0200, Marco Elver wrote:
> On Fri, 17 Jun 2022 at 01:13, Paul E. McKenney <paulmck@kernel.org> wrote:
> > On Mon, May 23, 2022 at 01:35:27PM +0200, Marco Elver wrote:
> > > On Mon, 23 May 2022 at 13:21, Kefeng Wang <wangkefeng.wang@huawei.com> wrote:
> > > >
> > > > The memory barrier dma_mb() is introduced by commit a76a37777f2c
> > > > ("iommu/arm-smmu-v3: Ensure queue is read after updating prod pointer"),
> > > > which is used to ensure that prior (both reads and writes) accesses
> > > > to memory by a CPU are ordered w.r.t. a subsequent MMIO write.
> > > >
> > > > Reviewed-by: Arnd Bergmann <arnd@arndb.de> # for asm-generic
> > > > Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
> > >
> > > Reviewed-by: Marco Elver <elver@google.com>
> >
> > Just checking...  Did these ever get picked up?  It was suggested
> > that they go up via the arm64 tree, if I remember correctly.
> 
> I don't see them in -next, and as far as I can tell, they're not in
> the arm64 tree.

Since v4 was posted during the merging window, it hasn't been queued for
5.19-rc1. I normally only merge patches with a Fixes tag during the -rc
period (though there are some exceptions). Mark commented in v1 that
such tag isn't necessary, so I thought I'd leave it for the 5.20 merging
window.

That said, the diffstat is small, so if it helps having this in 5.19, I
can queue it for -rc4.
Paul E. McKenney June 20, 2022, 9:02 p.m. UTC | #6
On Sun, Jun 19, 2022 at 10:45:22AM +0100, Catalin Marinas wrote:
> On Fri, Jun 17, 2022 at 12:18:41PM +0200, Marco Elver wrote:
> > On Fri, 17 Jun 2022 at 01:13, Paul E. McKenney <paulmck@kernel.org> wrote:
> > > On Mon, May 23, 2022 at 01:35:27PM +0200, Marco Elver wrote:
> > > > On Mon, 23 May 2022 at 13:21, Kefeng Wang <wangkefeng.wang@huawei.com> wrote:
> > > > >
> > > > > The memory barrier dma_mb() is introduced by commit a76a37777f2c
> > > > > ("iommu/arm-smmu-v3: Ensure queue is read after updating prod pointer"),
> > > > > which is used to ensure that prior (both reads and writes) accesses
> > > > > to memory by a CPU are ordered w.r.t. a subsequent MMIO write.
> > > > >
> > > > > Reviewed-by: Arnd Bergmann <arnd@arndb.de> # for asm-generic
> > > > > Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
> > > >
> > > > Reviewed-by: Marco Elver <elver@google.com>
> > >
> > > Just checking...  Did these ever get picked up?  It was suggested
> > > that they go up via the arm64 tree, if I remember correctly.
> > 
> > I don't see them in -next, and as far as I can tell, they're not in
> > the arm64 tree.
> 
> Since v4 was posted during the merging window, it hasn't been queued for
> 5.19-rc1. I normally only merge patches with a Fixes tag during the -rc
> period (though there are some exceptions). Mark commented in v1 that
> such tag isn't necessary, so I thought I'd leave it for the 5.20 merging
> window.
> 
> That said, the diffstat is small, so if it helps having this in 5.19, I
> can queue it for -rc4.

As long as it is not going to be lost, I am good.  ;-)

So v5.20 is fine.

							Thanx, Paul
diff mbox series

Patch

diff --git a/Documentation/memory-barriers.txt b/Documentation/memory-barriers.txt
index b12df9137e1c..832b5d36e279 100644
--- a/Documentation/memory-barriers.txt
+++ b/Documentation/memory-barriers.txt
@@ -1894,6 +1894,7 @@  There are some more advanced barrier functions:
 
  (*) dma_wmb();
  (*) dma_rmb();
+ (*) dma_mb();
 
      These are for use with consistent memory to guarantee the ordering
      of writes or reads of shared memory accessible to both the CPU and a
@@ -1925,11 +1926,11 @@  There are some more advanced barrier functions:
      The dma_rmb() allows us guarantee the device has released ownership
      before we read the data from the descriptor, and the dma_wmb() allows
      us to guarantee the data is written to the descriptor before the device
-     can see it now has ownership.  Note that, when using writel(), a prior
-     wmb() is not needed to guarantee that the cache coherent memory writes
-     have completed before writing to the MMIO region.  The cheaper
-     writel_relaxed() does not provide this guarantee and must not be used
-     here.
+     can see it now has ownership.  The dma_mb() implies both a dma_rmb() and
+     a dma_wmb().  Note that, when using writel(), a prior wmb() is not needed
+     to guarantee that the cache coherent memory writes have completed before
+     writing to the MMIO region.  The cheaper writel_relaxed() does not provide
+     this guarantee and must not be used here.
 
      See the subsection "Kernel I/O barrier effects" for more information on
      relaxed I/O accessors and the Documentation/core-api/dma-api.rst file for
diff --git a/include/asm-generic/barrier.h b/include/asm-generic/barrier.h
index fd7e8fbaeef1..961f4d88f9ef 100644
--- a/include/asm-generic/barrier.h
+++ b/include/asm-generic/barrier.h
@@ -38,6 +38,10 @@ 
 #define wmb()	do { kcsan_wmb(); __wmb(); } while (0)
 #endif
 
+#ifdef __dma_mb
+#define dma_mb()	do { kcsan_mb(); __dma_mb(); } while (0)
+#endif
+
 #ifdef __dma_rmb
 #define dma_rmb()	do { kcsan_rmb(); __dma_rmb(); } while (0)
 #endif
@@ -65,6 +69,10 @@ 
 #define wmb()	mb()
 #endif
 
+#ifndef dma_mb
+#define dma_mb()	mb()
+#endif
+
 #ifndef dma_rmb
 #define dma_rmb()	rmb()
 #endif