From patchwork Thu Dec 21 11:25:55 2017 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Jeff Layton X-Patchwork-Id: 10127509 Return-Path: Received: from mail.wl.linuxfoundation.org (pdx-wl-mail.web.codeaurora.org [172.30.200.125]) by pdx-korg-patchwork.web.codeaurora.org (Postfix) with ESMTP id C74D76019C for ; Thu, 21 Dec 2017 11:26:19 +0000 (UTC) Received: from mail.wl.linuxfoundation.org (localhost [127.0.0.1]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 8C38B29BCF for ; Thu, 21 Dec 2017 11:26:19 +0000 (UTC) Received: by mail.wl.linuxfoundation.org (Postfix, from userid 486) id 7B04029BD3; Thu, 21 Dec 2017 11:26:19 +0000 (UTC) X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on pdx-wl-mail.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-6.9 required=2.0 tests=BAYES_00,RCVD_IN_DNSWL_HI autolearn=unavailable version=3.3.1 Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 6512729BCF for ; Thu, 21 Dec 2017 11:26:18 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752066AbdLUL0C (ORCPT ); Thu, 21 Dec 2017 06:26:02 -0500 Received: from mail.kernel.org ([198.145.29.99]:56228 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1750884AbdLUL0A (ORCPT ); Thu, 21 Dec 2017 06:26:00 -0500 Received: from tleilax.poochiereds.net (cpe-71-70-156-158.nc.res.rr.com [71.70.156.158]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id 2BCE52192D; Thu, 21 Dec 2017 11:25:57 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 2BCE52192D Authentication-Results: mail.kernel.org; dmarc=none (p=none dis=none) header.from=kernel.org Authentication-Results: mail.kernel.org; spf=none smtp.mailfrom=jlayton@kernel.org Message-ID: <1513855555.3410.3.camel@kernel.org> Subject: Re: [PATCH v3 19/19] fs: handle inode->i_version more efficiently From: Jeff Layton To: Jan Kara Cc: Dave Chinner , "J. Bruce Fields" , linux-fsdevel@vger.kernel.org, linux-kernel@vger.kernel.org, viro@zeniv.linux.org.uk, linux-nfs@vger.kernel.org, neilb@suse.de, jack@suse.de, linux-ext4@vger.kernel.org, tytso@mit.edu, adilger.kernel@dilger.ca, linux-xfs@vger.kernel.org, darrick.wong@oracle.com, linux-btrfs@vger.kernel.org, clm@fb.com, jbacik@fb.com, dsterba@suse.com, linux-integrity@vger.kernel.org, zohar@linux.vnet.ibm.com, dmitry.kasatkin@gmail.com, linux-afs@lists.infradead.org, dhowells@redhat.com, jaltman@auristor.com Date: Thu, 21 Dec 2017 06:25:55 -0500 In-Reply-To: <20171220164150.GF31584@quack2.suse.cz> References: <20171218151156.14565-1-jlayton@kernel.org> <20171218151156.14565-20-jlayton@kernel.org> <20171218163418.GF30350@quack2.suse.cz> <1513617740.3422.30.camel@kernel.org> <20171218173651.GB12454@fieldses.org> <1513625720.3422.68.camel@kernel.org> <20171218220759.GF4094@dastard> <1513778586.4513.18.camel@kernel.org> <20171220164150.GF31584@quack2.suse.cz> X-Mailer: Evolution 3.26.3 (3.26.3-1.fc27) Mime-Version: 1.0 Sender: linux-btrfs-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-btrfs@vger.kernel.org X-Virus-Scanned: ClamAV using ClamSMTP On Wed, 2017-12-20 at 17:41 +0100, Jan Kara wrote: > On Wed 20-12-17 09:03:06, Jeff Layton wrote: > > On Tue, 2017-12-19 at 09:07 +1100, Dave Chinner wrote: > > > On Mon, Dec 18, 2017 at 02:35:20PM -0500, Jeff Layton wrote: > > > > [PATCH] SQUASH: add memory barriers around i_version accesses > > > > > > Why explicit memory barriers rather than annotating the operations > > > with the required semantics and getting the barriers the arch > > > requires automatically? I suspect this should be using > > > atomic_read_acquire() and atomic_cmpxchg_release(), because AFAICT > > > the atomic_cmpxchg needs to have release semantics to match the > > > acquire semantics needed for the load of the current value. > > > > > > From include/linux/atomics.h: > > > > > > * For compound atomics performing both a load and a store, ACQUIRE > > > * semantics apply only to the load and RELEASE semantics only to the > > > * store portion of the operation. Note that a failed cmpxchg_acquire > > > * does -not- imply any memory ordering constraints. > > > > > > Memory barriers hurt my brain. :/ > > > > > > At minimum, shouldn't the atomic op specific barriers be used rather > > > than full memory barriers? i.e: > > > > > > > They hurt my brain too. Yes, definitely atomic-specific barriers should > > be used here instead, since this is an atomic64_t now. > > > > After going over the docs again...my understanding has always been that > > you primarily need memory barriers to order accesses to different areas > > of memory. > > That is correct. > > > As Jan and I were discussing in the other thread, i_version is not > > synchronized with anything else. In this code, we're only dealing with a > > single 64-bit word. I don't think there are any races there wrt the API > > itself. > > There are not but it is like saying that lock implementation is correct > because the lock state does not get corrupted ;). Who cares about protected > data... > > > The "legacy" inode_inc_iversion() however _does_ have implied memory > > barriers from the i_lock. There could be some subtle de-facto ordering > > there, so I think we probably do want some barriers in here if only to > > preserve that. It's not likely to cost much, and may save us tracking > > down some fiddly bugs. > > > > What about this patch? Note that I've only added barriers to > > inode_maybe_inc_iversion. I don't see that we need it for the other > > functions, but please do tell me if I'm wrong there: > > > > --------------------8<--------------------------- > > > > [PATCH] SQUASH: add memory barriers around i_version accesses > > > > Signed-off-by: Jeff Layton > > --- > > include/linux/iversion.h | 53 +++++++++++++++++++++++++++++------------------- > > 1 file changed, 32 insertions(+), 21 deletions(-) > > > > diff --git a/include/linux/iversion.h b/include/linux/iversion.h > > index a9fbf99709df..02187a3bec3b 100644 > > --- a/include/linux/iversion.h > > +++ b/include/linux/iversion.h > > @@ -89,6 +89,23 @@ inode_set_iversion_raw(struct inode *inode, const u64 val) > > atomic64_set(&inode->i_version, val); > > } > > > > +/** > > + * inode_peek_iversion_raw - grab a "raw" iversion value > > + * @inode: inode from which i_version should be read > > + * > > + * Grab a "raw" inode->i_version value and return it. The i_version is not > > + * flagged or converted in any way. This is mostly used to access a self-managed > > + * i_version. > > + * > > + * With those filesystems, we want to treat the i_version as an entirely > > + * opaque value. > > + */ > > +static inline u64 > > +inode_peek_iversion_raw(const struct inode *inode) > > +{ > > + return atomic64_read(&inode->i_version); > > +} > > + > > /** > > * inode_set_iversion - set i_version to a particular value > > * @inode: inode to set > > @@ -152,7 +169,16 @@ inode_maybe_inc_iversion(struct inode *inode, bool force) > > { > > u64 cur, old, new; > > > > - cur = (u64)atomic64_read(&inode->i_version); > > + /* > > + * The i_version field is not strictly ordered with any other inode > > + * information, but the legacy inode_inc_iversion code used a spinlock > > + * to serialize increments. > > + * > > + * This code adds full memory barriers to ensure that any de-facto > > + * ordering with other info is preserved. > > + */ > > + smp_mb__before_atomic(); > > This should be just smp_mb(). __before_atomic() pairs with atomic > operations like atomic_inc(). atomic_read() is completely unordered > operation (happens to be plain memory read on x86) and so __before_atomic() > is not enough. > > > + cur = inode_peek_iversion_raw(inode); > > for (;;) { > > /* If flag is clear then we needn't do anything */ > > if (!force && !(cur & I_VERSION_QUERIED)) > > @@ -162,8 +188,10 @@ inode_maybe_inc_iversion(struct inode *inode, bool force) > > new = (cur & ~I_VERSION_QUERIED) + I_VERSION_INCREMENT; > > > > old = atomic64_cmpxchg(&inode->i_version, cur, new); > > - if (likely(old == cur)) > > + if (likely(old == cur)) { > > + smp_mb__after_atomic(); > > I don't think you need this. Cmpxchg is guaranteed to be full memory > barrier - from Documentation/atomic_t.txt: > - RMW operations that have a return value are fully ordered; > > > break; > > + } > > cur = old; > > } > > return true; > > ... > > > @@ -248,7 +259,7 @@ inode_query_iversion(struct inode *inode) > > { > > u64 cur, old, new; > > > > - cur = atomic64_read(&inode->i_version); > > + cur = inode_peek_iversion_raw(inode); > > for (;;) { > > /* If flag is already set, then no need to swap */ > > if (cur & I_VERSION_QUERIED) > > And here I'd expect smp_mb() after inode_peek_iversion_raw() (actually be > needed only if you are not going to do cmpxchg as that implies barrier as > well). "Safe" use of i_version would be: > > Update: > > modify inode > inode_maybe_inc_iversion(inode) > > Read: > > my_version = inode_query_iversion(inode) > get inode data > > And you need to make sure 'get inode data' does not get speculatively > evaluated before you actually sample i_version so that you are guaranteed > that if data changes, you will observe larger i_version in the future. > > Also please add a comment smp_mb() in inode_maybe_inc_iversion() like: > > /* This barrier pairs with the barrier in inode_query_iversion() */ > > and a similar comment to inode_query_iversion(). Because memory barriers > make sense only in pairs (see SMP BARRIER PAIRING in > Documentation/memory-barriers.txt). > Got it, I think. How about this (sorry for the unrelated deltas here): [PATCH] SQUASH: add memory barriers around i_version accesses Signed-off-by: Jeff Layton --- include/linux/iversion.h | 60 +++++++++++++++++++++++++++++++----------------- 1 file changed, 39 insertions(+), 21 deletions(-) diff --git a/include/linux/iversion.h b/include/linux/iversion.h index a9fbf99709df..1b3b5ed7c5b8 100644 --- a/include/linux/iversion.h +++ b/include/linux/iversion.h @@ -89,6 +89,23 @@ inode_set_iversion_raw(struct inode *inode, const u64 val) atomic64_set(&inode->i_version, val); } +/** + * inode_peek_iversion_raw - grab a "raw" iversion value + * @inode: inode from which i_version should be read + * + * Grab a "raw" inode->i_version value and return it. The i_version is not + * flagged or converted in any way. This is mostly used to access a self-managed + * i_version. + * + * With those filesystems, we want to treat the i_version as an entirely + * opaque value. + */ +static inline u64 +inode_peek_iversion_raw(const struct inode *inode) +{ + return atomic64_read(&inode->i_version); +} + /** * inode_set_iversion - set i_version to a particular value * @inode: inode to set @@ -152,7 +169,18 @@ inode_maybe_inc_iversion(struct inode *inode, bool force) { u64 cur, old, new; - cur = (u64)atomic64_read(&inode->i_version); + /* + * The i_version field is not strictly ordered with any other inode + * information, but the legacy inode_inc_iversion code used a spinlock + * to serialize increments. + * + * Here, we add full memory barriers to ensure that any de-facto + * ordering with other info is preserved. + * + * This barrier pairs with the barrier in inode_query_iversion() + */ + smp_mb(); + cur = inode_peek_iversion_raw(inode); for (;;) { /* If flag is clear then we needn't do anything */ if (!force && !(cur & I_VERSION_QUERIED)) @@ -183,23 +211,6 @@ inode_inc_iversion(struct inode *inode) inode_maybe_inc_iversion(inode, true); } -/** - * inode_peek_iversion_raw - grab a "raw" iversion value - * @inode: inode from which i_version should be read - * - * Grab a "raw" inode->i_version value and return it. The i_version is not - * flagged or converted in any way. This is mostly used to access a self-managed - * i_version. - * - * With those filesystems, we want to treat the i_version as an entirely - * opaque value. - */ -static inline u64 -inode_peek_iversion_raw(const struct inode *inode) -{ - return atomic64_read(&inode->i_version); -} - /** * inode_iversion_need_inc - is the i_version in need of being incremented? * @inode: inode to check @@ -248,15 +259,22 @@ inode_query_iversion(struct inode *inode) { u64 cur, old, new; - cur = atomic64_read(&inode->i_version); + cur = inode_peek_iversion_raw(inode); for (;;) { /* If flag is already set, then no need to swap */ - if (cur & I_VERSION_QUERIED) + if (cur & I_VERSION_QUERIED) { + /* + * This barrier (and the implicit barrier in the + * cmpxchg below) pairs with the barrier in + * inode_maybe_inc_iversion(). + */ + smp_mb(); break; + } new = cur | I_VERSION_QUERIED; old = atomic64_cmpxchg(&inode->i_version, cur, new); - if (old == cur) + if (likely(old == cur)) break; cur = old; }