Message ID | 1437994789-14133-1-git-send-email-jeff.layton@primarydata.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On Mon, 27 Jul 2015 06:59:49 -0400 Jeff Layton <jlayton@poochiereds.net> wrote: > Currently, we check to see if an open stateid needs updating, and then > update the stateid if so. The check and update however are not atomic, > so it's easily possible to end up finding an old seqid when we check > it only to have it updated by a newer one before we can get around to > updating it ourselves. > > We could try to play games with atomic ops here, but the simple fix > is to just ensure that we hold the per-stateid state_lock when > updating an open stateid. > > Signed-off-by: Jeff Layton <jeff.layton@primarydata.com> > --- > fs/nfs/nfs4proc.c | 15 ++++++++++----- > 1 file changed, 10 insertions(+), 5 deletions(-) > > diff --git a/fs/nfs/nfs4proc.c b/fs/nfs/nfs4proc.c > index 780accb962dd..bc6a7b5d81aa 100644 > --- a/fs/nfs/nfs4proc.c > +++ b/fs/nfs/nfs4proc.c > @@ -1234,14 +1234,17 @@ static void nfs_clear_open_stateid_locked(struct nfs4_state *state, > if (stateid == NULL) > return; > /* Handle races with OPEN */ > + spin_lock(&state->state_lock); > if (!nfs4_stateid_match_other(stateid, &state->open_stateid) || > !nfs4_stateid_is_newer(stateid, &state->open_stateid)) { > nfs_resync_open_stateid_locked(state); > + spin_unlock(&state->state_lock); > return; > } > if (test_bit(NFS_DELEGATED_STATE, &state->flags) == 0) > nfs4_stateid_copy(&state->stateid, stateid); > nfs4_stateid_copy(&state->open_stateid, stateid); > + spin_unlock(&state->state_lock); > } > > static void nfs_clear_open_stateid(struct nfs4_state *state, nfs4_stateid *stateid, fmode_t fmode) > @@ -1265,11 +1268,13 @@ static void nfs_set_open_stateid_locked(struct nfs4_state *state, nfs4_stateid * > case FMODE_READ|FMODE_WRITE: > set_bit(NFS_O_RDWR_STATE, &state->flags); > } > - if (!nfs_need_update_open_stateid(state, stateid)) > - return; > - if (test_bit(NFS_DELEGATED_STATE, &state->flags) == 0) > - nfs4_stateid_copy(&state->stateid, stateid); > - nfs4_stateid_copy(&state->open_stateid, stateid); > + spin_lock(&state->state_lock); > + if (nfs_need_update_open_stateid(state, stateid)) { > + if (test_bit(NFS_DELEGATED_STATE, &state->flags) == 0) > + nfs4_stateid_copy(&state->stateid, stateid); > + nfs4_stateid_copy(&state->open_stateid, stateid); > + } > + spin_unlock(&state->state_lock); > } > > static void __update_open_stateid(struct nfs4_state *state, nfs4_stateid *open_stateid, const nfs4_stateid *deleg_stateid, fmode_t fmode) Longer explanation for the RFC... I had a bug report where the client was sending a ton of parallel opens, and would sometimes get confused and either lose a newer stateid or overwrite it with a older one. This patch from Trond seems to have fixed the reproducer that we have: [PATCH] NFSv4: We must set NFS_OPEN_STATE flag in nfs_resync_open_stateid_locked ...but in looking over the code it seems pretty clear that there is a potential race between when we compare seqids in stateids and when they get copied. This patch should close the race for open stateids, but it's possible there are similar races for other types as well. I haven't had the time to go over the code in detail to be sure, and I don't have a reproducer that triggers this. Should we worry about this at all and try to fix it, or just let it go and assume that if it does occur the client will recover?
On Mon, Jul 27, 2015 at 7:07 AM, Jeff Layton <jeff.layton@primarydata.com> wrote: > > On Mon, 27 Jul 2015 06:59:49 -0400 > Jeff Layton <jlayton@poochiereds.net> wrote: > > > Currently, we check to see if an open stateid needs updating, and then > > update the stateid if so. The check and update however are not atomic, > > so it's easily possible to end up finding an old seqid when we check > > it only to have it updated by a newer one before we can get around to > > updating it ourselves. What am I missing? AFAICS, we always hold a write lock on the state->seqlock in those functions. The state->state_lock protects the state->lock_states list. It shouldn't have any function as far as protecting open stateids. Trond -- To unsubscribe from this list: send the line "unsubscribe linux-nfs" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On Mon, 27 Jul 2015 09:44:07 -0400 Trond Myklebust <trond.myklebust@primarydata.com> wrote: > On Mon, Jul 27, 2015 at 7:07 AM, Jeff Layton > <jeff.layton@primarydata.com> wrote: > > > > On Mon, 27 Jul 2015 06:59:49 -0400 > > Jeff Layton <jlayton@poochiereds.net> wrote: > > > > > Currently, we check to see if an open stateid needs updating, and then > > > update the stateid if so. The check and update however are not atomic, > > > so it's easily possible to end up finding an old seqid when we check > > > it only to have it updated by a newer one before we can get around to > > > updating it ourselves. > > What am I missing? AFAICS, we always hold a write lock on the > state->seqlock in those functions. > > The state->state_lock protects the state->lock_states list. It > shouldn't have any function as far as protecting open stateids. > > Trond Ahh my bad. You aren't missing anything. I was thinking that the write_seqlock calls would just bump a counter, and missed the fact that they have a spinlock embedded within them. In that case, there is no bug here and this patch can be dropped. Thanks!
diff --git a/fs/nfs/nfs4proc.c b/fs/nfs/nfs4proc.c index 780accb962dd..bc6a7b5d81aa 100644 --- a/fs/nfs/nfs4proc.c +++ b/fs/nfs/nfs4proc.c @@ -1234,14 +1234,17 @@ static void nfs_clear_open_stateid_locked(struct nfs4_state *state, if (stateid == NULL) return; /* Handle races with OPEN */ + spin_lock(&state->state_lock); if (!nfs4_stateid_match_other(stateid, &state->open_stateid) || !nfs4_stateid_is_newer(stateid, &state->open_stateid)) { nfs_resync_open_stateid_locked(state); + spin_unlock(&state->state_lock); return; } if (test_bit(NFS_DELEGATED_STATE, &state->flags) == 0) nfs4_stateid_copy(&state->stateid, stateid); nfs4_stateid_copy(&state->open_stateid, stateid); + spin_unlock(&state->state_lock); } static void nfs_clear_open_stateid(struct nfs4_state *state, nfs4_stateid *stateid, fmode_t fmode) @@ -1265,11 +1268,13 @@ static void nfs_set_open_stateid_locked(struct nfs4_state *state, nfs4_stateid * case FMODE_READ|FMODE_WRITE: set_bit(NFS_O_RDWR_STATE, &state->flags); } - if (!nfs_need_update_open_stateid(state, stateid)) - return; - if (test_bit(NFS_DELEGATED_STATE, &state->flags) == 0) - nfs4_stateid_copy(&state->stateid, stateid); - nfs4_stateid_copy(&state->open_stateid, stateid); + spin_lock(&state->state_lock); + if (nfs_need_update_open_stateid(state, stateid)) { + if (test_bit(NFS_DELEGATED_STATE, &state->flags) == 0) + nfs4_stateid_copy(&state->stateid, stateid); + nfs4_stateid_copy(&state->open_stateid, stateid); + } + spin_unlock(&state->state_lock); } static void __update_open_stateid(struct nfs4_state *state, nfs4_stateid *open_stateid, const nfs4_stateid *deleg_stateid, fmode_t fmode)
Currently, we check to see if an open stateid needs updating, and then update the stateid if so. The check and update however are not atomic, so it's easily possible to end up finding an old seqid when we check it only to have it updated by a newer one before we can get around to updating it ourselves. We could try to play games with atomic ops here, but the simple fix is to just ensure that we hold the per-stateid state_lock when updating an open stateid. Signed-off-by: Jeff Layton <jeff.layton@primarydata.com> --- fs/nfs/nfs4proc.c | 15 ++++++++++----- 1 file changed, 10 insertions(+), 5 deletions(-)