Message ID | 1403290612-15341-1-git-send-email-jlayton@primarydata.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Thanks, Jeff. Acked-by: J. Bruce Fields <bfields@redhat.com> --b. On Fri, Jun 20, 2014 at 02:56:52PM -0400, Jeff Layton wrote: > From: Jeff Layton <jlayton@redhat.com> > > This is commit a0ef5e19684f0447da9ff0654a12019c484f57ca in mainline. > > While the commit message below doesn't lay this out, we've subsequently > found that there are some cases where an entry that's still in use can > be freed prematurely if a particular operation takes a *very* long time > (on the order of minutes) and/or the server is very busy and doesn't > have a lot of memory dedicated to the DRC. This patch eliminates that > possibility, so it's actually more than just a cleanup. > > The regression crept in in v3.9, and this patch went into mainline in > v3.14. Please apply this to any stable kernel between those two > mainline releases. > > Original patch description follows: > > -------------------------------[snip]---------------------------- > > Currently when we are processing a request, we try to scrape an expired > or over-limit entry off the list in preference to allocating a new one > from the slab. > > This is unnecessarily complicated. Just use the slab layer. > > Signed-off-by: Jeff Layton <jlayton@redhat.com> > Signed-off-by: J. Bruce Fields <bfields@redhat.com> > --- > fs/nfsd/nfscache.c | 36 ++++-------------------------------- > 1 file changed, 4 insertions(+), 32 deletions(-) > > diff --git a/fs/nfsd/nfscache.c b/fs/nfsd/nfscache.c > index ec8d97ddc635..02e8e9ad5750 100644 > --- a/fs/nfsd/nfscache.c > +++ b/fs/nfsd/nfscache.c > @@ -129,13 +129,6 @@ nfsd_reply_cache_alloc(void) > } > > static void > -nfsd_reply_cache_unhash(struct svc_cacherep *rp) > -{ > - hlist_del_init(&rp->c_hash); > - list_del_init(&rp->c_lru); > -} > - > -static void > nfsd_reply_cache_free_locked(struct svc_cacherep *rp) > { > if (rp->c_type == RC_REPLBUFF && rp->c_replvec.iov_base) { > @@ -402,22 +395,8 @@ nfsd_cache_lookup(struct svc_rqst *rqstp) > > /* > * Since the common case is a cache miss followed by an insert, > - * preallocate an entry. First, try to reuse the first entry on the LRU > - * if it works, then go ahead and prune the LRU list. > + * preallocate an entry. > */ > - spin_lock(&cache_lock); > - if (!list_empty(&lru_head)) { > - rp = list_first_entry(&lru_head, struct svc_cacherep, c_lru); > - if (nfsd_cache_entry_expired(rp) || > - num_drc_entries >= max_drc_entries) { > - nfsd_reply_cache_unhash(rp); > - prune_cache_entries(); > - goto search_cache; > - } > - } > - > - /* No expired ones available, allocate a new one. */ > - spin_unlock(&cache_lock); > rp = nfsd_reply_cache_alloc(); > spin_lock(&cache_lock); > if (likely(rp)) { > @@ -425,7 +404,9 @@ nfsd_cache_lookup(struct svc_rqst *rqstp) > drc_mem_usage += sizeof(*rp); > } > > -search_cache: > + /* go ahead and prune the cache */ > + prune_cache_entries(); > + > found = nfsd_cache_search(rqstp, csum); > if (found) { > if (likely(rp)) > @@ -439,15 +420,6 @@ search_cache: > goto out; > } > > - /* > - * We're keeping the one we just allocated. Are we now over the > - * limit? Prune one off the tip of the LRU in trade for the one we > - * just allocated if so. > - */ > - if (num_drc_entries >= max_drc_entries) > - nfsd_reply_cache_free_locked(list_first_entry(&lru_head, > - struct svc_cacherep, c_lru)); > - > nfsdstats.rcmisses++; > rqstp->rq_cacherep = rp; > rp->c_state = RC_INPROG; > -- > 1.9.3 > -- To unsubscribe from this list: send the line "unsubscribe linux-nfs" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On 06/20/2014 08:56 PM, Jeff Layton wrote: > From: Jeff Layton <jlayton@redhat.com> > > This is commit a0ef5e19684f0447da9ff0654a12019c484f57ca in mainline. > > While the commit message below doesn't lay this out, we've subsequently > found that there are some cases where an entry that's still in use can > be freed prematurely if a particular operation takes a *very* long time > (on the order of minutes) and/or the server is very busy and doesn't > have a lot of memory dedicated to the DRC. This patch eliminates that > possibility, so it's actually more than just a cleanup. > > The regression crept in in v3.9, and this patch went into mainline in > v3.14. Please apply this to any stable kernel between those two > mainline releases. Now applied to 3.12. Thanks.
On Wed, 25 Jun 2014 15:32:56 +0200 Jiri Slaby <jslaby@suse.cz> wrote: > On 06/20/2014 08:56 PM, Jeff Layton wrote: > > From: Jeff Layton <jlayton@redhat.com> > > > > This is commit a0ef5e19684f0447da9ff0654a12019c484f57ca in mainline. > > > > While the commit message below doesn't lay this out, we've subsequently > > found that there are some cases where an entry that's still in use can > > be freed prematurely if a particular operation takes a *very* long time > > (on the order of minutes) and/or the server is very busy and doesn't > > have a lot of memory dedicated to the DRC. This patch eliminates that > > possibility, so it's actually more than just a cleanup. > > > > The regression crept in in v3.9, and this patch went into mainline in > > v3.14. Please apply this to any stable kernel between those two > > mainline releases. > > Now applied to 3.12. Thanks. > > Thank you! It also turns out that we're going to want to pull 1b19453d1c6abcfa7c312ba6c9f11a277568fc94 from mainline into stable kernels as well. There's one more possibility for an entry to be freed while still in use that that patch will fix. I'll resend that to stable@vger later today... Thanks!
On Wed, Jun 25, 2014 at 09:36:32AM -0400, Jeff Layton wrote: > On Wed, 25 Jun 2014 15:32:56 +0200 > Jiri Slaby <jslaby@suse.cz> wrote: > > > On 06/20/2014 08:56 PM, Jeff Layton wrote: > > > From: Jeff Layton <jlayton@redhat.com> > > > > > > This is commit a0ef5e19684f0447da9ff0654a12019c484f57ca in mainline. > > > > > > While the commit message below doesn't lay this out, we've subsequently > > > found that there are some cases where an entry that's still in use can > > > be freed prematurely if a particular operation takes a *very* long time > > > (on the order of minutes) and/or the server is very busy and doesn't > > > have a lot of memory dedicated to the DRC. This patch eliminates that > > > possibility, so it's actually more than just a cleanup. > > > > > > The regression crept in in v3.9, and this patch went into mainline in > > > v3.14. Please apply this to any stable kernel between those two > > > mainline releases. > > > > Now applied to 3.12. Thanks. > > > > > > Thank you! It also turns out that we're going to want to pull > 1b19453d1c6abcfa7c312ba6c9f11a277568fc94 from mainline into stable > kernels as well. There's one more possibility for an entry to be freed > while still in use that that patch will fix. > > I'll resend that to stable@vger later today... > > Thanks! > -- > Jeff Layton <jlayton@primarydata.com> > -- > To unsubscribe from this list: send the line "unsubscribe stable" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html Thanks, I'll queue both patches for the 3.11 kernel. Cheers, -- Luís -- To unsubscribe from this list: send the line "unsubscribe linux-nfs" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
diff --git a/fs/nfsd/nfscache.c b/fs/nfsd/nfscache.c index ec8d97ddc635..02e8e9ad5750 100644 --- a/fs/nfsd/nfscache.c +++ b/fs/nfsd/nfscache.c @@ -129,13 +129,6 @@ nfsd_reply_cache_alloc(void) } static void -nfsd_reply_cache_unhash(struct svc_cacherep *rp) -{ - hlist_del_init(&rp->c_hash); - list_del_init(&rp->c_lru); -} - -static void nfsd_reply_cache_free_locked(struct svc_cacherep *rp) { if (rp->c_type == RC_REPLBUFF && rp->c_replvec.iov_base) { @@ -402,22 +395,8 @@ nfsd_cache_lookup(struct svc_rqst *rqstp) /* * Since the common case is a cache miss followed by an insert, - * preallocate an entry. First, try to reuse the first entry on the LRU - * if it works, then go ahead and prune the LRU list. + * preallocate an entry. */ - spin_lock(&cache_lock); - if (!list_empty(&lru_head)) { - rp = list_first_entry(&lru_head, struct svc_cacherep, c_lru); - if (nfsd_cache_entry_expired(rp) || - num_drc_entries >= max_drc_entries) { - nfsd_reply_cache_unhash(rp); - prune_cache_entries(); - goto search_cache; - } - } - - /* No expired ones available, allocate a new one. */ - spin_unlock(&cache_lock); rp = nfsd_reply_cache_alloc(); spin_lock(&cache_lock); if (likely(rp)) { @@ -425,7 +404,9 @@ nfsd_cache_lookup(struct svc_rqst *rqstp) drc_mem_usage += sizeof(*rp); } -search_cache: + /* go ahead and prune the cache */ + prune_cache_entries(); + found = nfsd_cache_search(rqstp, csum); if (found) { if (likely(rp)) @@ -439,15 +420,6 @@ search_cache: goto out; } - /* - * We're keeping the one we just allocated. Are we now over the - * limit? Prune one off the tip of the LRU in trade for the one we - * just allocated if so. - */ - if (num_drc_entries >= max_drc_entries) - nfsd_reply_cache_free_locked(list_first_entry(&lru_head, - struct svc_cacherep, c_lru)); - nfsdstats.rcmisses++; rqstp->rq_cacherep = rp; rp->c_state = RC_INPROG;