Message ID | 20210907212347.1977686-1-minchan@kernel.org (mailing list archive) |
---|---|
State | New |
Headers | show |
Series | [v3] mm: fs: invalidate bh_lrus for only cold path | expand |
Andrew, Could you take a look? On Tue, Sep 07, 2021 at 02:23:47PM -0700, Minchan Kim wrote: > kernel test robot reported the regression of fio.write_iops[1] > with [2]. > > Since lru_add_drain is called frequently, invalidate bh_lrus > there could increase bh_lrus cache miss ratio, which needs > more IO in the end. > > This patch moves the bh_lrus invalidation from the hot path( > e.g., zap_page_range, pagevec_release) to cold path(i.e., > lru_add_drain_all, lru_cache_disable). > > [1] https://lore.kernel.org/lkml/20210520083144.GD14190@xsang-OptiPlex-9020/ > [2] 8cc621d2f45d, mm: fs: invalidate BH LRU during page migration > Reviewed-by: Chris Goldsworthy <cgoldswo@codeaurora.org> > Reported-by: kernel test robot <oliver.sang@intel.com> > Signed-off-by: Minchan Kim <minchan@kernel.org> > --- > * v2: https://lore.kernel.org/lkml/20210601145425.1396981-1-minchan@kernel.org/ > * v1: https://lore.kernel.org/lkml/YK0oQ76zX0uVZCwQ@google.com/ > fs/buffer.c | 8 ++++++-- > include/linux/buffer_head.h | 4 ++-- > mm/swap.c | 19 ++++++++++++++++--- > 3 files changed, 24 insertions(+), 7 deletions(-) > > diff --git a/fs/buffer.c b/fs/buffer.c > index ab7573d72dd7..c615387aedca 100644 > --- a/fs/buffer.c > +++ b/fs/buffer.c > @@ -1425,12 +1425,16 @@ void invalidate_bh_lrus(void) > } > EXPORT_SYMBOL_GPL(invalidate_bh_lrus); > > -void invalidate_bh_lrus_cpu(int cpu) > +/* > + * It's called from workqueue context so we need a bh_lru_lock to close > + * the race with preemption/irq. > + */ > +void invalidate_bh_lrus_cpu(void) > { > struct bh_lru *b; > > bh_lru_lock(); > - b = per_cpu_ptr(&bh_lrus, cpu); > + b = this_cpu_ptr(&bh_lrus); > __invalidate_bh_lrus(b); > bh_lru_unlock(); > } > diff --git a/include/linux/buffer_head.h b/include/linux/buffer_head.h > index 6486d3c19463..36f33685c8c0 100644 > --- a/include/linux/buffer_head.h > +++ b/include/linux/buffer_head.h > @@ -194,7 +194,7 @@ void __breadahead_gfp(struct block_device *, sector_t block, unsigned int size, > struct buffer_head *__bread_gfp(struct block_device *, > sector_t block, unsigned size, gfp_t gfp); > void invalidate_bh_lrus(void); > -void invalidate_bh_lrus_cpu(int cpu); > +void invalidate_bh_lrus_cpu(void); > bool has_bh_in_lru(int cpu, void *dummy); > struct buffer_head *alloc_buffer_head(gfp_t gfp_flags); > void free_buffer_head(struct buffer_head * bh); > @@ -408,7 +408,7 @@ static inline int inode_has_buffers(struct inode *inode) { return 0; } > static inline void invalidate_inode_buffers(struct inode *inode) {} > static inline int remove_inode_buffers(struct inode *inode) { return 1; } > static inline int sync_mapping_buffers(struct address_space *mapping) { return 0; } > -static inline void invalidate_bh_lrus_cpu(int cpu) {} > +static inline void invalidate_bh_lrus_cpu(void) {} > static inline bool has_bh_in_lru(int cpu, void *dummy) { return false; } > #define buffer_heads_over_limit 0 > > diff --git a/mm/swap.c b/mm/swap.c > index 897200d27dd0..af3cad4e5378 100644 > --- a/mm/swap.c > +++ b/mm/swap.c > @@ -620,7 +620,6 @@ void lru_add_drain_cpu(int cpu) > pagevec_lru_move_fn(pvec, lru_lazyfree_fn); > > activate_page_drain(cpu); > - invalidate_bh_lrus_cpu(cpu); > } > > /** > @@ -703,6 +702,20 @@ void lru_add_drain(void) > local_unlock(&lru_pvecs.lock); > } > > +/* > + * It's called from per-cpu workqueue context in SMP case so > + * lru_add_drain_cpu and invalidate_bh_lrus_cpu should run on > + * the same cpu. It shouldn't be a problem in !SMP case since > + * the core is only one and the locks will disable preemption. > + */ > +static void lru_add_and_bh_lrus_drain(void) > +{ > + local_lock(&lru_pvecs.lock); > + lru_add_drain_cpu(smp_processor_id()); > + local_unlock(&lru_pvecs.lock); > + invalidate_bh_lrus_cpu(); > +} > + > void lru_add_drain_cpu_zone(struct zone *zone) > { > local_lock(&lru_pvecs.lock); > @@ -717,7 +730,7 @@ static DEFINE_PER_CPU(struct work_struct, lru_add_drain_work); > > static void lru_add_drain_per_cpu(struct work_struct *dummy) > { > - lru_add_drain(); > + lru_add_and_bh_lrus_drain(); > } > > /* > @@ -858,7 +871,7 @@ void lru_cache_disable(void) > */ > __lru_add_drain_all(true); > #else > - lru_add_drain(); > + lru_add_and_bh_lrus_drain(); > #endif > } > > -- > 2.33.0.309.g3052b89438-goog >
On Tue, Sep 7, 2021 at 2:24 PM Minchan Kim <minchan@kernel.org> wrote: > > kernel test robot reported the regression of fio.write_iops[1] > with [2]. > > Since lru_add_drain is called frequently, invalidate bh_lrus > there could increase bh_lrus cache miss ratio, which needs > more IO in the end. > > This patch moves the bh_lrus invalidation from the hot path( > e.g., zap_page_range, pagevec_release) to cold path(i.e., > lru_add_drain_all, lru_cache_disable). Was this confirmed to fix the regression? I only see the "tested with 5.14" that the regression was still there https://lore.kernel.org/lkml/034fc860-d0d0-0c61-09d2-3c41c4f020c6@intel.com/ I don't see a confirmation that this patch fixed it. It looks likely, but if you have the confirmation somewhere, it would help to link that too. Or did I miss it? Linus
On Mon, Sep 20, 2021 at 04:29:52PM -0700, Linus Torvalds wrote: > On Tue, Sep 7, 2021 at 2:24 PM Minchan Kim <minchan@kernel.org> wrote: > > > > kernel test robot reported the regression of fio.write_iops[1] > > with [2]. > > > > Since lru_add_drain is called frequently, invalidate bh_lrus > > there could increase bh_lrus cache miss ratio, which needs > > more IO in the end. > > > > This patch moves the bh_lrus invalidation from the hot path( > > e.g., zap_page_range, pagevec_release) to cold path(i.e., > > lru_add_drain_all, lru_cache_disable). > > Was this confirmed to fix the regression? > > I only see the "tested with 5.14" that the regression was still there > > https://lore.kernel.org/lkml/034fc860-d0d0-0c61-09d2-3c41c4f020c6@intel.com/ > > I don't see a confirmation that this patch fixed it. > > It looks likely, but if you have the confirmation somewhere, it would > help to link that too. > > Or did I miss it? I have no idea why I couldn't find the reply from lore.kernel.org/lkml/ The message id was 89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com in the thread(https://lore.kernel.org/lkml/20210520083144.GD14190@xsang-OptiPlex-9020/) Xing, Zhengjun zhengjun.xing@intel.com confirmed Quote - ``` Hi Minchan, ... I test the patch, the regression reduced to -2.9%. ``` Since then, I modified the patch a little more and couldn't find the message from the public URL. Thus, do we need to confirm it again against latest kernel? Xing, could you confirm whether this patch fixes the regression?
On Mon, Sep 20, 2021 at 4:50 PM Minchan Kim <minchan@kernel.org> wrote: > > I have no idea why I couldn't find the reply from lore.kernel.org/lkml/ > The message id was 89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com > in the thread(https://lore.kernel.org/lkml/20210520083144.GD14190@xsang-OptiPlex-9020/) Hmm. That message-id isn't found anywhere. Maybe it was sent just to you personally? [ Goes off and looks ] Nope, I can see it in my mails too, and yeah, I see To: Minchan Kim <minchan@kernel.org>, kernel test robot <oliver.sang@intel.com> Cc: Linus Torvalds <torvalds@linux-foundation.org>, Chris Goldsworthy <cgoldswo@codeaurora.org>, Laura Abbott <labbott@kernel.org>, David Hildenbrand <david@redhat.com>, John Dias <joaodias@google.com>, Matthew Wilcox <willy@infradead.org>, Michal Hocko <mhocko@suse.com>, Suren Baghdasaryan <surenb@google.com>, Vlastimil Babka <vbabka@suse.cz>, Andrew Morton <akpm@linux-foundation.org>, LKML <linux-kernel@vger.kernel.org>, lkp@lists.01.org, lkp@intel.com References: <20210520083144.GD14190@xsang-OptiPlex-9020> <YKasEeXCr9R5yzCr@google.com> From: "Xing, Zhengjun" <zhengjun.xing@intel.com> Message-ID: <89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com> but lore has no idea about it: https://lore.kernel.org/all/89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com/ just says "Message-ID <89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com> not found". Strange. Something presumably caused the list to drop that email. Linus
On Mon, Sep 20, 2021 at 05:23:38PM -0700, Linus Torvalds wrote: > Hmm. That message-id isn't found anywhere. > > Maybe it was sent just to you personally? > > [ Goes off and looks ] > > Nope, I can see it in my mails too, and yeah, I see > > To: Minchan Kim <minchan@kernel.org>, kernel test robot > <oliver.sang@intel.com> > Cc: Linus Torvalds <torvalds@linux-foundation.org>, Chris > Goldsworthy <cgoldswo@codeaurora.org>, Laura Abbott > <labbott@kernel.org>, David Hildenbrand <david@redhat.com>, John Dias > <joaodias@google.com>, Matthew Wilcox <willy@infradead.org>, Michal > Hocko <mhocko@suse.com>, Suren Baghdasaryan <surenb@google.com>, > Vlastimil Babka <vbabka@suse.cz>, Andrew Morton > <akpm@linux-foundation.org>, LKML <linux-kernel@vger.kernel.org>, > lkp@lists.01.org, lkp@intel.com > References: <20210520083144.GD14190@xsang-OptiPlex-9020> > <YKasEeXCr9R5yzCr@google.com> > From: "Xing, Zhengjun" <zhengjun.xing@intel.com> > Message-ID: <89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com> > > but lore has no idea about it: > > https://lore.kernel.org/all/89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com/ > > just says "Message-ID <89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com> > not found". > > Strange. > > Something presumably caused the list to drop that email. I think vger is backed up. I've flagged this with John, hopefully he'll be able to figure out what's keeping things from hitting the archives. -K
diff --git a/fs/buffer.c b/fs/buffer.c index ab7573d72dd7..c615387aedca 100644 --- a/fs/buffer.c +++ b/fs/buffer.c @@ -1425,12 +1425,16 @@ void invalidate_bh_lrus(void) } EXPORT_SYMBOL_GPL(invalidate_bh_lrus); -void invalidate_bh_lrus_cpu(int cpu) +/* + * It's called from workqueue context so we need a bh_lru_lock to close + * the race with preemption/irq. + */ +void invalidate_bh_lrus_cpu(void) { struct bh_lru *b; bh_lru_lock(); - b = per_cpu_ptr(&bh_lrus, cpu); + b = this_cpu_ptr(&bh_lrus); __invalidate_bh_lrus(b); bh_lru_unlock(); } diff --git a/include/linux/buffer_head.h b/include/linux/buffer_head.h index 6486d3c19463..36f33685c8c0 100644 --- a/include/linux/buffer_head.h +++ b/include/linux/buffer_head.h @@ -194,7 +194,7 @@ void __breadahead_gfp(struct block_device *, sector_t block, unsigned int size, struct buffer_head *__bread_gfp(struct block_device *, sector_t block, unsigned size, gfp_t gfp); void invalidate_bh_lrus(void); -void invalidate_bh_lrus_cpu(int cpu); +void invalidate_bh_lrus_cpu(void); bool has_bh_in_lru(int cpu, void *dummy); struct buffer_head *alloc_buffer_head(gfp_t gfp_flags); void free_buffer_head(struct buffer_head * bh); @@ -408,7 +408,7 @@ static inline int inode_has_buffers(struct inode *inode) { return 0; } static inline void invalidate_inode_buffers(struct inode *inode) {} static inline int remove_inode_buffers(struct inode *inode) { return 1; } static inline int sync_mapping_buffers(struct address_space *mapping) { return 0; } -static inline void invalidate_bh_lrus_cpu(int cpu) {} +static inline void invalidate_bh_lrus_cpu(void) {} static inline bool has_bh_in_lru(int cpu, void *dummy) { return false; } #define buffer_heads_over_limit 0 diff --git a/mm/swap.c b/mm/swap.c index 897200d27dd0..af3cad4e5378 100644 --- a/mm/swap.c +++ b/mm/swap.c @@ -620,7 +620,6 @@ void lru_add_drain_cpu(int cpu) pagevec_lru_move_fn(pvec, lru_lazyfree_fn); activate_page_drain(cpu); - invalidate_bh_lrus_cpu(cpu); } /** @@ -703,6 +702,20 @@ void lru_add_drain(void) local_unlock(&lru_pvecs.lock); } +/* + * It's called from per-cpu workqueue context in SMP case so + * lru_add_drain_cpu and invalidate_bh_lrus_cpu should run on + * the same cpu. It shouldn't be a problem in !SMP case since + * the core is only one and the locks will disable preemption. + */ +static void lru_add_and_bh_lrus_drain(void) +{ + local_lock(&lru_pvecs.lock); + lru_add_drain_cpu(smp_processor_id()); + local_unlock(&lru_pvecs.lock); + invalidate_bh_lrus_cpu(); +} + void lru_add_drain_cpu_zone(struct zone *zone) { local_lock(&lru_pvecs.lock); @@ -717,7 +730,7 @@ static DEFINE_PER_CPU(struct work_struct, lru_add_drain_work); static void lru_add_drain_per_cpu(struct work_struct *dummy) { - lru_add_drain(); + lru_add_and_bh_lrus_drain(); } /* @@ -858,7 +871,7 @@ void lru_cache_disable(void) */ __lru_add_drain_all(true); #else - lru_add_drain(); + lru_add_and_bh_lrus_drain(); #endif }