Message ID | 20180305133743.12746-2-guro@fb.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
[+CC linux-api] On 03/05/2018 02:37 PM, Roman Gushchin wrote: > This patch introduces a concept of indirectly reclaimable memory > and adds the corresponding memory counter and /proc/vmstat item. > > Indirectly reclaimable memory is any sort of memory, used by > the kernel (except of reclaimable slabs), which is actually > reclaimable, i.e. will be released under memory pressure. > > The counter is in bytes, as it's not always possible to > count such objects in pages. The name contains BYTES > by analogy to NR_KERNEL_STACK_KB. > > Signed-off-by: Roman Gushchin <guro@fb.com> > Cc: Andrew Morton <akpm@linux-foundation.org> > Cc: Alexander Viro <viro@zeniv.linux.org.uk> > Cc: Michal Hocko <mhocko@suse.com> > Cc: Johannes Weiner <hannes@cmpxchg.org> > Cc: linux-fsdevel@vger.kernel.org > Cc: linux-kernel@vger.kernel.org > Cc: linux-mm@kvack.org > Cc: kernel-team@fb.com Hmm, looks like I'm late and this user-visible API change was just merged. But it's for rc1, so we can still change it, hopefully? One problem I see with the counter is that it's in bytes, but among counters that use pages, and the name doesn't indicate it. Then, I don't see why users should care about the "indirectly" part, as that's just an implementation detail. It is reclaimable and that's what matters, right? (I also wanted to complain about lack of Documentation/... update, but looks like there's no general file about vmstat, ugh) I also kind of liked the idea from v1 rfc posting that there would be a separate set of reclaimable kmalloc-X caches for these kind of allocations. Besides accounting, it should also help reduce memory fragmentation. The right variant of cache would be detected via __GFP_RECLAIMABLE. With that in mind, can we at least for now put the (manually maintained) byte counter in a variable that's not directly exposed via /proc/vmstat, and then when printing nr_slab_reclaimable, simply add the value (divided by PAGE_SIZE), and when printing nr_slab_unreclaimable, subtract the same value. This way we would be simply making the existing counters more precise, in line with their semantics. Thoughts? Vlastimil > --- > include/linux/mmzone.h | 1 + > mm/vmstat.c | 1 + > 2 files changed, 2 insertions(+) > > diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h > index e09fe563d5dc..15e783f29e21 100644 > --- a/include/linux/mmzone.h > +++ b/include/linux/mmzone.h > @@ -180,6 +180,7 @@ enum node_stat_item { > NR_VMSCAN_IMMEDIATE, /* Prioritise for reclaim when writeback ends */ > NR_DIRTIED, /* page dirtyings since bootup */ > NR_WRITTEN, /* page writings since bootup */ > + NR_INDIRECTLY_RECLAIMABLE_BYTES, /* measured in bytes */ > NR_VM_NODE_STAT_ITEMS > }; > > diff --git a/mm/vmstat.c b/mm/vmstat.c > index 40b2db6db6b1..b6b5684f31fe 100644 > --- a/mm/vmstat.c > +++ b/mm/vmstat.c > @@ -1161,6 +1161,7 @@ const char * const vmstat_text[] = { > "nr_vmscan_immediate_reclaim", > "nr_dirtied", > "nr_written", > + "nr_indirectly_reclaimable", > > /* enum writeback_stat_item counters */ > "nr_dirty_threshold", >
On Wed, Apr 11, 2018 at 03:16:08PM +0200, Vlastimil Babka wrote: > [+CC linux-api] > > On 03/05/2018 02:37 PM, Roman Gushchin wrote: > > This patch introduces a concept of indirectly reclaimable memory > > and adds the corresponding memory counter and /proc/vmstat item. > > > > Indirectly reclaimable memory is any sort of memory, used by > > the kernel (except of reclaimable slabs), which is actually > > reclaimable, i.e. will be released under memory pressure. > > > > The counter is in bytes, as it's not always possible to > > count such objects in pages. The name contains BYTES > > by analogy to NR_KERNEL_STACK_KB. > > > > Signed-off-by: Roman Gushchin <guro@fb.com> > > Cc: Andrew Morton <akpm@linux-foundation.org> > > Cc: Alexander Viro <viro@zeniv.linux.org.uk> > > Cc: Michal Hocko <mhocko@suse.com> > > Cc: Johannes Weiner <hannes@cmpxchg.org> > > Cc: linux-fsdevel@vger.kernel.org > > Cc: linux-kernel@vger.kernel.org > > Cc: linux-mm@kvack.org > > Cc: kernel-team@fb.com > > Hmm, looks like I'm late and this user-visible API change was just > merged. But it's for rc1, so we can still change it, hopefully? > > One problem I see with the counter is that it's in bytes, but among > counters that use pages, and the name doesn't indicate it. Here I just followed "nr_kernel_stack" path, which is measured in kB, but this is not mentioned in the field name. > Then, I don't > see why users should care about the "indirectly" part, as that's just an > implementation detail. It is reclaimable and that's what matters, right? > (I also wanted to complain about lack of Documentation/... update, but > looks like there's no general file about vmstat, ugh) I agree, that it's a bit weird, and it's probably better to not expose it at all; but this is how all vm counters work. We do expose them all in /proc/vmstat. A good number of them is useless until you are not a mm developer, so it's arguable more "debug info" rather than "api". It's definitely not a reason to make them messy. Does "nr_indirectly_reclaimable_bytes" look better to you? > > I also kind of liked the idea from v1 rfc posting that there would be a > separate set of reclaimable kmalloc-X caches for these kind of > allocations. Besides accounting, it should also help reduce memory > fragmentation. The right variant of cache would be detected via > __GFP_RECLAIMABLE. Well, the downside is that we have to introduce X new caches just for this particular problem. I'm not strictly against the idea, but not convinced that it's much better. > > With that in mind, can we at least for now put the (manually maintained) > byte counter in a variable that's not directly exposed via /proc/vmstat, > and then when printing nr_slab_reclaimable, simply add the value > (divided by PAGE_SIZE), and when printing nr_slab_unreclaimable, > subtract the same value. This way we would be simply making the existing > counters more precise, in line with their semantics. Idk, I don't like the idea of adding a counter outside of the vm counters infrastructure, and I definitely wouldn't touch the exposed nr_slab_reclaimable and nr_slab_unreclaimable fields. We do have some stats in /proc/slabinfo, /proc/meminfo and /sys/kernel/slab and I think that we should keep it consistent. Thanks! > > Thoughts? > Vlastimil > > > --- > > include/linux/mmzone.h | 1 + > > mm/vmstat.c | 1 + > > 2 files changed, 2 insertions(+) > > > > diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h > > index e09fe563d5dc..15e783f29e21 100644 > > --- a/include/linux/mmzone.h > > +++ b/include/linux/mmzone.h > > @@ -180,6 +180,7 @@ enum node_stat_item { > > NR_VMSCAN_IMMEDIATE, /* Prioritise for reclaim when writeback ends */ > > NR_DIRTIED, /* page dirtyings since bootup */ > > NR_WRITTEN, /* page writings since bootup */ > > + NR_INDIRECTLY_RECLAIMABLE_BYTES, /* measured in bytes */ > > NR_VM_NODE_STAT_ITEMS > > }; > > > > diff --git a/mm/vmstat.c b/mm/vmstat.c > > index 40b2db6db6b1..b6b5684f31fe 100644 > > --- a/mm/vmstat.c > > +++ b/mm/vmstat.c > > @@ -1161,6 +1161,7 @@ const char * const vmstat_text[] = { > > "nr_vmscan_immediate_reclaim", > > "nr_dirtied", > > "nr_written", > > + "nr_indirectly_reclaimable", > > > > /* enum writeback_stat_item counters */ > > "nr_dirty_threshold", > > >
On 04/11/2018 03:56 PM, Roman Gushchin wrote: > On Wed, Apr 11, 2018 at 03:16:08PM +0200, Vlastimil Babka wrote: >> [+CC linux-api] >> >> On 03/05/2018 02:37 PM, Roman Gushchin wrote: >>> This patch introduces a concept of indirectly reclaimable memory >>> and adds the corresponding memory counter and /proc/vmstat item. >>> >>> Indirectly reclaimable memory is any sort of memory, used by >>> the kernel (except of reclaimable slabs), which is actually >>> reclaimable, i.e. will be released under memory pressure. >>> >>> The counter is in bytes, as it's not always possible to >>> count such objects in pages. The name contains BYTES >>> by analogy to NR_KERNEL_STACK_KB. >>> >>> Signed-off-by: Roman Gushchin <guro@fb.com> >>> Cc: Andrew Morton <akpm@linux-foundation.org> >>> Cc: Alexander Viro <viro@zeniv.linux.org.uk> >>> Cc: Michal Hocko <mhocko@suse.com> >>> Cc: Johannes Weiner <hannes@cmpxchg.org> >>> Cc: linux-fsdevel@vger.kernel.org >>> Cc: linux-kernel@vger.kernel.org >>> Cc: linux-mm@kvack.org >>> Cc: kernel-team@fb.com >> >> Hmm, looks like I'm late and this user-visible API change was just >> merged. But it's for rc1, so we can still change it, hopefully? >> >> One problem I see with the counter is that it's in bytes, but among >> counters that use pages, and the name doesn't indicate it. > > Here I just followed "nr_kernel_stack" path, which is measured in kB, > but this is not mentioned in the field name. Oh, didn't know. Bad example to follow :P >> Then, I don't >> see why users should care about the "indirectly" part, as that's just an >> implementation detail. It is reclaimable and that's what matters, right? >> (I also wanted to complain about lack of Documentation/... update, but >> looks like there's no general file about vmstat, ugh) > > I agree, that it's a bit weird, and it's probably better to not expose > it at all; but this is how all vm counters work. We do expose them all > in /proc/vmstat. A good number of them is useless until you are not a > mm developer, so it's arguable more "debug info" rather than "api". Yeah the problem is that once tools start rely on them, they fall under the "do not break userspace" rule, however we call them. So being cautious and conservative can't hurt. > It's definitely not a reason to make them messy. > Does "nr_indirectly_reclaimable_bytes" look better to you? It still has has the "indirecly" part and feels arbitrary :/ >> >> I also kind of liked the idea from v1 rfc posting that there would be a >> separate set of reclaimable kmalloc-X caches for these kind of >> allocations. Besides accounting, it should also help reduce memory >> fragmentation. The right variant of cache would be detected via >> __GFP_RECLAIMABLE. > > Well, the downside is that we have to introduce X new caches > just for this particular problem. I'm not strictly against the idea, > but not convinced that it's much better. Maybe we can find more cases that would benefit from it. Heck, even slab itself allocates some management structures from the generic kmalloc caches, and if they are used for reclaimable caches, they could be tracked as reclaimable as well. >> >> With that in mind, can we at least for now put the (manually maintained) >> byte counter in a variable that's not directly exposed via /proc/vmstat, >> and then when printing nr_slab_reclaimable, simply add the value >> (divided by PAGE_SIZE), and when printing nr_slab_unreclaimable, >> subtract the same value. This way we would be simply making the existing >> counters more precise, in line with their semantics. > > Idk, I don't like the idea of adding a counter outside of the vm counters > infrastructure, and I definitely wouldn't touch the exposed > nr_slab_reclaimable and nr_slab_unreclaimable fields. We would be just making the reported values more precise wrt reality. > We do have some stats in /proc/slabinfo, /proc/meminfo and /sys/kernel/slab > and I think that we should keep it consistent. Right, meminfo would be adjusted the same. slabinfo doesn't indicate which caches are reclaimable, so there will be no change. /sys/kernel/slab/cache/reclaim_account does, but I doubt anything will break. > Thanks! > >> >> Thoughts? >> Vlastimil >> >>> --- >>> include/linux/mmzone.h | 1 + >>> mm/vmstat.c | 1 + >>> 2 files changed, 2 insertions(+) >>> >>> diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h >>> index e09fe563d5dc..15e783f29e21 100644 >>> --- a/include/linux/mmzone.h >>> +++ b/include/linux/mmzone.h >>> @@ -180,6 +180,7 @@ enum node_stat_item { >>> NR_VMSCAN_IMMEDIATE, /* Prioritise for reclaim when writeback ends */ >>> NR_DIRTIED, /* page dirtyings since bootup */ >>> NR_WRITTEN, /* page writings since bootup */ >>> + NR_INDIRECTLY_RECLAIMABLE_BYTES, /* measured in bytes */ >>> NR_VM_NODE_STAT_ITEMS >>> }; >>> >>> diff --git a/mm/vmstat.c b/mm/vmstat.c >>> index 40b2db6db6b1..b6b5684f31fe 100644 >>> --- a/mm/vmstat.c >>> +++ b/mm/vmstat.c >>> @@ -1161,6 +1161,7 @@ const char * const vmstat_text[] = { >>> "nr_vmscan_immediate_reclaim", >>> "nr_dirtied", >>> "nr_written", >>> + "nr_indirectly_reclaimable", >>> >>> /* enum writeback_stat_item counters */ >>> "nr_dirty_threshold", >>> >> >
On Thu 12-04-18 08:52:52, Vlastimil Babka wrote: > On 04/11/2018 03:56 PM, Roman Gushchin wrote: > > On Wed, Apr 11, 2018 at 03:16:08PM +0200, Vlastimil Babka wrote: [...] > >> With that in mind, can we at least for now put the (manually maintained) > >> byte counter in a variable that's not directly exposed via /proc/vmstat, > >> and then when printing nr_slab_reclaimable, simply add the value > >> (divided by PAGE_SIZE), and when printing nr_slab_unreclaimable, > >> subtract the same value. This way we would be simply making the existing > >> counters more precise, in line with their semantics. > > > > Idk, I don't like the idea of adding a counter outside of the vm counters > > infrastructure, and I definitely wouldn't touch the exposed > > nr_slab_reclaimable and nr_slab_unreclaimable fields. Why? > We would be just making the reported values more precise wrt reality. I was suggesting something similar in an earlier discussion. I am not really happy about the new exposed counter either. It is just arbitrary by name yet very specific for this particular usecase. What is a poor user supposed to do with the new counter? Can this be used for any calculations?
On Thu, Apr 12, 2018 at 01:52:17PM +0200, Michal Hocko wrote: > On Thu 12-04-18 08:52:52, Vlastimil Babka wrote: > > On 04/11/2018 03:56 PM, Roman Gushchin wrote: > > > On Wed, Apr 11, 2018 at 03:16:08PM +0200, Vlastimil Babka wrote: > [...] > > >> With that in mind, can we at least for now put the (manually maintained) > > >> byte counter in a variable that's not directly exposed via /proc/vmstat, > > >> and then when printing nr_slab_reclaimable, simply add the value > > >> (divided by PAGE_SIZE), and when printing nr_slab_unreclaimable, > > >> subtract the same value. This way we would be simply making the existing > > >> counters more precise, in line with their semantics. > > > > > > Idk, I don't like the idea of adding a counter outside of the vm counters > > > infrastructure, and I definitely wouldn't touch the exposed > > > nr_slab_reclaimable and nr_slab_unreclaimable fields. > > Why? Both nr_slab_reclaimable and nr_slab_unreclaimable have a very simple meaning: they are numbers of pages used by corresponding slab caches. In the answer to the very first version of this patchset Andrew suggested to generalize the idea to allow further accounting of non-kmalloc() allocations. I like the idea, even if don't have a good example right now. The problem with external names existed for many years before we've accidentally hit it, so if we don't have other examples right now, it doesn't mean that we wouldn't have them in the future. > > > We would be just making the reported values more precise wrt reality. > > I was suggesting something similar in an earlier discussion. I am not > really happy about the new exposed counter either. It is just arbitrary > by name yet very specific for this particular usecase. > > What is a poor user supposed to do with the new counter? Can this be > used for any calculations? For me the most important part is to fix the overcommit logic, because it's a real security and production issue. Adjusting MemAvailable is important too. I really open here for any concrete suggestions on how to do it without exporting of a new value, and without adding too much complexity to the code (e.g. skipping this particular mm counter on printing will be quite messy). Thanks!
On Thu 12-04-18 15:38:33, Roman Gushchin wrote: > On Thu, Apr 12, 2018 at 01:52:17PM +0200, Michal Hocko wrote: > > On Thu 12-04-18 08:52:52, Vlastimil Babka wrote: > > > On 04/11/2018 03:56 PM, Roman Gushchin wrote: > > > > On Wed, Apr 11, 2018 at 03:16:08PM +0200, Vlastimil Babka wrote: > > [...] > > > >> With that in mind, can we at least for now put the (manually maintained) > > > >> byte counter in a variable that's not directly exposed via /proc/vmstat, > > > >> and then when printing nr_slab_reclaimable, simply add the value > > > >> (divided by PAGE_SIZE), and when printing nr_slab_unreclaimable, > > > >> subtract the same value. This way we would be simply making the existing > > > >> counters more precise, in line with their semantics. > > > > > > > > Idk, I don't like the idea of adding a counter outside of the vm counters > > > > infrastructure, and I definitely wouldn't touch the exposed > > > > nr_slab_reclaimable and nr_slab_unreclaimable fields. > > > > Why? > > Both nr_slab_reclaimable and nr_slab_unreclaimable have a very simple > meaning: they are numbers of pages used by corresponding slab caches. Right, but if names are reclaimable then they should end up in the reclaimable slabs and to be accounted as such. Objects themselves are not sufficient to reclaim the accounted memory. > In the answer to the very first version of this patchset > Andrew suggested to generalize the idea to allow further > accounting of non-kmalloc() allocations. > I like the idea, even if don't have a good example right now. Well, I have to disagree here. It sounds completely ad-hoc without a reasoable semantic. Or how does it help users when they do not know what is the indirect dependency and how to trigger it. > The problem with external names existed for many years before > we've accidentally hit it, so if we don't have other examples > right now, it doesn't mean that we wouldn't have them in the future. > > > > > > We would be just making the reported values more precise wrt reality. > > > > I was suggesting something similar in an earlier discussion. I am not > > really happy about the new exposed counter either. It is just arbitrary > > by name yet very specific for this particular usecase. > > > > What is a poor user supposed to do with the new counter? Can this be > > used for any calculations? > > For me the most important part is to fix the overcommit logic, because it's > a real security and production issue. Sure, the problem is ugly. Not the first one when the unaccounted kernel allocation can eat a lot of memory. We have many other such. The usual answer was to use kmemcg accounting.
On Thu, Apr 12, 2018 at 08:52:52AM +0200, Vlastimil Babka wrote: > On 04/11/2018 03:56 PM, Roman Gushchin wrote: > > On Wed, Apr 11, 2018 at 03:16:08PM +0200, Vlastimil Babka wrote: > >> [+CC linux-api] > >> > >> On 03/05/2018 02:37 PM, Roman Gushchin wrote: > >>> This patch introduces a concept of indirectly reclaimable memory > >>> and adds the corresponding memory counter and /proc/vmstat item. > >>> > >>> Indirectly reclaimable memory is any sort of memory, used by > >>> the kernel (except of reclaimable slabs), which is actually > >>> reclaimable, i.e. will be released under memory pressure. > >>> > >>> The counter is in bytes, as it's not always possible to > >>> count such objects in pages. The name contains BYTES > >>> by analogy to NR_KERNEL_STACK_KB. > >>> > >>> Signed-off-by: Roman Gushchin <guro@fb.com> > >>> Cc: Andrew Morton <akpm@linux-foundation.org> > >>> Cc: Alexander Viro <viro@zeniv.linux.org.uk> > >>> Cc: Michal Hocko <mhocko@suse.com> > >>> Cc: Johannes Weiner <hannes@cmpxchg.org> > >>> Cc: linux-fsdevel@vger.kernel.org > >>> Cc: linux-kernel@vger.kernel.org > >>> Cc: linux-mm@kvack.org > >>> Cc: kernel-team@fb.com > >> > >> Hmm, looks like I'm late and this user-visible API change was just > >> merged. But it's for rc1, so we can still change it, hopefully? > >> > >> One problem I see with the counter is that it's in bytes, but among > >> counters that use pages, and the name doesn't indicate it. > > > > Here I just followed "nr_kernel_stack" path, which is measured in kB, > > but this is not mentioned in the field name. > > Oh, didn't know. Bad example to follow :P > > >> Then, I don't > >> see why users should care about the "indirectly" part, as that's just an > >> implementation detail. It is reclaimable and that's what matters, right? > >> (I also wanted to complain about lack of Documentation/... update, but > >> looks like there's no general file about vmstat, ugh) > > > > I agree, that it's a bit weird, and it's probably better to not expose > > it at all; but this is how all vm counters work. We do expose them all > > in /proc/vmstat. A good number of them is useless until you are not a > > mm developer, so it's arguable more "debug info" rather than "api". > > Yeah the problem is that once tools start rely on them, they fall under > the "do not break userspace" rule, however we call them. So being > cautious and conservative can't hurt. > > > It's definitely not a reason to make them messy. > > Does "nr_indirectly_reclaimable_bytes" look better to you? > > It still has has the "indirecly" part and feels arbitrary :/ > > >> > >> I also kind of liked the idea from v1 rfc posting that there would be a > >> separate set of reclaimable kmalloc-X caches for these kind of > >> allocations. Besides accounting, it should also help reduce memory > >> fragmentation. The right variant of cache would be detected via > >> __GFP_RECLAIMABLE. > > > > Well, the downside is that we have to introduce X new caches > > just for this particular problem. I'm not strictly against the idea, > > but not convinced that it's much better. > > Maybe we can find more cases that would benefit from it. Heck, even slab > itself allocates some management structures from the generic kmalloc > caches, and if they are used for reclaimable caches, they could be > tracked as reclaimable as well. This is a good catch! > > >> > >> With that in mind, can we at least for now put the (manually maintained) > >> byte counter in a variable that's not directly exposed via /proc/vmstat, > >> and then when printing nr_slab_reclaimable, simply add the value > >> (divided by PAGE_SIZE), and when printing nr_slab_unreclaimable, > >> subtract the same value. This way we would be simply making the existing > >> counters more precise, in line with their semantics. > > > > Idk, I don't like the idea of adding a counter outside of the vm counters > > infrastructure, and I definitely wouldn't touch the exposed > > nr_slab_reclaimable and nr_slab_unreclaimable fields. > > We would be just making the reported values more precise wrt reality. It depends on if we believe that only slab memory can be reclaimable or not. If yes, this is true, otherwise not. My guess is that some drivers (e.g. networking) might have buffers, which are reclaimable under mempressure, and are allocated using the page allocator. But I have to look closer... > > We do have some stats in /proc/slabinfo, /proc/meminfo and /sys/kernel/slab > > and I think that we should keep it consistent. > > Right, meminfo would be adjusted the same. slabinfo doesn't indicate > which caches are reclaimable, so there will be no change. > /sys/kernel/slab/cache/reclaim_account does, but I doubt anything will > break. It also can be found out of the corresponding directory name in sysfs: $ ls -la /sys/kernel/slab/dentr* lrwxrwxrwx. 1 root root 0 Apr 11 14:45 /sys/kernel/slab/dentry -> :aA-0000192 ^ this is the "reclaimable" flag Not saying that something will break. Thanks!
On Thu 12-04-18 15:57:03, Roman Gushchin wrote: > On Thu, Apr 12, 2018 at 08:52:52AM +0200, Vlastimil Babka wrote: [...] > > We would be just making the reported values more precise wrt reality. > > It depends on if we believe that only slab memory can be reclaimable > or not. If yes, this is true, otherwise not. > > My guess is that some drivers (e.g. networking) might have buffers, > which are reclaimable under mempressure, and are allocated using > the page allocator. But I have to look closer... Well, we have many direct page allocator users which are not accounted in vmstat. Some of those use their specific accounting (e.g. network buffers, some fs metadata a many others). In the ideal world MM layer would know about those but... Anyway, this particular case is quite clear, no? We _use_ kmalloc so this is slab allocator. We just misaccount it.
On Thu, Apr 12, 2018 at 8:27 PM, Roman Gushchin <guro@fb.com> wrote: > On Thu, Apr 12, 2018 at 08:52:52AM +0200, Vlastimil Babka wrote: >> On 04/11/2018 03:56 PM, Roman Gushchin wrote: >> > On Wed, Apr 11, 2018 at 03:16:08PM +0200, Vlastimil Babka wrote: >> >> [+CC linux-api] >> >> >> >> On 03/05/2018 02:37 PM, Roman Gushchin wrote: >> >>> This patch introduces a concept of indirectly reclaimable memory >> >>> and adds the corresponding memory counter and /proc/vmstat item. >> >>> >> >>> Indirectly reclaimable memory is any sort of memory, used by >> >>> the kernel (except of reclaimable slabs), which is actually >> >>> reclaimable, i.e. will be released under memory pressure. >> >>> >> >>> The counter is in bytes, as it's not always possible to >> >>> count such objects in pages. The name contains BYTES >> >>> by analogy to NR_KERNEL_STACK_KB. >> >>> >> >>> Signed-off-by: Roman Gushchin <guro@fb.com> >> >>> Cc: Andrew Morton <akpm@linux-foundation.org> >> >>> Cc: Alexander Viro <viro@zeniv.linux.org.uk> >> >>> Cc: Michal Hocko <mhocko@suse.com> >> >>> Cc: Johannes Weiner <hannes@cmpxchg.org> >> >>> Cc: linux-fsdevel@vger.kernel.org >> >>> Cc: linux-kernel@vger.kernel.org >> >>> Cc: linux-mm@kvack.org >> >>> Cc: kernel-team@fb.com >> >> >> >> Hmm, looks like I'm late and this user-visible API change was just >> >> merged. But it's for rc1, so we can still change it, hopefully? >> >> >> >> One problem I see with the counter is that it's in bytes, but among >> >> counters that use pages, and the name doesn't indicate it. >> > >> > Here I just followed "nr_kernel_stack" path, which is measured in kB, >> > but this is not mentioned in the field name. >> >> Oh, didn't know. Bad example to follow :P >> >> >> Then, I don't >> >> see why users should care about the "indirectly" part, as that's just an >> >> implementation detail. It is reclaimable and that's what matters, right? >> >> (I also wanted to complain about lack of Documentation/... update, but >> >> looks like there's no general file about vmstat, ugh) >> > >> > I agree, that it's a bit weird, and it's probably better to not expose >> > it at all; but this is how all vm counters work. We do expose them all >> > in /proc/vmstat. A good number of them is useless until you are not a >> > mm developer, so it's arguable more "debug info" rather than "api". >> >> Yeah the problem is that once tools start rely on them, they fall under >> the "do not break userspace" rule, however we call them. So being >> cautious and conservative can't hurt. >> >> > It's definitely not a reason to make them messy. >> > Does "nr_indirectly_reclaimable_bytes" look better to you? >> >> It still has has the "indirecly" part and feels arbitrary :/ >> >> >> >> >> I also kind of liked the idea from v1 rfc posting that there would be a >> >> separate set of reclaimable kmalloc-X caches for these kind of >> >> allocations. Besides accounting, it should also help reduce memory >> >> fragmentation. The right variant of cache would be detected via >> >> __GFP_RECLAIMABLE. >> > >> > Well, the downside is that we have to introduce X new caches >> > just for this particular problem. I'm not strictly against the idea, >> > but not convinced that it's much better. >> >> Maybe we can find more cases that would benefit from it. Heck, even slab >> itself allocates some management structures from the generic kmalloc >> caches, and if they are used for reclaimable caches, they could be >> tracked as reclaimable as well. > > This is a good catch! > >> >> >> >> >> With that in mind, can we at least for now put the (manually maintained) >> >> byte counter in a variable that's not directly exposed via /proc/vmstat, >> >> and then when printing nr_slab_reclaimable, simply add the value >> >> (divided by PAGE_SIZE), and when printing nr_slab_unreclaimable, >> >> subtract the same value. This way we would be simply making the existing >> >> counters more precise, in line with their semantics. >> > >> > Idk, I don't like the idea of adding a counter outside of the vm counters >> > infrastructure, and I definitely wouldn't touch the exposed >> > nr_slab_reclaimable and nr_slab_unreclaimable fields. >> >> We would be just making the reported values more precise wrt reality. > > It depends on if we believe that only slab memory can be reclaimable > or not. If yes, this is true, otherwise not. > > My guess is that some drivers (e.g. networking) might have buffers, > which are reclaimable under mempressure, and are allocated using > the page allocator. But I have to look closer... > One such case I have encountered is that of the ION page pool. The page pool registers a shrinker. When not in any memory pressure page pool can go high and thus cause an mmap to fail when OVERCOMMIT_GUESS is set. I can send a patch to account ION page pool pages in NR_INDIRECTLY_RECLAIMABLE_BYTES. Thanks, Vinayak
On 4/13/2018 5:43 PM, vinayak menon wrote: > On Thu, Apr 12, 2018 at 8:27 PM, Roman Gushchin <guro@fb.com> wrote: >> On Thu, Apr 12, 2018 at 08:52:52AM +0200, Vlastimil Babka wrote: >>> On 04/11/2018 03:56 PM, Roman Gushchin wrote: >>>> On Wed, Apr 11, 2018 at 03:16:08PM +0200, Vlastimil Babka wrote: >>>>> [+CC linux-api] >>>>> >>>>> On 03/05/2018 02:37 PM, Roman Gushchin wrote: >>>>>> This patch introduces a concept of indirectly reclaimable memory >>>>>> and adds the corresponding memory counter and /proc/vmstat item. >>>>>> >>>>>> Indirectly reclaimable memory is any sort of memory, used by >>>>>> the kernel (except of reclaimable slabs), which is actually >>>>>> reclaimable, i.e. will be released under memory pressure. >>>>>> >>>>>> The counter is in bytes, as it's not always possible to >>>>>> count such objects in pages. The name contains BYTES >>>>>> by analogy to NR_KERNEL_STACK_KB. >>>>>> >>>>>> Signed-off-by: Roman Gushchin <guro@fb.com> >>>>>> Cc: Andrew Morton <akpm@linux-foundation.org> >>>>>> Cc: Alexander Viro <viro@zeniv.linux.org.uk> >>>>>> Cc: Michal Hocko <mhocko@suse.com> >>>>>> Cc: Johannes Weiner <hannes@cmpxchg.org> >>>>>> Cc: linux-fsdevel@vger.kernel.org >>>>>> Cc: linux-kernel@vger.kernel.org >>>>>> Cc: linux-mm@kvack.org >>>>>> Cc: kernel-team@fb.com >>>>> >>>>> Hmm, looks like I'm late and this user-visible API change was just >>>>> merged. But it's for rc1, so we can still change it, hopefully? >>>>> >>>>> One problem I see with the counter is that it's in bytes, but among >>>>> counters that use pages, and the name doesn't indicate it. >>>> >>>> Here I just followed "nr_kernel_stack" path, which is measured in kB, >>>> but this is not mentioned in the field name. >>> >>> Oh, didn't know. Bad example to follow :P >>> >>>>> Then, I don't >>>>> see why users should care about the "indirectly" part, as that's just an >>>>> implementation detail. It is reclaimable and that's what matters, right? >>>>> (I also wanted to complain about lack of Documentation/... update, but >>>>> looks like there's no general file about vmstat, ugh) >>>> >>>> I agree, that it's a bit weird, and it's probably better to not expose >>>> it at all; but this is how all vm counters work. We do expose them all >>>> in /proc/vmstat. A good number of them is useless until you are not a >>>> mm developer, so it's arguable more "debug info" rather than "api". >>> >>> Yeah the problem is that once tools start rely on them, they fall under >>> the "do not break userspace" rule, however we call them. So being >>> cautious and conservative can't hurt. >>> >>>> It's definitely not a reason to make them messy. >>>> Does "nr_indirectly_reclaimable_bytes" look better to you? >>> >>> It still has has the "indirecly" part and feels arbitrary :/ >>> >>>>> >>>>> I also kind of liked the idea from v1 rfc posting that there would be a >>>>> separate set of reclaimable kmalloc-X caches for these kind of >>>>> allocations. Besides accounting, it should also help reduce memory >>>>> fragmentation. The right variant of cache would be detected via >>>>> __GFP_RECLAIMABLE. >>>> >>>> Well, the downside is that we have to introduce X new caches >>>> just for this particular problem. I'm not strictly against the idea, >>>> but not convinced that it's much better. >>> >>> Maybe we can find more cases that would benefit from it. Heck, even slab >>> itself allocates some management structures from the generic kmalloc >>> caches, and if they are used for reclaimable caches, they could be >>> tracked as reclaimable as well. >> >> This is a good catch! >> >>> >>>>> >>>>> With that in mind, can we at least for now put the (manually maintained) >>>>> byte counter in a variable that's not directly exposed via /proc/vmstat, >>>>> and then when printing nr_slab_reclaimable, simply add the value >>>>> (divided by PAGE_SIZE), and when printing nr_slab_unreclaimable, >>>>> subtract the same value. This way we would be simply making the existing >>>>> counters more precise, in line with their semantics. >>>> >>>> Idk, I don't like the idea of adding a counter outside of the vm counters >>>> infrastructure, and I definitely wouldn't touch the exposed >>>> nr_slab_reclaimable and nr_slab_unreclaimable fields. >>> >>> We would be just making the reported values more precise wrt reality. >> >> It depends on if we believe that only slab memory can be reclaimable >> or not. If yes, this is true, otherwise not. >> >> My guess is that some drivers (e.g. networking) might have buffers, >> which are reclaimable under mempressure, and are allocated using >> the page allocator. But I have to look closer... >> > > One such case I have encountered is that of the ION page pool. The page pool > registers a shrinker. When not in any memory pressure page pool can go high > and thus cause an mmap to fail when OVERCOMMIT_GUESS is set. I can send > a patch to account ION page pool pages in NR_INDIRECTLY_RECLAIMABLE_BYTES. > > Thanks, > Vinayak > As Vinayak mentioned NR_INDIRECTLY_RECLAIMABLE_BYTES can be used to solve the issue with ION page pool when OVERCOMMIT_GUESS is set, the patch for the same can be found here https://lkml.org/lkml/2018/4/24/1288
On Wed, Apr 25, 2018 at 09:19:29AM +0530, Vijayanand Jitta wrote: > >>>> Idk, I don't like the idea of adding a counter outside of the vm counters > >>>> infrastructure, and I definitely wouldn't touch the exposed > >>>> nr_slab_reclaimable and nr_slab_unreclaimable fields. > >>> > >>> We would be just making the reported values more precise wrt reality. > >> > >> It depends on if we believe that only slab memory can be reclaimable > >> or not. If yes, this is true, otherwise not. > >> > >> My guess is that some drivers (e.g. networking) might have buffers, > >> which are reclaimable under mempressure, and are allocated using > >> the page allocator. But I have to look closer... > >> > > > > One such case I have encountered is that of the ION page pool. The page pool > > registers a shrinker. When not in any memory pressure page pool can go high > > and thus cause an mmap to fail when OVERCOMMIT_GUESS is set. I can send > > a patch to account ION page pool pages in NR_INDIRECTLY_RECLAIMABLE_BYTES. Perfect! This is exactly what I've expected. > > > > Thanks, > > Vinayak > > > > As Vinayak mentioned NR_INDIRECTLY_RECLAIMABLE_BYTES can be used to solve the issue > with ION page pool when OVERCOMMIT_GUESS is set, the patch for the same can be > found here https://lkml.org/lkml/2018/4/24/1288 This makes perfect sense to me. Please, fell free to add: Acked-by: Roman Gushchin <guro@fb.com> Thank you!
On 04/25/2018 02:52 PM, Roman Gushchin wrote: > On Wed, Apr 25, 2018 at 09:19:29AM +0530, Vijayanand Jitta wrote: >>>>>> Idk, I don't like the idea of adding a counter outside of the vm counters >>>>>> infrastructure, and I definitely wouldn't touch the exposed >>>>>> nr_slab_reclaimable and nr_slab_unreclaimable fields. >>>>> >>>>> We would be just making the reported values more precise wrt reality. >>>> >>>> It depends on if we believe that only slab memory can be reclaimable >>>> or not. If yes, this is true, otherwise not. >>>> >>>> My guess is that some drivers (e.g. networking) might have buffers, >>>> which are reclaimable under mempressure, and are allocated using >>>> the page allocator. But I have to look closer... >>>> >>> >>> One such case I have encountered is that of the ION page pool. The page pool >>> registers a shrinker. When not in any memory pressure page pool can go high >>> and thus cause an mmap to fail when OVERCOMMIT_GUESS is set. I can send >>> a patch to account ION page pool pages in NR_INDIRECTLY_RECLAIMABLE_BYTES. FYI, we have discussed this at LSF/MM and agreed to try the kmalloc reclaimable caches idea. The existing counter could then remain for page allocator users such as ION. It's a bit weird to have it in bytes and not pages then, IMHO. What if we hid it from /proc/vmstat now so it doesn't become ABI, and later convert it to page granularity and expose it under a name such as "nr_other_reclaimable" ? Vlastimil > Perfect! > This is exactly what I've expected. > >>> >>> Thanks, >>> Vinayak >>> >> >> As Vinayak mentioned NR_INDIRECTLY_RECLAIMABLE_BYTES can be used to solve the issue >> with ION page pool when OVERCOMMIT_GUESS is set, the patch for the same can be >> found here https://lkml.org/lkml/2018/4/24/1288 > > This makes perfect sense to me. > > Please, fell free to add: > Acked-by: Roman Gushchin <guro@fb.com> > > Thank you! >
On Fri, Apr 13, 2018 at 05:43:39PM +0530, vinayak menon wrote: > One such case I have encountered is that of the ION page pool. The page pool > registers a shrinker. When not in any memory pressure page pool can go high > and thus cause an mmap to fail when OVERCOMMIT_GUESS is set. I can send > a patch to account ION page pool pages in NR_INDIRECTLY_RECLAIMABLE_BYTES. Why not just account them as NR_SLAB_RECLAIMABLE? I know it's not slab, but other than that mis-naming, it seems like it'll do the right thing.
On Wed, Apr 25, 2018 at 05:47:26PM +0200, Vlastimil Babka wrote: > On 04/25/2018 02:52 PM, Roman Gushchin wrote: > > On Wed, Apr 25, 2018 at 09:19:29AM +0530, Vijayanand Jitta wrote: > >>>>>> Idk, I don't like the idea of adding a counter outside of the vm counters > >>>>>> infrastructure, and I definitely wouldn't touch the exposed > >>>>>> nr_slab_reclaimable and nr_slab_unreclaimable fields. > >>>>> > >>>>> We would be just making the reported values more precise wrt reality. > >>>> > >>>> It depends on if we believe that only slab memory can be reclaimable > >>>> or not. If yes, this is true, otherwise not. > >>>> > >>>> My guess is that some drivers (e.g. networking) might have buffers, > >>>> which are reclaimable under mempressure, and are allocated using > >>>> the page allocator. But I have to look closer... > >>>> > >>> > >>> One such case I have encountered is that of the ION page pool. The page pool > >>> registers a shrinker. When not in any memory pressure page pool can go high > >>> and thus cause an mmap to fail when OVERCOMMIT_GUESS is set. I can send > >>> a patch to account ION page pool pages in NR_INDIRECTLY_RECLAIMABLE_BYTES. > > FYI, we have discussed this at LSF/MM and agreed to try the kmalloc > reclaimable caches idea. The existing counter could then remain for page > allocator users such as ION. It's a bit weird to have it in bytes and > not pages then, IMHO. What if we hid it from /proc/vmstat now so it > doesn't become ABI, and later convert it to page granularity and expose > it under a name such as "nr_other_reclaimable" ? I've nothing against hiding it from /proc/vmstat, as long as we keep the counter in place and the main issue resolved. Maybe it's better to add nr_reclaimable = nr_slab_reclaimable + nr_other_reclaimable, which will have a simpler meaning that nr_other_reclaimable (what is other?). Thanks!
On 04/25/2018 05:55 PM, Matthew Wilcox wrote: > On Fri, Apr 13, 2018 at 05:43:39PM +0530, vinayak menon wrote: >> One such case I have encountered is that of the ION page pool. The page pool >> registers a shrinker. When not in any memory pressure page pool can go high >> and thus cause an mmap to fail when OVERCOMMIT_GUESS is set. I can send >> a patch to account ION page pool pages in NR_INDIRECTLY_RECLAIMABLE_BYTES. > > Why not just account them as NR_SLAB_RECLAIMABLE? I know it's not slab, but > other than that mis-naming, it seems like it'll do the right thing. Hm I think it would be confusing for anyone trying to correlate the number with /proc/slabinfo - the numbers there wouldn't add up.
On 04/25/2018 06:48 PM, Roman Gushchin wrote: > On Wed, Apr 25, 2018 at 05:47:26PM +0200, Vlastimil Babka wrote: >> On 04/25/2018 02:52 PM, Roman Gushchin wrote: >>> On Wed, Apr 25, 2018 at 09:19:29AM +0530, Vijayanand Jitta wrote: >>>>>>>> Idk, I don't like the idea of adding a counter outside of the vm counters >>>>>>>> infrastructure, and I definitely wouldn't touch the exposed >>>>>>>> nr_slab_reclaimable and nr_slab_unreclaimable fields. >>>>>>> >>>>>>> We would be just making the reported values more precise wrt reality. >>>>>> >>>>>> It depends on if we believe that only slab memory can be reclaimable >>>>>> or not. If yes, this is true, otherwise not. >>>>>> >>>>>> My guess is that some drivers (e.g. networking) might have buffers, >>>>>> which are reclaimable under mempressure, and are allocated using >>>>>> the page allocator. But I have to look closer... >>>>>> >>>>> >>>>> One such case I have encountered is that of the ION page pool. The page pool >>>>> registers a shrinker. When not in any memory pressure page pool can go high >>>>> and thus cause an mmap to fail when OVERCOMMIT_GUESS is set. I can send >>>>> a patch to account ION page pool pages in NR_INDIRECTLY_RECLAIMABLE_BYTES. >> >> FYI, we have discussed this at LSF/MM and agreed to try the kmalloc >> reclaimable caches idea. The existing counter could then remain for page >> allocator users such as ION. It's a bit weird to have it in bytes and >> not pages then, IMHO. What if we hid it from /proc/vmstat now so it >> doesn't become ABI, and later convert it to page granularity and expose >> it under a name such as "nr_other_reclaimable" ? > > I've nothing against hiding it from /proc/vmstat, as long as we keep > the counter in place and the main issue resolved. Sure. > Maybe it's better to add nr_reclaimable = nr_slab_reclaimable + nr_other_reclaimable, > which will have a simpler meaning that nr_other_reclaimable (what is other?). "other" can be changed, sure. nr_reclaimable is possible if we change slab to adjust that counter as well - vmstat code doesn't support arbitrary calculations when printing. > Thanks! >
On Wed, Apr 25, 2018 at 07:02:42PM +0200, Vlastimil Babka wrote: > On 04/25/2018 06:48 PM, Roman Gushchin wrote: > > On Wed, Apr 25, 2018 at 05:47:26PM +0200, Vlastimil Babka wrote: > >> On 04/25/2018 02:52 PM, Roman Gushchin wrote: > >>> On Wed, Apr 25, 2018 at 09:19:29AM +0530, Vijayanand Jitta wrote: > >>>>>>>> Idk, I don't like the idea of adding a counter outside of the vm counters > >>>>>>>> infrastructure, and I definitely wouldn't touch the exposed > >>>>>>>> nr_slab_reclaimable and nr_slab_unreclaimable fields. > >>>>>>> > >>>>>>> We would be just making the reported values more precise wrt reality. > >>>>>> > >>>>>> It depends on if we believe that only slab memory can be reclaimable > >>>>>> or not. If yes, this is true, otherwise not. > >>>>>> > >>>>>> My guess is that some drivers (e.g. networking) might have buffers, > >>>>>> which are reclaimable under mempressure, and are allocated using > >>>>>> the page allocator. But I have to look closer... > >>>>>> > >>>>> > >>>>> One such case I have encountered is that of the ION page pool. The page pool > >>>>> registers a shrinker. When not in any memory pressure page pool can go high > >>>>> and thus cause an mmap to fail when OVERCOMMIT_GUESS is set. I can send > >>>>> a patch to account ION page pool pages in NR_INDIRECTLY_RECLAIMABLE_BYTES. > >> > >> FYI, we have discussed this at LSF/MM and agreed to try the kmalloc > >> reclaimable caches idea. The existing counter could then remain for page > >> allocator users such as ION. It's a bit weird to have it in bytes and > >> not pages then, IMHO. What if we hid it from /proc/vmstat now so it > >> doesn't become ABI, and later convert it to page granularity and expose > >> it under a name such as "nr_other_reclaimable" ? > > > > I've nothing against hiding it from /proc/vmstat, as long as we keep > > the counter in place and the main issue resolved. > > Sure. > > > Maybe it's better to add nr_reclaimable = nr_slab_reclaimable + nr_other_reclaimable, > > which will have a simpler meaning that nr_other_reclaimable (what is other?). > > "other" can be changed, sure. nr_reclaimable is possible if we change > slab to adjust that counter as well - vmstat code doesn't support > arbitrary calculations when printing. Sure, but even just hiding a value isn't that easy now. So we have to touch the vmstat printing code anyway. Thanks!
diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h index e09fe563d5dc..15e783f29e21 100644 --- a/include/linux/mmzone.h +++ b/include/linux/mmzone.h @@ -180,6 +180,7 @@ enum node_stat_item { NR_VMSCAN_IMMEDIATE, /* Prioritise for reclaim when writeback ends */ NR_DIRTIED, /* page dirtyings since bootup */ NR_WRITTEN, /* page writings since bootup */ + NR_INDIRECTLY_RECLAIMABLE_BYTES, /* measured in bytes */ NR_VM_NODE_STAT_ITEMS }; diff --git a/mm/vmstat.c b/mm/vmstat.c index 40b2db6db6b1..b6b5684f31fe 100644 --- a/mm/vmstat.c +++ b/mm/vmstat.c @@ -1161,6 +1161,7 @@ const char * const vmstat_text[] = { "nr_vmscan_immediate_reclaim", "nr_dirtied", "nr_written", + "nr_indirectly_reclaimable", /* enum writeback_stat_item counters */ "nr_dirty_threshold",
This patch introduces a concept of indirectly reclaimable memory and adds the corresponding memory counter and /proc/vmstat item. Indirectly reclaimable memory is any sort of memory, used by the kernel (except of reclaimable slabs), which is actually reclaimable, i.e. will be released under memory pressure. The counter is in bytes, as it's not always possible to count such objects in pages. The name contains BYTES by analogy to NR_KERNEL_STACK_KB. Signed-off-by: Roman Gushchin <guro@fb.com> Cc: Andrew Morton <akpm@linux-foundation.org> Cc: Alexander Viro <viro@zeniv.linux.org.uk> Cc: Michal Hocko <mhocko@suse.com> Cc: Johannes Weiner <hannes@cmpxchg.org> Cc: linux-fsdevel@vger.kernel.org Cc: linux-kernel@vger.kernel.org Cc: linux-mm@kvack.org Cc: kernel-team@fb.com --- include/linux/mmzone.h | 1 + mm/vmstat.c | 1 + 2 files changed, 2 insertions(+)