Message ID | 20240103003011.211382-1-v-songbaohua@oppo.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | zram: easy the allocation of zcomp_strm's buffers with 2 pages | expand |
On (24/01/03 13:30), Barry Song wrote: > There is no need to keep zcomp_strm's buffers contiguous physically. > And rarely, 1-order allocation can fail while buddy is seriously > fragmented. Dunno. Some of these don't sound like convincing reasons, I'm afraid. We don't allocate compression streams all the time, we do it once per-CPU. And if the system is under such a terrible memory pressure then one probably should not use zram at all, because zsmalloc needs pages for its pool. I also wonder whether Android uses HW compression, in which case we may need to have physically contig pages. Not to mention TLB shootdowns that virt contig pages add to the picture. [..] > @@ -37,7 +38,7 @@ static void zcomp_strm_free(struct zcomp_strm *zstrm) > { > if (!IS_ERR_OR_NULL(zstrm->tfm)) > crypto_free_comp(zstrm->tfm); > - free_pages((unsigned long)zstrm->buffer, 1); > + vfree(zstrm->buffer); > zstrm->tfm = NULL; > zstrm->buffer = NULL; > } > @@ -53,7 +54,7 @@ static int zcomp_strm_init(struct zcomp_strm *zstrm, struct zcomp *comp) > * allocate 2 pages. 1 for compressed data, plus 1 extra for the > * case when compressed size is larger than the original one > */ > - zstrm->buffer = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, 1); > + zstrm->buffer = vzalloc(2 * PAGE_SIZE); > if (IS_ERR_OR_NULL(zstrm->tfm) || !zstrm->buffer) { > zcomp_strm_free(zstrm); > return -ENOMEM;
On Sat, Jan 6, 2024 at 9:30 AM Sergey Senozhatsky <senozhatsky@chromium.org> wrote: > > On (24/01/03 13:30), Barry Song wrote: > > There is no need to keep zcomp_strm's buffers contiguous physically. > > And rarely, 1-order allocation can fail while buddy is seriously > > fragmented. > > Dunno. Some of these don't sound like convincing reasons, I'm afraid. > We don't allocate compression streams all the time, we do it once > per-CPU. And if the system is under such a terrible memory pressure We actually do it many times actually because we free it while unplugging and re-allocate it during hotplugging. this can happen quite often for systems like Android using hotplug for power management. > then one probably should not use zram at all, because zsmalloc needs > pages for its pool. In my humble opinion, 1-order allocation and 0-order allocation are different things, 1-order is still more difficult though it is easier than 2-order which was a big pain causing allocation latency for tasks' kernel stacks and negatively affecting user experience. it has now been replaced by vmalloc and makes life easier :-) > > I also wonder whether Android uses HW compression, in which case we > may need to have physically contig pages. Not to mention TLB shootdowns > that virt contig pages add to the picture. I don't understand how HW compression and TLB shootdown are related as zRAM is using a traditional comp API. We are always passing a virtual address, traditional HW drivers use their own buffers to do DMA. int crypto_comp_compress(struct crypto_comp *comp, const u8 *src, unsigned int slen, u8 *dst, unsigned int *dlen); int crypto_comp_decompress(struct crypto_comp *comp, const u8 *src, unsigned int slen, u8 *dst, unsigned int *dlen); In new acomp API, we are passing a sg - users' buffers to drivers directly, sg_init_one(&input, src, entry->length); sg_init_table(&output, 1); sg_set_page(&output, page, PAGE_SIZE, 0); acomp_request_set_params(acomp_ctx->req, &input, &output, entry->length, dlen); ret = crypto_wait_req(crypto_acomp_decompress(acomp_ctx->req), &acomp_ctx->wait); but i agree one-nents sg might have some advantage in scompress case after we move to new acomp APIs if we have this patch I sent recently [patch 3/3], https://lore.kernel.org/linux-mm/20240103095006.608744-1-21cnbao@gmail.com/ For the current zRAM code, I guess HW compression/TLB is not a concern. > > [..] > > @@ -37,7 +38,7 @@ static void zcomp_strm_free(struct zcomp_strm *zstrm) > > { > > if (!IS_ERR_OR_NULL(zstrm->tfm)) > > crypto_free_comp(zstrm->tfm); > > - free_pages((unsigned long)zstrm->buffer, 1); > > + vfree(zstrm->buffer); > > zstrm->tfm = NULL; > > zstrm->buffer = NULL; > > } > > @@ -53,7 +54,7 @@ static int zcomp_strm_init(struct zcomp_strm *zstrm, struct zcomp *comp) > > * allocate 2 pages. 1 for compressed data, plus 1 extra for the > > * case when compressed size is larger than the original one > > */ > > - zstrm->buffer = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, 1); > > + zstrm->buffer = vzalloc(2 * PAGE_SIZE); > > if (IS_ERR_OR_NULL(zstrm->tfm) || !zstrm->buffer) { > > zcomp_strm_free(zstrm); > > return -ENOMEM; Thanks Barry
On (24/01/06 15:38), Barry Song wrote: > On Sat, Jan 6, 2024 at 9:30 AM Sergey Senozhatsky > <senozhatsky@chromium.org> wrote: > > > > On (24/01/03 13:30), Barry Song wrote: > > > There is no need to keep zcomp_strm's buffers contiguous physically. > > > And rarely, 1-order allocation can fail while buddy is seriously > > > fragmented. > > > > Dunno. Some of these don't sound like convincing reasons, I'm afraid. > > We don't allocate compression streams all the time, we do it once > > per-CPU. And if the system is under such a terrible memory pressure > > We actually do it many times actually because we free it while unplugging and > re-allocate it during hotplugging. this can happen quite often for systems like > Android using hotplug for power management. Okay, makes sense. Do you see these problems in real life? I don't recall any reports. > > then one probably should not use zram at all, because zsmalloc needs > > pages for its pool. > > In my humble opinion, 1-order allocation and 0-order allocation are different > things, 1-order is still more difficult though it is easier than > 2-order which was > a big pain causing allocation latency for tasks' kernel stacks and negatively > affecting user experience. it has now been replaced by vmalloc and makes > life easier :-) Sure. > > I also wonder whether Android uses HW compression, in which case we > > may need to have physically contig pages. Not to mention TLB shootdowns > > that virt contig pages add to the picture. > > I don't understand how HW compression and TLB shootdown are related as zRAM > is using a traditional comp API. Oh, those are not related. TLB shootdowns are what now will be added to all compressions/decompressions, so it's sort of extra cost. HW compression (which android may be doing?) is another story. Did you run any perf tests on zram w/ and w/o the patch? > We are always passing a virtual address, traditional HW drivers use their own > buffers to do DMA. > > int crypto_comp_compress(struct crypto_comp *comp, > const u8 *src, unsigned int slen, > u8 *dst, unsigned int *dlen); > int crypto_comp_decompress(struct crypto_comp *comp, > const u8 *src, unsigned int slen, > u8 *dst, unsigned int *dlen); > > In new acomp API, we are passing a sg - users' buffers to drivers directly, > sg_init_one(&input, src, entry->length); > sg_init_table(&output, 1); > sg_set_page(&output, page, PAGE_SIZE, 0); > acomp_request_set_params(acomp_ctx->req, &input, &output, entry->length, dlen); > ret = crypto_wait_req(crypto_acomp_decompress(acomp_ctx->req), > &acomp_ctx->wait); > > but i agree one-nents sg might have some advantage in scompress case Right. > after we move > to new acomp APIs if we have this patch I sent recently [patch 3/3], > https://lore.kernel.org/linux-mm/20240103095006.608744-1-21cnbao@gmail.com/ Nice.
On Mon, Jan 15, 2024 at 10:35 AM Sergey Senozhatsky <senozhatsky@chromium.org> wrote: > > On (24/01/06 15:38), Barry Song wrote: > > On Sat, Jan 6, 2024 at 9:30 AM Sergey Senozhatsky > > <senozhatsky@chromium.org> wrote: > > > > > > On (24/01/03 13:30), Barry Song wrote: > > > > There is no need to keep zcomp_strm's buffers contiguous physically. > > > > And rarely, 1-order allocation can fail while buddy is seriously > > > > fragmented. > > > > > > Dunno. Some of these don't sound like convincing reasons, I'm afraid. > > > We don't allocate compression streams all the time, we do it once > > > per-CPU. And if the system is under such a terrible memory pressure > > > > We actually do it many times actually because we free it while unplugging and > > re-allocate it during hotplugging. this can happen quite often for systems like > > Android using hotplug for power management. > > Okay, makes sense. > Do you see these problems in real life? I don't recall any reports. i don't have problems with the current zram which supports normal pages only. but in our out-of-tree code, we have enhanced zram/zsmalloc to support large folios compression/decompression, which will make zram work much better with large anon folios/mTHP things on which Ryan Roberts is working on. I mean, a large folio with for example 16 normal pages can be saved as one object in zram. In millions of phones, we have deployed this approach and seen huge improvement on compression ratio and cpu consumption decrease. in that case, we need a larger per-cpu buffer, and have seen frequent failure on allocation. that inspired me to send this patch in advance. > > > > then one probably should not use zram at all, because zsmalloc needs > > > pages for its pool. > > > > In my humble opinion, 1-order allocation and 0-order allocation are different > > things, 1-order is still more difficult though it is easier than > > 2-order which was > > a big pain causing allocation latency for tasks' kernel stacks and negatively > > affecting user experience. it has now been replaced by vmalloc and makes > > life easier :-) > > Sure. > > > > I also wonder whether Android uses HW compression, in which case we > > > may need to have physically contig pages. Not to mention TLB shootdowns > > > that virt contig pages add to the picture. > > > > I don't understand how HW compression and TLB shootdown are related as zRAM > > is using a traditional comp API. > > Oh, those are not related. TLB shootdowns are what now will be added to > all compressions/decompressions, so it's sort of extra cost. i am sorry i still don't understand where the tlb shootdowns come from. we don't unmap this per-cpu buffers during compression and decompression, do we ? am i missing something? > HW compression (which android may be doing?) is another story. > > Did you run any perf tests on zram w/ and w/o the patch? > > > We are always passing a virtual address, traditional HW drivers use their own > > buffers to do DMA. > > > > int crypto_comp_compress(struct crypto_comp *comp, > > const u8 *src, unsigned int slen, > > u8 *dst, unsigned int *dlen); > > int crypto_comp_decompress(struct crypto_comp *comp, > > const u8 *src, unsigned int slen, > > u8 *dst, unsigned int *dlen); > > > > In new acomp API, we are passing a sg - users' buffers to drivers directly, > > sg_init_one(&input, src, entry->length); > > sg_init_table(&output, 1); > > sg_set_page(&output, page, PAGE_SIZE, 0); > > acomp_request_set_params(acomp_ctx->req, &input, &output, entry->length, dlen); > > ret = crypto_wait_req(crypto_acomp_decompress(acomp_ctx->req), > > &acomp_ctx->wait); > > > > but i agree one-nents sg might have some advantage in scompress case > > Right. > > > after we move > > to new acomp APIs if we have this patch I sent recently [patch 3/3], > > https://lore.kernel.org/linux-mm/20240103095006.608744-1-21cnbao@gmail.com/ > > Nice. Thanks Barry
On (24/01/29 10:46), Barry Song wrote: > On Mon, Jan 15, 2024 at 10:35 AM Sergey Senozhatsky > <senozhatsky@chromium.org> wrote: > > > > On (24/01/06 15:38), Barry Song wrote: > > > On Sat, Jan 6, 2024 at 9:30 AM Sergey Senozhatsky > > > <senozhatsky@chromium.org> wrote: > > > > > > > > On (24/01/03 13:30), Barry Song wrote: > > > > > There is no need to keep zcomp_strm's buffers contiguous physically. > > > > > And rarely, 1-order allocation can fail while buddy is seriously > > > > > fragmented. > > > > [..] > > Okay, makes sense. > > Do you see these problems in real life? I don't recall any reports. > > i don't have problems with the current zram which supports normal pages only. > > but in our out-of-tree code, we have enhanced zram/zsmalloc to support large > folios compression/decompression, which will make zram work much better > with large anon folios/mTHP things on which Ryan Roberts is working on. > > I mean, a large folio with for example 16 normal pages can be saved as > one object in zram. > In millions of phones, we have deployed this approach and seen huge improvement > on compression ratio and cpu consumption decrease. in that case, we > need a larger > per-cpu buffer, and have seen frequent failure on allocation. that > inspired me to send > this patch in advance. May I please ask you to resend this patch with updated commit mesasge? E.g. mention cpu offlinig/onlining, etc. [..] > > > > I also wonder whether Android uses HW compression, in which case we > > > > may need to have physically contig pages. Not to mention TLB shootdowns > > > > that virt contig pages add to the picture. > > > > > > I don't understand how HW compression and TLB shootdown are related as zRAM > > > is using a traditional comp API. > > > > Oh, those are not related. TLB shootdowns are what now will be added to > > all compressions/decompressions, so it's sort of extra cost. > > i am sorry i still don't understand where the tlb shootdowns come > from. we don't unmap > this per-cpu buffers during compression and decompression, do we ? > > am i missing something? No, I guess you are right.
On Tue, Feb 6, 2024 at 9:55 AM Sergey Senozhatsky <senozhatsky@chromium.org> wrote: > > On (24/01/29 10:46), Barry Song wrote: > > On Mon, Jan 15, 2024 at 10:35 AM Sergey Senozhatsky > > <senozhatsky@chromium.org> wrote: > > > > > > On (24/01/06 15:38), Barry Song wrote: > > > > On Sat, Jan 6, 2024 at 9:30 AM Sergey Senozhatsky > > > > <senozhatsky@chromium.org> wrote: > > > > > > > > > > On (24/01/03 13:30), Barry Song wrote: > > > > > > There is no need to keep zcomp_strm's buffers contiguous physically. > > > > > > And rarely, 1-order allocation can fail while buddy is seriously > > > > > > fragmented. > > > > > > [..] > > > Okay, makes sense. > > > Do you see these problems in real life? I don't recall any reports. > > > > i don't have problems with the current zram which supports normal pages only. > > > > but in our out-of-tree code, we have enhanced zram/zsmalloc to support large > > folios compression/decompression, which will make zram work much better > > with large anon folios/mTHP things on which Ryan Roberts is working on. > > > > I mean, a large folio with for example 16 normal pages can be saved as > > one object in zram. > > In millions of phones, we have deployed this approach and seen huge improvement > > on compression ratio and cpu consumption decrease. in that case, we > > need a larger > > per-cpu buffer, and have seen frequent failure on allocation. that > > inspired me to send > > this patch in advance. > > May I please ask you to resend this patch with updated commit mesasge? > E.g. mention cpu offlinig/onlining, etc. I will send v2 to add this information in the commit message. thanks! > > [..] > > > > > I also wonder whether Android uses HW compression, in which case we > > > > > may need to have physically contig pages. Not to mention TLB shootdowns > > > > > that virt contig pages add to the picture. > > > > > > > > I don't understand how HW compression and TLB shootdown are related as zRAM > > > > is using a traditional comp API. > > > > > > Oh, those are not related. TLB shootdowns are what now will be added to > > > all compressions/decompressions, so it's sort of extra cost. > > > > i am sorry i still don't understand where the tlb shootdowns come > > from. we don't unmap > > this per-cpu buffers during compression and decompression, do we ? > > > > am i missing something? > > No, I guess you are right. Best regards Barry
diff --git a/drivers/block/zram/zcomp.c b/drivers/block/zram/zcomp.c index 55af4efd7983..8237b08c49d8 100644 --- a/drivers/block/zram/zcomp.c +++ b/drivers/block/zram/zcomp.c @@ -11,6 +11,7 @@ #include <linux/sched.h> #include <linux/cpu.h> #include <linux/crypto.h> +#include <linux/vmalloc.h> #include "zcomp.h" @@ -37,7 +38,7 @@ static void zcomp_strm_free(struct zcomp_strm *zstrm) { if (!IS_ERR_OR_NULL(zstrm->tfm)) crypto_free_comp(zstrm->tfm); - free_pages((unsigned long)zstrm->buffer, 1); + vfree(zstrm->buffer); zstrm->tfm = NULL; zstrm->buffer = NULL; } @@ -53,7 +54,7 @@ static int zcomp_strm_init(struct zcomp_strm *zstrm, struct zcomp *comp) * allocate 2 pages. 1 for compressed data, plus 1 extra for the * case when compressed size is larger than the original one */ - zstrm->buffer = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, 1); + zstrm->buffer = vzalloc(2 * PAGE_SIZE); if (IS_ERR_OR_NULL(zstrm->tfm) || !zstrm->buffer) { zcomp_strm_free(zstrm); return -ENOMEM;
There is no need to keep zcomp_strm's buffers contiguous physically. And rarely, 1-order allocation can fail while buddy is seriously fragmented. Signed-off-by: Barry Song <v-songbaohua@oppo.com> --- drivers/block/zram/zcomp.c | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-)