diff mbox series

[v6] module: replace module_layout with module_memory

Message ID 20230201173654.2475531-1-song@kernel.org (mailing list archive)
State New, archived
Headers show
Series [v6] module: replace module_layout with module_memory | expand

Commit Message

Song Liu Feb. 1, 2023, 5:36 p.m. UTC
module_layout manages different types of memory (text, data, rodata, etc.)
in one allocation, which is problematic for some reasons:

1. It is hard to enable CONFIG_STRICT_MODULE_RWX.
2. It is hard to use huge pages in modules (and not break strict rwx).
3. Many archs uses module_layout for arch-specific data, but it is not
   obvious how these data are used (are they RO, RX, or RW?)

Improve the scenario by replacing 2 (or 3) module_layout per module with
up to 7 module_memory per module:

        MOD_TEXT,
        MOD_DATA,
        MOD_RODATA,
        MOD_RO_AFTER_INIT,
        MOD_INIT_TEXT,
        MOD_INIT_DATA,
        MOD_INIT_RODATA,

and allocating them separately. This adds slightly more entries to
mod_tree (from up to 3 entries per module, to up to 7 entries per
module). However, this at most adds a small constant overhead to
__module_address(), which is expected to be fast.

Various archs use module_layout for different data. These data are put
into different module_memory based on their location in module_layout.
IOW, data that used to go with text is allocated with MOD_MEM_TYPE_TEXT;
data that used to go with data is allocated with MOD_MEM_TYPE_DATA, etc.

module_memory simplifies quite some of the module code. For example,
ARCH_WANTS_MODULES_DATA_IN_VMALLOC is a lot cleaner, as it just uses a
different allocator for the data. kernel/module/strict_rwx.c is also
much cleaner with module_memory.

Signed-off-by: Song Liu <song@kernel.org>
Cc: Luis Chamberlain <mcgrof@kernel.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Guenter Roeck <linux@roeck-us.net>
Cc: Christophe Leroy <christophe.leroy@csgroup.eu>

---

This is the preparation work for the type aware module_alloc() discussed
in [1]. While this work is not covered much in the discussion, it is a
critical step of the effort.

As this part grows pretty big (~1000 lines, + and -), I would like get
some feedback on it, so that I know it is on the right track.

Please share your comments. Thanks!

Test coverage: Tested on x86_64.
Build tested by kernel test bot in [2]. The only regression in [2] was a
typo in parisc, which is also fixed.

[1] https://lore.kernel.org/linux-mm/20221107223921.3451913-1-song@kernel.org/T/#u
[2] https://lore.kernel.org/linux-raid/63b8827e.clJQX2wg+I+tiX7m%25lkp@intel.com/T/#u

Changes v1 => v2:
1. Add data_addr_[min|max] for CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
   case.

Changes v2 => v3:
1. Fix and remove the last use of INIT_OFFSET_MASK.
2. Add more information in the commit log. (Luis Chamberlain)
3. Rebase and fix issues in x86/calltrunks.
4. Minor cleanup.

Changes v3 => v4:
1. Shorten enum/variable names, so that the code are easier to read.
   (Christophe Leroy)
2. Remove an used variable. (Guenter Roeck, Christophe Leroy)

Changes v4 => v5:
1. Simplify some code some code. (Peter Zijlstra, Christophe Leroy)
2. Remove module_check_misalignment(), which is not useful any more.

Changes v5 => v6:
1. Improve mod_mem_type_is_* and for_*mod_mem_type marcos.
   (Peter Zijlstra).
---
 arch/alpha/kernel/module.c      |   2 +-
 arch/arc/kernel/unwind.c        |   9 +-
 arch/arm/kernel/module-plts.c   |   2 +-
 arch/arm64/kernel/module-plts.c |   2 +-
 arch/ia64/kernel/module.c       |  24 ++-
 arch/mips/kernel/vpe.c          |  11 +-
 arch/parisc/kernel/module.c     |  51 ++---
 arch/powerpc/kernel/module_32.c |   7 +-
 arch/s390/kernel/module.c       |  26 +--
 arch/x86/kernel/callthunks.c    |   4 +-
 arch/x86/kernel/module.c        |   4 +-
 include/linux/module.h          |  86 +++++---
 kernel/module/internal.h        |  40 ++--
 kernel/module/kallsyms.c        |  56 ++---
 kernel/module/kdb.c             |  17 +-
 kernel/module/main.c            | 355 ++++++++++++++++++--------------
 kernel/module/procfs.c          |  16 +-
 kernel/module/strict_rwx.c      | 100 ++-------
 kernel/module/tree_lookup.c     |  39 ++--
 19 files changed, 432 insertions(+), 419 deletions(-)

Comments

Christophe Leroy Feb. 1, 2023, 6:34 p.m. UTC | #1
Hi,

On what tree / branch does this patch applies ?

I have tried with modules/modules-next, linus/master, with v6.1. It 
conflicts on all of them.

'git am -3' even says that 'It does not apply to blobs recorded in its 
index.'

Christophe

Le 01/02/2023 à 18:36, Song Liu a écrit :
> module_layout manages different types of memory (text, data, rodata, etc.)
> in one allocation, which is problematic for some reasons:
> 
> 1. It is hard to enable CONFIG_STRICT_MODULE_RWX.
> 2. It is hard to use huge pages in modules (and not break strict rwx).
> 3. Many archs uses module_layout for arch-specific data, but it is not
>     obvious how these data are used (are they RO, RX, or RW?)
> 
> Improve the scenario by replacing 2 (or 3) module_layout per module with
> up to 7 module_memory per module:
> 
>          MOD_TEXT,
>          MOD_DATA,
>          MOD_RODATA,
>          MOD_RO_AFTER_INIT,
>          MOD_INIT_TEXT,
>          MOD_INIT_DATA,
>          MOD_INIT_RODATA,
> 
> and allocating them separately. This adds slightly more entries to
> mod_tree (from up to 3 entries per module, to up to 7 entries per
> module). However, this at most adds a small constant overhead to
> __module_address(), which is expected to be fast.
> 
> Various archs use module_layout for different data. These data are put
> into different module_memory based on their location in module_layout.
> IOW, data that used to go with text is allocated with MOD_MEM_TYPE_TEXT;
> data that used to go with data is allocated with MOD_MEM_TYPE_DATA, etc.
> 
> module_memory simplifies quite some of the module code. For example,
> ARCH_WANTS_MODULES_DATA_IN_VMALLOC is a lot cleaner, as it just uses a
> different allocator for the data. kernel/module/strict_rwx.c is also
> much cleaner with module_memory.
> 
> Signed-off-by: Song Liu <song@kernel.org>
> Cc: Luis Chamberlain <mcgrof@kernel.org>
> Cc: Thomas Gleixner <tglx@linutronix.de>
> Cc: Peter Zijlstra <peterz@infradead.org>
> Cc: Guenter Roeck <linux@roeck-us.net>
> Cc: Christophe Leroy <christophe.leroy@csgroup.eu>
> 
> ---
> 
> This is the preparation work for the type aware module_alloc() discussed
> in [1]. While this work is not covered much in the discussion, it is a
> critical step of the effort.
> 
> As this part grows pretty big (~1000 lines, + and -), I would like get
> some feedback on it, so that I know it is on the right track.
> 
> Please share your comments. Thanks!
> 
> Test coverage: Tested on x86_64.
> Build tested by kernel test bot in [2]. The only regression in [2] was a
> typo in parisc, which is also fixed.
> 
> [1] https://lore.kernel.org/linux-mm/20221107223921.3451913-1-song@kernel.org/T/#u
> [2] https://lore.kernel.org/linux-raid/63b8827e.clJQX2wg+I+tiX7m%25lkp@intel.com/T/#u
> 
> Changes v1 => v2:
> 1. Add data_addr_[min|max] for CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
>     case.
> 
> Changes v2 => v3:
> 1. Fix and remove the last use of INIT_OFFSET_MASK.
> 2. Add more information in the commit log. (Luis Chamberlain)
> 3. Rebase and fix issues in x86/calltrunks.
> 4. Minor cleanup.
> 
> Changes v3 => v4:
> 1. Shorten enum/variable names, so that the code are easier to read.
>     (Christophe Leroy)
> 2. Remove an used variable. (Guenter Roeck, Christophe Leroy)
> 
> Changes v4 => v5:
> 1. Simplify some code some code. (Peter Zijlstra, Christophe Leroy)
> 2. Remove module_check_misalignment(), which is not useful any more.
> 
> Changes v5 => v6:
> 1. Improve mod_mem_type_is_* and for_*mod_mem_type marcos.
>     (Peter Zijlstra).
> ---
>   arch/alpha/kernel/module.c      |   2 +-
>   arch/arc/kernel/unwind.c        |   9 +-
>   arch/arm/kernel/module-plts.c   |   2 +-
>   arch/arm64/kernel/module-plts.c |   2 +-
>   arch/ia64/kernel/module.c       |  24 ++-
>   arch/mips/kernel/vpe.c          |  11 +-
>   arch/parisc/kernel/module.c     |  51 ++---
>   arch/powerpc/kernel/module_32.c |   7 +-
>   arch/s390/kernel/module.c       |  26 +--
>   arch/x86/kernel/callthunks.c    |   4 +-
>   arch/x86/kernel/module.c        |   4 +-
>   include/linux/module.h          |  86 +++++---
>   kernel/module/internal.h        |  40 ++--
>   kernel/module/kallsyms.c        |  56 ++---
>   kernel/module/kdb.c             |  17 +-
>   kernel/module/main.c            | 355 ++++++++++++++++++--------------
>   kernel/module/procfs.c          |  16 +-
>   kernel/module/strict_rwx.c      | 100 ++-------
>   kernel/module/tree_lookup.c     |  39 ++--
>   19 files changed, 432 insertions(+), 419 deletions(-)
> 
> diff --git a/arch/alpha/kernel/module.c b/arch/alpha/kernel/module.c
> index 5b60c248de9e..9109213abc09 100644
> --- a/arch/alpha/kernel/module.c
> +++ b/arch/alpha/kernel/module.c
> @@ -148,7 +148,7 @@ apply_relocate_add(Elf64_Shdr *sechdrs, const char *strtab,
>   
>   	/* The small sections were sorted to the end of the segment.
>   	   The following should definitely cover them.  */
> -	gp = (u64)me->core_layout.base + me->core_layout.size - 0x8000;
> +	gp = (u64)me->mem[MOD_DATA].base + me->mem[MOD_DATA].size - 0x8000;
>   	got = sechdrs[me->arch.gotsecindex].sh_addr;
>   
>   	for (i = 0; i < n; i++) {
> diff --git a/arch/arc/kernel/unwind.c b/arch/arc/kernel/unwind.c
> index 200270a94558..933451f4494f 100644
> --- a/arch/arc/kernel/unwind.c
> +++ b/arch/arc/kernel/unwind.c
> @@ -369,6 +369,8 @@ void *unwind_add_table(struct module *module, const void *table_start,
>   		       unsigned long table_size)
>   {
>   	struct unwind_table *table;
> +	struct module_memory *mod_mem_core_text;
> +	struct module_memory *mod_mem_init_text;
>   
>   	if (table_size <= 0)
>   		return NULL;
> @@ -377,9 +379,12 @@ void *unwind_add_table(struct module *module, const void *table_start,
>   	if (!table)
>   		return NULL;
>   
> +	mod_mem_core_text = &module->mem[MOD_TEXT];
> +	mod_mem_init_text = &module->mem[MOD_INIT_TEXT];
> +
>   	init_unwind_table(table, module->name,
> -			  module->core_layout.base, module->core_layout.size,
> -			  module->init_layout.base, module->init_layout.size,
> +			  mod_mem_core_text->base, mod_mem_core_text->size,
> +			  mod_mem_init_text->base, mod_mem_init_text->size,
>   			  table_start, table_size,
>   			  NULL, 0);
>   
> diff --git a/arch/arm/kernel/module-plts.c b/arch/arm/kernel/module-plts.c
> index af7c322ebed6..9d4ecb6b1412 100644
> --- a/arch/arm/kernel/module-plts.c
> +++ b/arch/arm/kernel/module-plts.c
> @@ -30,7 +30,7 @@ static const u32 fixed_plts[] = {
>   
>   static bool in_init(const struct module *mod, unsigned long loc)
>   {
> -	return loc - (u32)mod->init_layout.base < mod->init_layout.size;
> +	return within_module_init(loc, mod);
>   }
>   
>   static void prealloc_fixed(struct mod_plt_sec *pltsec, struct plt_entries *plt)
> diff --git a/arch/arm64/kernel/module-plts.c b/arch/arm64/kernel/module-plts.c
> index 5a0a8f552a61..fedffa83a283 100644
> --- a/arch/arm64/kernel/module-plts.c
> +++ b/arch/arm64/kernel/module-plts.c
> @@ -67,7 +67,7 @@ static bool plt_entries_equal(const struct plt_entry *a,
>   
>   static bool in_init(const struct module *mod, void *loc)
>   {
> -	return (u64)loc - (u64)mod->init_layout.base < mod->init_layout.size;
> +	return within_module_init(loc, mod);
>   }
>   
>   u64 module_emit_plt_entry(struct module *mod, Elf64_Shdr *sechdrs,
> diff --git a/arch/ia64/kernel/module.c b/arch/ia64/kernel/module.c
> index 8f62cf97f691..3661135da9d9 100644
> --- a/arch/ia64/kernel/module.c
> +++ b/arch/ia64/kernel/module.c
> @@ -485,19 +485,19 @@ module_frob_arch_sections (Elf_Ehdr *ehdr, Elf_Shdr *sechdrs, char *secstrings,
>   	return 0;
>   }
>   
> -static inline int
> +static inline bool
>   in_init (const struct module *mod, uint64_t addr)
>   {
> -	return addr - (uint64_t) mod->init_layout.base < mod->init_layout.size;
> +	return within_module_init(addr, mod);
>   }
>   
> -static inline int
> +static inline bool
>   in_core (const struct module *mod, uint64_t addr)
>   {
> -	return addr - (uint64_t) mod->core_layout.base < mod->core_layout.size;
> +	return within_module_core(addr, mod);
>   }
>   
> -static inline int
> +static inline bool
>   is_internal (const struct module *mod, uint64_t value)
>   {
>   	return in_init(mod, value) || in_core(mod, value);
> @@ -677,7 +677,8 @@ do_reloc (struct module *mod, uint8_t r_type, Elf64_Sym *sym, uint64_t addend,
>   		break;
>   
>   	      case RV_BDREL:
> -		val -= (uint64_t) (in_init(mod, val) ? mod->init_layout.base : mod->core_layout.base);
> +		val -= (uint64_t) (in_init(mod, val) ? mod->mem[MOD_INIT_TEXT].base :
> +				   mod->mem[MOD_TEXT].base);
>   		break;
>   
>   	      case RV_LTV:
> @@ -812,15 +813,18 @@ apply_relocate_add (Elf64_Shdr *sechdrs, const char *strtab, unsigned int symind
>   		 *     addresses have been selected...
>   		 */
>   		uint64_t gp;
> -		if (mod->core_layout.size > MAX_LTOFF)
> +		struct module_memory *mod_mem;
> +
> +		mod_mem = &mod->mem[MOD_DATA];
> +		if (mod_mem->size > MAX_LTOFF)
>   			/*
>   			 * This takes advantage of fact that SHF_ARCH_SMALL gets allocated
>   			 * at the end of the module.
>   			 */
> -			gp = mod->core_layout.size - MAX_LTOFF / 2;
> +			gp = mod_mem->size - MAX_LTOFF / 2;
>   		else
> -			gp = mod->core_layout.size / 2;
> -		gp = (uint64_t) mod->core_layout.base + ((gp + 7) & -8);
> +			gp = mod_mem->size / 2;
> +		gp = (uint64_t) mod_mem->base + ((gp + 7) & -8);
>   		mod->arch.gp = gp;
>   		DEBUGP("%s: placing gp at 0x%lx\n", __func__, gp);
>   	}
> diff --git a/arch/mips/kernel/vpe.c b/arch/mips/kernel/vpe.c
> index 13294972707b..ab114c686f9d 100644
> --- a/arch/mips/kernel/vpe.c
> +++ b/arch/mips/kernel/vpe.c
> @@ -199,18 +199,17 @@ static void layout_sections(struct module *mod, const Elf_Ehdr *hdr,
>   	for (m = 0; m < ARRAY_SIZE(masks); ++m) {
>   		for (i = 0; i < hdr->e_shnum; ++i) {
>   			Elf_Shdr *s = &sechdrs[i];
> +			struct module_memory *mod_mem;
> +
> +			mod_mem = &mod->mem[MOD_TEXT];
>   
>   			if ((s->sh_flags & masks[m][0]) != masks[m][0]
>   			    || (s->sh_flags & masks[m][1])
>   			    || s->sh_entsize != ~0UL)
>   				continue;
>   			s->sh_entsize =
> -				get_offset((unsigned long *)&mod->core_layout.size, s);
> +				get_offset((unsigned long *)&mod_mem->size, s);
>   		}
> -
> -		if (m == 0)
> -			mod->core_layout.text_size = mod->core_layout.size;
> -
>   	}
>   }
>   
> @@ -641,7 +640,7 @@ static int vpe_elfload(struct vpe *v)
>   		layout_sections(&mod, hdr, sechdrs, secstrings);
>   	}
>   
> -	v->load_addr = alloc_progmem(mod.core_layout.size);
> +	v->load_addr = alloc_progmem(mod.mod_mem[MOD_TEXT].size);
>   	if (!v->load_addr)
>   		return -ENOMEM;
>   
> diff --git a/arch/parisc/kernel/module.c b/arch/parisc/kernel/module.c
> index 7df140545b22..f6e38c4d3904 100644
> --- a/arch/parisc/kernel/module.c
> +++ b/arch/parisc/kernel/module.c
> @@ -27,9 +27,9 @@
>    *      We are not doing SEGREL32 handling correctly. According to the ABI, we
>    *      should do a value offset, like this:
>    *			if (in_init(me, (void *)val))
> - *				val -= (uint32_t)me->init_layout.base;
> + *				val -= (uint32_t)me->mem[MOD_INIT_TEXT].base;
>    *			else
> - *				val -= (uint32_t)me->core_layout.base;
> + *				val -= (uint32_t)me->mem[MOD_TEXT].base;
>    *	However, SEGREL32 is used only for PARISC unwind entries, and we want
>    *	those entries to have an absolute address, and not just an offset.
>    *
> @@ -76,25 +76,6 @@
>    * allows us to allocate up to 4095 GOT entries. */
>   #define MAX_GOTS	4095
>   
> -/* three functions to determine where in the module core
> - * or init pieces the location is */
> -static inline int in_init(struct module *me, void *loc)
> -{
> -	return (loc >= me->init_layout.base &&
> -		loc <= (me->init_layout.base + me->init_layout.size));
> -}
> -
> -static inline int in_core(struct module *me, void *loc)
> -{
> -	return (loc >= me->core_layout.base &&
> -		loc <= (me->core_layout.base + me->core_layout.size));
> -}
> -
> -static inline int in_local(struct module *me, void *loc)
> -{
> -	return in_init(me, loc) || in_core(me, loc);
> -}
> -
>   #ifndef CONFIG_64BIT
>   struct got_entry {
>   	Elf32_Addr addr;
> @@ -302,6 +283,7 @@ int module_frob_arch_sections(CONST Elf_Ehdr *hdr,
>   {
>   	unsigned long gots = 0, fdescs = 0, len;
>   	unsigned int i;
> +	struct module_memory *mod_mem;
>   
>   	len = hdr->e_shnum * sizeof(me->arch.section[0]);
>   	me->arch.section = kzalloc(len, GFP_KERNEL);
> @@ -346,14 +328,15 @@ int module_frob_arch_sections(CONST Elf_Ehdr *hdr,
>   		me->arch.section[s].stub_entries += count;
>   	}
>   
> +	mod_mem = &me->mem[MOD_TEXT];
>   	/* align things a bit */
> -	me->core_layout.size = ALIGN(me->core_layout.size, 16);
> -	me->arch.got_offset = me->core_layout.size;
> -	me->core_layout.size += gots * sizeof(struct got_entry);
> +	mod_mem->size = ALIGN(mod_mem->size, 16);
> +	me->arch.got_offset = mod_mem->size;
> +	mod_mem->size += gots * sizeof(struct got_entry);
>   
> -	me->core_layout.size = ALIGN(me->core_layout.size, 16);
> -	me->arch.fdesc_offset = me->core_layout.size;
> -	me->core_layout.size += fdescs * sizeof(Elf_Fdesc);
> +	mod_mem->size = ALIGN(mod_mem->size, 16);
> +	me->arch.fdesc_offset = mod_mem->size;
> +	mod_mem->size += fdescs * sizeof(Elf_Fdesc);
>   
>   	me->arch.got_max = gots;
>   	me->arch.fdesc_max = fdescs;
> @@ -371,7 +354,7 @@ static Elf64_Word get_got(struct module *me, unsigned long value, long addend)
>   
>   	BUG_ON(value == 0);
>   
> -	got = me->core_layout.base + me->arch.got_offset;
> +	got = me->mem[MOD_TEXT].base + me->arch.got_offset;
>   	for (i = 0; got[i].addr; i++)
>   		if (got[i].addr == value)
>   			goto out;
> @@ -389,7 +372,7 @@ static Elf64_Word get_got(struct module *me, unsigned long value, long addend)
>   #ifdef CONFIG_64BIT
>   static Elf_Addr get_fdesc(struct module *me, unsigned long value)
>   {
> -	Elf_Fdesc *fdesc = me->core_layout.base + me->arch.fdesc_offset;
> +	Elf_Fdesc *fdesc = me->mem[MOD_TEXT].base + me->arch.fdesc_offset;
>   
>   	if (!value) {
>   		printk(KERN_ERR "%s: zero OPD requested!\n", me->name);
> @@ -407,7 +390,7 @@ static Elf_Addr get_fdesc(struct module *me, unsigned long value)
>   
>   	/* Create new one */
>   	fdesc->addr = value;
> -	fdesc->gp = (Elf_Addr)me->core_layout.base + me->arch.got_offset;
> +	fdesc->gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
>   	return (Elf_Addr)fdesc;
>   }
>   #endif /* CONFIG_64BIT */
> @@ -742,7 +725,7 @@ int apply_relocate_add(Elf_Shdr *sechdrs,
>   			       loc, val);
>   			val += addend;
>   			/* can we reach it locally? */
> -			if (in_local(me, (void *)val)) {
> +			if (within_module(val, me)) {
>   				/* this is the case where the symbol is local
>   				 * to the module, but in a different section,
>   				 * so stub the jump in case it's more than 22
> @@ -801,7 +784,7 @@ int apply_relocate_add(Elf_Shdr *sechdrs,
>   			break;
>   		case R_PARISC_FPTR64:
>   			/* 64-bit function address */
> -			if(in_local(me, (void *)(val + addend))) {
> +			if (within_module(val + addend, me)) {
>   				*loc64 = get_fdesc(me, val+addend);
>   				pr_debug("FDESC for %s at %llx points to %llx\n",
>   				       strtab + sym->st_name, *loc64,
> @@ -839,7 +822,7 @@ register_unwind_table(struct module *me,
>   
>   	table = (unsigned char *)sechdrs[me->arch.unwind_section].sh_addr;
>   	end = table + sechdrs[me->arch.unwind_section].sh_size;
> -	gp = (Elf_Addr)me->core_layout.base + me->arch.got_offset;
> +	gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
>   
>   	pr_debug("register_unwind_table(), sect = %d at 0x%p - 0x%p (gp=0x%lx)\n",
>   	       me->arch.unwind_section, table, end, gp);
> @@ -977,7 +960,7 @@ void module_arch_cleanup(struct module *mod)
>   #ifdef CONFIG_64BIT
>   void *dereference_module_function_descriptor(struct module *mod, void *ptr)
>   {
> -	unsigned long start_opd = (Elf64_Addr)mod->core_layout.base +
> +	unsigned long start_opd = (Elf64_Addr)mod->mem[MOD_TEXT].base +
>   				   mod->arch.fdesc_offset;
>   	unsigned long end_opd = start_opd +
>   				mod->arch.fdesc_count * sizeof(Elf64_Fdesc);
> diff --git a/arch/powerpc/kernel/module_32.c b/arch/powerpc/kernel/module_32.c
> index ea6536171778..2729179e9dc0 100644
> --- a/arch/powerpc/kernel/module_32.c
> +++ b/arch/powerpc/kernel/module_32.c
> @@ -163,8 +163,7 @@ static uint32_t do_plt_call(void *location,
>   
>   	pr_debug("Doing plt for call to 0x%x at 0x%x\n", val, (unsigned int)location);
>   	/* Init, or core PLT? */
> -	if (location >= mod->core_layout.base
> -	    && location < mod->core_layout.base + mod->core_layout.size)
> +	if (within_module_init(location, mod))
>   		entry = (void *)sechdrs[mod->arch.core_plt_section].sh_addr;
>   	else
>   		entry = (void *)sechdrs[mod->arch.init_plt_section].sh_addr;
> @@ -322,14 +321,14 @@ notrace int module_trampoline_target(struct module *mod, unsigned long addr,
>   
>   int module_finalize_ftrace(struct module *module, const Elf_Shdr *sechdrs)
>   {
> -	module->arch.tramp = do_plt_call(module->core_layout.base,
> +	module->arch.tramp = do_plt_call(module->mem[MOD_TEXT].base,
>   					 (unsigned long)ftrace_caller,
>   					 sechdrs, module);
>   	if (!module->arch.tramp)
>   		return -ENOENT;
>   
>   #ifdef CONFIG_DYNAMIC_FTRACE_WITH_REGS
> -	module->arch.tramp_regs = do_plt_call(module->core_layout.base,
> +	module->arch.tramp_regs = do_plt_call(module->mem[MOD_TEXT].base,
>   					      (unsigned long)ftrace_regs_caller,
>   					      sechdrs, module);
>   	if (!module->arch.tramp_regs)
> diff --git a/arch/s390/kernel/module.c b/arch/s390/kernel/module.c
> index 2d159b32885b..3ed455b785ad 100644
> --- a/arch/s390/kernel/module.c
> +++ b/arch/s390/kernel/module.c
> @@ -126,6 +126,7 @@ int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
>   	Elf_Rela *rela;
>   	char *strings;
>   	int nrela, i, j;
> +	struct module_memory *mod_mem;
>   
>   	/* Find symbol table and string table. */
>   	symtab = NULL;
> @@ -173,14 +174,15 @@ int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
>   
>   	/* Increase core size by size of got & plt and set start
>   	   offsets for got and plt. */
> -	me->core_layout.size = ALIGN(me->core_layout.size, 4);
> -	me->arch.got_offset = me->core_layout.size;
> -	me->core_layout.size += me->arch.got_size;
> -	me->arch.plt_offset = me->core_layout.size;
> +	mod_mem = &me->mem[MOD_TEXT];
> +	mod_mem->size = ALIGN(mod_mem->size, 4);
> +	me->arch.got_offset = mod_mem->size;
> +	mod_mem->size += me->arch.got_size;
> +	me->arch.plt_offset = mod_mem->size;
>   	if (me->arch.plt_size) {
>   		if (IS_ENABLED(CONFIG_EXPOLINE) && !nospec_disable)
>   			me->arch.plt_size += PLT_ENTRY_SIZE;
> -		me->core_layout.size += me->arch.plt_size;
> +		mod_mem->size += me->arch.plt_size;
>   	}
>   	return 0;
>   }
> @@ -304,7 +306,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
>   	case R_390_GOTPLT64:	/* 64 bit offset to jump slot.	*/
>   	case R_390_GOTPLTENT:	/* 32 bit rel. offset to jump slot >> 1. */
>   		if (info->got_initialized == 0) {
> -			Elf_Addr *gotent = me->core_layout.base +
> +			Elf_Addr *gotent = me->mem[MOD_TEXT].base +
>   					   me->arch.got_offset +
>   					   info->got_offset;
>   
> @@ -329,7 +331,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
>   			rc = apply_rela_bits(loc, val, 0, 64, 0, write);
>   		else if (r_type == R_390_GOTENT ||
>   			 r_type == R_390_GOTPLTENT) {
> -			val += (Elf_Addr) me->core_layout.base - loc;
> +			val += (Elf_Addr) me->mem[MOD_TEXT].base - loc;
>   			rc = apply_rela_bits(loc, val, 1, 32, 1, write);
>   		}
>   		break;
> @@ -345,7 +347,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
>   			char *plt_base;
>   			char *ip;
>   
> -			plt_base = me->core_layout.base + me->arch.plt_offset;
> +			plt_base = me->mem[MOD_TEXT].base + me->arch.plt_offset;
>   			ip = plt_base + info->plt_offset;
>   			*(int *)insn = 0x0d10e310;	/* basr 1,0  */
>   			*(int *)&insn[4] = 0x100c0004;	/* lg	1,12(1) */
> @@ -375,7 +377,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
>   			       val - loc + 0xffffUL < 0x1ffffeUL) ||
>   			      (r_type == R_390_PLT32DBL &&
>   			       val - loc + 0xffffffffULL < 0x1fffffffeULL)))
> -				val = (Elf_Addr) me->core_layout.base +
> +				val = (Elf_Addr) me->mem[MOD_TEXT].base +
>   					me->arch.plt_offset +
>   					info->plt_offset;
>   			val += rela->r_addend - loc;
> @@ -397,7 +399,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
>   	case R_390_GOTOFF32:	/* 32 bit offset to GOT.  */
>   	case R_390_GOTOFF64:	/* 64 bit offset to GOT. */
>   		val = val + rela->r_addend -
> -			((Elf_Addr) me->core_layout.base + me->arch.got_offset);
> +			((Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset);
>   		if (r_type == R_390_GOTOFF16)
>   			rc = apply_rela_bits(loc, val, 0, 16, 0, write);
>   		else if (r_type == R_390_GOTOFF32)
> @@ -407,7 +409,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
>   		break;
>   	case R_390_GOTPC:	/* 32 bit PC relative offset to GOT. */
>   	case R_390_GOTPCDBL:	/* 32 bit PC rel. off. to GOT shifted by 1. */
> -		val = (Elf_Addr) me->core_layout.base + me->arch.got_offset +
> +		val = (Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset +
>   			rela->r_addend - loc;
>   		if (r_type == R_390_GOTPC)
>   			rc = apply_rela_bits(loc, val, 1, 32, 0, write);
> @@ -515,7 +517,7 @@ int module_finalize(const Elf_Ehdr *hdr,
>   	    !nospec_disable && me->arch.plt_size) {
>   		unsigned int *ij;
>   
> -		ij = me->core_layout.base + me->arch.plt_offset +
> +		ij = me->mem[MOD_TEXT].base + me->arch.plt_offset +
>   			me->arch.plt_size - PLT_ENTRY_SIZE;
>   		ij[0] = 0xc6000000;	/* exrl	%r0,.+10	*/
>   		ij[1] = 0x0005a7f4;	/* j	.		*/
> diff --git a/arch/x86/kernel/callthunks.c b/arch/x86/kernel/callthunks.c
> index ffea98f9064b..22ab13966427 100644
> --- a/arch/x86/kernel/callthunks.c
> +++ b/arch/x86/kernel/callthunks.c
> @@ -330,8 +330,8 @@ void noinline callthunks_patch_module_calls(struct callthunk_sites *cs,
>   					    struct module *mod)
>   {
>   	struct core_text ct = {
> -		.base = (unsigned long)mod->core_layout.base,
> -		.end  = (unsigned long)mod->core_layout.base + mod->core_layout.size,
> +		.base = (unsigned long)mod->mem[MOD_TEXT].base,
> +		.end  = (unsigned long)mod->mem[MOD_TEXT].base + mod->mem[MOD_TEXT].size,
>   		.name = mod->name,
>   	};
>   
> diff --git a/arch/x86/kernel/module.c b/arch/x86/kernel/module.c
> index 705fb2a41d7d..dbc8e226c464 100644
> --- a/arch/x86/kernel/module.c
> +++ b/arch/x86/kernel/module.c
> @@ -339,8 +339,8 @@ int module_finalize(const Elf_Ehdr *hdr,
>   	}
>   	if (locks) {
>   		void *lseg = (void *)locks->sh_addr;
> -		void *text = me->core_layout.base;
> -		void *text_end = text + me->core_layout.text_size;
> +		void *text = me->mem[MOD_TEXT].base;
> +		void *text_end = text + me->mem[MOD_TEXT].size;
>   		alternatives_smp_module_add(me, me->name,
>   					    lseg, lseg + locks->sh_size,
>   					    text, text_end);
> diff --git a/include/linux/module.h b/include/linux/module.h
> index 8c5909c0076c..ccc9de8a93db 100644
> --- a/include/linux/module.h
> +++ b/include/linux/module.h
> @@ -320,17 +320,47 @@ struct mod_tree_node {
>   	struct latch_tree_node node;
>   };
>   
> -struct module_layout {
> -	/* The actual code + data. */
> +enum mod_mem_type {
> +	MOD_TEXT = 0,
> +	MOD_DATA,
> +	MOD_RODATA,
> +	MOD_RO_AFTER_INIT,
> +	MOD_INIT_TEXT,
> +	MOD_INIT_DATA,
> +	MOD_INIT_RODATA,
> +
> +	MOD_MEM_NUM_TYPES,
> +	MOD_INVALID = -1,
> +};
> +
> +#define mod_mem_type_is_init(type)	\
> +	((type) == MOD_INIT_TEXT ||	\
> +	 (type) == MOD_INIT_DATA ||	\
> +	 (type) == MOD_INIT_RODATA)
> +
> +#define mod_mem_type_is_core(type) (!mod_mem_type_is_init(type))
> +
> +#define mod_mem_type_is_text(type)	\
> +	 ((type) == MOD_TEXT ||		\
> +	  (type) == MOD_INIT_TEXT)
> +
> +#define mod_mem_type_is_data(type) (!mod_mem_type_is_text(type))
> +
> +#define mod_mem_type_is_core_data(type)	\
> +	(mod_mem_type_is_core(type) &&	\
> +	 mod_mem_type_is_data(type))
> +
> +#define for_each_mod_mem_type(type)			\
> +	for (enum mod_mem_type (type) = 0;		\
> +	     (type) < MOD_MEM_NUM_TYPES; (type)++)
> +
> +#define for_class_mod_mem_type(type, class)		\
> +	for_each_mod_mem_type(type)			\
> +		if (mod_mem_type_is_##class(type))
> +
> +struct module_memory {
>   	void *base;
> -	/* Total size. */
>   	unsigned int size;
> -	/* The size of the executable code.  */
> -	unsigned int text_size;
> -	/* Size of RO section of the module (text+rodata) */
> -	unsigned int ro_size;
> -	/* Size of RO after init section */
> -	unsigned int ro_after_init_size;
>   
>   #ifdef CONFIG_MODULES_TREE_LOOKUP
>   	struct mod_tree_node mtn;
> @@ -339,9 +369,9 @@ struct module_layout {
>   
>   #ifdef CONFIG_MODULES_TREE_LOOKUP
>   /* Only touch one cacheline for common rbtree-for-core-layout case. */
> -#define __module_layout_align ____cacheline_aligned
> +#define __module_memory_align ____cacheline_aligned
>   #else
> -#define __module_layout_align
> +#define __module_memory_align
>   #endif
>   
>   struct mod_kallsyms {
> @@ -418,12 +448,8 @@ struct module {
>   	/* Startup function. */
>   	int (*init)(void);
>   
> -	/* Core layout: rbtree is accessed frequently, so keep together. */
> -	struct module_layout core_layout __module_layout_align;
> -	struct module_layout init_layout;
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -	struct module_layout data_layout;
> -#endif
> +	/* rbtree is accessed frequently, so keep together. */
> +	struct module_memory mem[MOD_MEM_NUM_TYPES] __module_memory_align;
>   
>   	/* Arch-specific module values */
>   	struct mod_arch_specific arch;
> @@ -573,23 +599,33 @@ bool __is_module_percpu_address(unsigned long addr, unsigned long *can_addr);
>   bool is_module_percpu_address(unsigned long addr);
>   bool is_module_text_address(unsigned long addr);
>   
> +static inline bool within_module_mem_type(unsigned long addr,
> +					  const struct module *mod,
> +					  enum mod_mem_type type)
> +{
> +	unsigned long base, size;
> +
> +	base = (unsigned long)mod->mem[type].base;
> +	size = mod->mem[type].size;
> +	return addr - base < size;
> +}
> +
>   static inline bool within_module_core(unsigned long addr,
>   				      const struct module *mod)
>   {
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -	if ((unsigned long)mod->data_layout.base <= addr &&
> -	    addr < (unsigned long)mod->data_layout.base + mod->data_layout.size)
> +	for_class_mod_mem_type(type, core)
> +		if (within_module_mem_type(addr, mod, type))
>   			return true;
> -#endif
> -	return (unsigned long)mod->core_layout.base <= addr &&
> -	       addr < (unsigned long)mod->core_layout.base + mod->core_layout.size;
> +	return false;
>   }
>   
>   static inline bool within_module_init(unsigned long addr,
>   				      const struct module *mod)
>   {
> -	return (unsigned long)mod->init_layout.base <= addr &&
> -	       addr < (unsigned long)mod->init_layout.base + mod->init_layout.size;
> +	for_class_mod_mem_type(type, init)
> +		if (within_module_mem_type(addr, mod, type))
> +			return true;
> +	return false;
>   }
>   
>   static inline bool within_module(unsigned long addr, const struct module *mod)
> diff --git a/kernel/module/internal.h b/kernel/module/internal.h
> index 2e2bf236f558..c6c44ceca07a 100644
> --- a/kernel/module/internal.h
> +++ b/kernel/module/internal.h
> @@ -17,27 +17,19 @@
>   #define ARCH_SHF_SMALL 0
>   #endif
>   
> -/* If this is set, the section belongs in the init part of the module */
> -#define INIT_OFFSET_MASK (1UL << (BITS_PER_LONG - 1))
> -/* Maximum number of characters written by module_flags() */
> -#define MODULE_FLAGS_BUF_SIZE (TAINT_FLAGS_COUNT + 4)
> -
> -#ifndef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -#define	data_layout core_layout
> -#endif
> -
>   /*
> - * Modules' sections will be aligned on page boundaries
> - * to ensure complete separation of code and data, but
> - * only when CONFIG_STRICT_MODULE_RWX=y
> + * Use highest 4 bits of sh_entsize to store the mod_mem_type of this
> + * section. This leaves 28 bits for offset on 32-bit systems, which is
> + * about 256 MiB (WARN_ON_ONCE if we exceed that).
>    */
> -static inline unsigned int strict_align(unsigned int size)
> -{
> -	if (IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
> -		return PAGE_ALIGN(size);
> -	else
> -		return size;
> -}
> +
> +#define SH_ENTSIZE_TYPE_BITS	4
> +#define SH_ENTSIZE_TYPE_SHIFT	(BITS_PER_LONG - SH_ENTSIZE_TYPE_BITS)
> +#define SH_ENTSIZE_TYPE_MASK	((1UL << SH_ENTSIZE_TYPE_BITS) - 1)
> +#define SH_ENTSIZE_OFFSET_MASK	((1UL << (BITS_PER_LONG - SH_ENTSIZE_TYPE_BITS)) - 1)
> +
> +/* Maximum number of characters written by module_flags() */
> +#define MODULE_FLAGS_BUF_SIZE (TAINT_FLAGS_COUNT + 4)
>   
>   extern struct mutex module_mutex;
>   extern struct list_head modules;
> @@ -101,8 +93,8 @@ int try_to_force_load(struct module *mod, const char *reason);
>   bool find_symbol(struct find_symbol_arg *fsa);
>   struct module *find_module_all(const char *name, size_t len, bool even_unformed);
>   int cmp_name(const void *name, const void *sym);
> -long module_get_offset(struct module *mod, unsigned int *size, Elf_Shdr *sechdr,
> -		       unsigned int section);
> +long module_get_offset_and_type(struct module *mod, enum mod_mem_type type,
> +				Elf_Shdr *sechdr, unsigned int section);
>   char *module_flags(struct module *mod, char *buf, bool show_state);
>   size_t module_flags_taint(unsigned long taints, char *buf);
>   
> @@ -190,10 +182,13 @@ struct mod_tree_root {
>   #endif
>   	unsigned long addr_min;
>   	unsigned long addr_max;
> +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> +	unsigned long data_addr_min;
> +	unsigned long data_addr_max;
> +#endif
>   };
>   
>   extern struct mod_tree_root mod_tree;
> -extern struct mod_tree_root mod_data_tree;
>   
>   #ifdef CONFIG_MODULES_TREE_LOOKUP
>   void mod_tree_insert(struct module *mod);
> @@ -224,7 +219,6 @@ void module_enable_nx(const struct module *mod);
>   void module_enable_x(const struct module *mod);
>   int module_enforce_rwx_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
>   				char *secstrings, struct module *mod);
> -bool module_check_misalignment(const struct module *mod);
>   
>   #ifdef CONFIG_MODULE_SIG
>   int module_sig_check(struct load_info *info, int flags);
> diff --git a/kernel/module/kallsyms.c b/kernel/module/kallsyms.c
> index 4523f99b0358..acdca7692961 100644
> --- a/kernel/module/kallsyms.c
> +++ b/kernel/module/kallsyms.c
> @@ -78,6 +78,7 @@ static bool is_core_symbol(const Elf_Sym *src, const Elf_Shdr *sechdrs,
>   			   unsigned int shnum, unsigned int pcpundx)
>   {
>   	const Elf_Shdr *sec;
> +	enum mod_mem_type type;
>   
>   	if (src->st_shndx == SHN_UNDEF ||
>   	    src->st_shndx >= shnum ||
> @@ -90,11 +91,12 @@ static bool is_core_symbol(const Elf_Sym *src, const Elf_Shdr *sechdrs,
>   #endif
>   
>   	sec = sechdrs + src->st_shndx;
> +	type = sec->sh_entsize >> SH_ENTSIZE_TYPE_SHIFT;
>   	if (!(sec->sh_flags & SHF_ALLOC)
>   #ifndef CONFIG_KALLSYMS_ALL
>   	    || !(sec->sh_flags & SHF_EXECINSTR)
>   #endif
> -	    || (sec->sh_entsize & INIT_OFFSET_MASK))
> +	    || mod_mem_type_is_init(type))
>   		return false;
>   
>   	return true;
> @@ -113,11 +115,13 @@ void layout_symtab(struct module *mod, struct load_info *info)
>   	Elf_Shdr *strsect = info->sechdrs + info->index.str;
>   	const Elf_Sym *src;
>   	unsigned int i, nsrc, ndst, strtab_size = 0;
> +	struct module_memory *mod_mem_data = &mod->mem[MOD_DATA];
> +	struct module_memory *mod_mem_init_data = &mod->mem[MOD_INIT_DATA];
>   
>   	/* Put symbol section at end of init part of module. */
>   	symsect->sh_flags |= SHF_ALLOC;
> -	symsect->sh_entsize = module_get_offset(mod, &mod->init_layout.size, symsect,
> -						info->index.sym) | INIT_OFFSET_MASK;
> +	symsect->sh_entsize = module_get_offset_and_type(mod, MOD_INIT_DATA,
> +							 symsect, info->index.sym);
>   	pr_debug("\t%s\n", info->secstrings + symsect->sh_name);
>   
>   	src = (void *)info->hdr + symsect->sh_offset;
> @@ -134,28 +138,27 @@ void layout_symtab(struct module *mod, struct load_info *info)
>   	}
>   
>   	/* Append room for core symbols at end of core part. */
> -	info->symoffs = ALIGN(mod->data_layout.size, symsect->sh_addralign ?: 1);
> -	info->stroffs = mod->data_layout.size = info->symoffs + ndst * sizeof(Elf_Sym);
> -	mod->data_layout.size += strtab_size;
> +	info->symoffs = ALIGN(mod_mem_data->size, symsect->sh_addralign ?: 1);
> +	info->stroffs = mod_mem_data->size = info->symoffs + ndst * sizeof(Elf_Sym);
> +	mod_mem_data->size += strtab_size;
>   	/* Note add_kallsyms() computes strtab_size as core_typeoffs - stroffs */
> -	info->core_typeoffs = mod->data_layout.size;
> -	mod->data_layout.size += ndst * sizeof(char);
> -	mod->data_layout.size = strict_align(mod->data_layout.size);
> +	info->core_typeoffs = mod_mem_data->size;
> +	mod_mem_data->size += ndst * sizeof(char);
>   
>   	/* Put string table section at end of init part of module. */
>   	strsect->sh_flags |= SHF_ALLOC;
> -	strsect->sh_entsize = module_get_offset(mod, &mod->init_layout.size, strsect,
> -						info->index.str) | INIT_OFFSET_MASK;
> +	strsect->sh_entsize = module_get_offset_and_type(mod, MOD_INIT_DATA,
> +							 strsect, info->index.str);
>   	pr_debug("\t%s\n", info->secstrings + strsect->sh_name);
>   
>   	/* We'll tack temporary mod_kallsyms on the end. */
> -	mod->init_layout.size = ALIGN(mod->init_layout.size,
> +	mod_mem_init_data->size = ALIGN(mod_mem_init_data->size,
>   					__alignof__(struct mod_kallsyms));
> -	info->mod_kallsyms_init_off = mod->init_layout.size;
> -	mod->init_layout.size += sizeof(struct mod_kallsyms);
> -	info->init_typeoffs = mod->init_layout.size;
> -	mod->init_layout.size += nsrc * sizeof(char);
> -	mod->init_layout.size = strict_align(mod->init_layout.size);
> +	info->mod_kallsyms_init_off = mod_mem_init_data->size;
> +
> +	mod_mem_init_data->size += sizeof(struct mod_kallsyms);
> +	info->init_typeoffs = mod_mem_init_data->size;
> +	mod_mem_init_data->size += nsrc * sizeof(char);
>   }
>   
>   /*
> @@ -171,9 +174,11 @@ void add_kallsyms(struct module *mod, const struct load_info *info)
>   	char *s;
>   	Elf_Shdr *symsec = &info->sechdrs[info->index.sym];
>   	unsigned long strtab_size;
> +	void *data_base = mod->mem[MOD_DATA].base;
> +	void *init_data_base = mod->mem[MOD_INIT_DATA].base;
>   
>   	/* Set up to point into init section. */
> -	mod->kallsyms = (void __rcu *)mod->init_layout.base +
> +	mod->kallsyms = (void __rcu *)init_data_base +
>   		info->mod_kallsyms_init_off;
>   
>   	rcu_read_lock();
> @@ -183,15 +188,15 @@ void add_kallsyms(struct module *mod, const struct load_info *info)
>   	/* Make sure we get permanent strtab: don't use info->strtab. */
>   	rcu_dereference(mod->kallsyms)->strtab =
>   		(void *)info->sechdrs[info->index.str].sh_addr;
> -	rcu_dereference(mod->kallsyms)->typetab = mod->init_layout.base + info->init_typeoffs;
> +	rcu_dereference(mod->kallsyms)->typetab = init_data_base + info->init_typeoffs;
>   
>   	/*
>   	 * Now populate the cut down core kallsyms for after init
>   	 * and set types up while we still have access to sections.
>   	 */
> -	mod->core_kallsyms.symtab = dst = mod->data_layout.base + info->symoffs;
> -	mod->core_kallsyms.strtab = s = mod->data_layout.base + info->stroffs;
> -	mod->core_kallsyms.typetab = mod->data_layout.base + info->core_typeoffs;
> +	mod->core_kallsyms.symtab = dst = data_base + info->symoffs;
> +	mod->core_kallsyms.strtab = s = data_base + info->stroffs;
> +	mod->core_kallsyms.typetab = data_base + info->core_typeoffs;
>   	strtab_size = info->core_typeoffs - info->stroffs;
>   	src = rcu_dereference(mod->kallsyms)->symtab;
>   	for (ndst = i = 0; i < rcu_dereference(mod->kallsyms)->num_symtab; i++) {
> @@ -267,12 +272,15 @@ static const char *find_kallsyms_symbol(struct module *mod,
>   	unsigned int i, best = 0;
>   	unsigned long nextval, bestval;
>   	struct mod_kallsyms *kallsyms = rcu_dereference_sched(mod->kallsyms);
> +	struct module_memory *mod_mem;
>   
>   	/* At worse, next value is at end of module */
>   	if (within_module_init(addr, mod))
> -		nextval = (unsigned long)mod->init_layout.base + mod->init_layout.text_size;
> +		mod_mem = &mod->mem[MOD_INIT_TEXT];
>   	else
> -		nextval = (unsigned long)mod->core_layout.base + mod->core_layout.text_size;
> +		mod_mem = &mod->mem[MOD_TEXT];
> +
> +	nextval = (unsigned long)mod_mem->base + mod_mem->size;
>   
>   	bestval = kallsyms_symbol_value(&kallsyms->symtab[best]);
>   
> diff --git a/kernel/module/kdb.c b/kernel/module/kdb.c
> index f4317f92e189..995c32d3698f 100644
> --- a/kernel/module/kdb.c
> +++ b/kernel/module/kdb.c
> @@ -26,10 +26,11 @@ int kdb_lsmod(int argc, const char **argv)
>   		if (mod->state == MODULE_STATE_UNFORMED)
>   			continue;
>   
> -		kdb_printf("%-20s%8u", mod->name, mod->core_layout.size);
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -		kdb_printf("/%8u", mod->data_layout.size);
> -#endif
> +		kdb_printf("%-20s%8u", mod->name, mod->mem[MOD_TEXT].size);
> +		kdb_printf("/%8u", mod->mem[MOD_RODATA].size);
> +		kdb_printf("/%8u", mod->mem[MOD_RO_AFTER_INIT].size);
> +		kdb_printf("/%8u", mod->mem[MOD_DATA].size);
> +
>   		kdb_printf("  0x%px ", (void *)mod);
>   #ifdef CONFIG_MODULE_UNLOAD
>   		kdb_printf("%4d ", module_refcount(mod));
> @@ -40,10 +41,10 @@ int kdb_lsmod(int argc, const char **argv)
>   			kdb_printf(" (Loading)");
>   		else
>   			kdb_printf(" (Live)");
> -		kdb_printf(" 0x%px", mod->core_layout.base);
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -		kdb_printf("/0x%px", mod->data_layout.base);
> -#endif
> +		kdb_printf(" 0x%px", mod->mem[MOD_TEXT].base);
> +		kdb_printf("/0x%px", mod->mem[MOD_RODATA].base);
> +		kdb_printf("/0x%px", mod->mem[MOD_RO_AFTER_INIT].base);
> +		kdb_printf("/0x%px", mod->mem[MOD_DATA].base);
>   
>   #ifdef CONFIG_MODULE_UNLOAD
>   		{
> diff --git a/kernel/module/main.c b/kernel/module/main.c
> index d3be89de706d..9f890d14072c 100644
> --- a/kernel/module/main.c
> +++ b/kernel/module/main.c
> @@ -80,12 +80,6 @@ struct mod_tree_root mod_tree __cacheline_aligned = {
>   	.addr_min = -1UL,
>   };
>   
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -struct mod_tree_root mod_data_tree __cacheline_aligned = {
> -	.addr_min = -1UL,
> -};
> -#endif
> -
>   struct symsearch {
>   	const struct kernel_symbol *start, *stop;
>   	const s32 *crcs;
> @@ -93,14 +87,24 @@ struct symsearch {
>   };
>   
>   /*
> - * Bounds of module text, for speeding up __module_address.
> + * Bounds of module memory, for speeding up __module_address.
>    * Protected by module_mutex.
>    */
> -static void __mod_update_bounds(void *base, unsigned int size, struct mod_tree_root *tree)
> +static void __mod_update_bounds(enum mod_mem_type type __maybe_unused, void *base,
> +				unsigned int size, struct mod_tree_root *tree)
>   {
>   	unsigned long min = (unsigned long)base;
>   	unsigned long max = min + size;
>   
> +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> +	if (mod_mem_type_is_core_data(type)) {
> +		if (min < tree->data_addr_min)
> +			tree->data_addr_min = min;
> +		if (max > tree->data_addr_max)
> +			tree->data_addr_max = max;
> +		return;
> +	}
> +#endif
>   	if (min < tree->addr_min)
>   		tree->addr_min = min;
>   	if (max > tree->addr_max)
> @@ -109,12 +113,12 @@ static void __mod_update_bounds(void *base, unsigned int size, struct mod_tree_r
>   
>   static void mod_update_bounds(struct module *mod)
>   {
> -	__mod_update_bounds(mod->core_layout.base, mod->core_layout.size, &mod_tree);
> -	if (mod->init_layout.size)
> -		__mod_update_bounds(mod->init_layout.base, mod->init_layout.size, &mod_tree);
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -	__mod_update_bounds(mod->data_layout.base, mod->data_layout.size, &mod_data_tree);
> -#endif
> +	for_each_mod_mem_type(type) {
> +		struct module_memory *mod_mem = &mod->mem[type];
> +
> +		if (mod_mem->size)
> +			__mod_update_bounds(type, mod_mem->base, mod_mem->size, &mod_tree);
> +	}
>   }
>   
>   /* Block module loading/unloading? */
> @@ -923,12 +927,27 @@ static ssize_t store_uevent(struct module_attribute *mattr,
>   struct module_attribute module_uevent =
>   	__ATTR(uevent, 0200, NULL, store_uevent);
>   
> +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> +
>   static ssize_t show_coresize(struct module_attribute *mattr,
>   			     struct module_kobject *mk, char *buffer)
>   {
> -	return sprintf(buffer, "%u\n", mk->mod->core_layout.size);
> +	return sprintf(buffer, "%u\n", mk->mod->mem[MOD_TEXT].size);
>   }
>   
> +#else
> +
> +static ssize_t show_coresize(struct module_attribute *mattr,
> +			     struct module_kobject *mk, char *buffer)
> +{
> +	unsigned int size = 0;
> +
> +	for_class_mod_mem_type(type, core)
> +		size += mk->mod->mem[type].size;
> +	return sprintf(buffer, "%u\n", size);
> +}
> +#endif
> +
>   static struct module_attribute modinfo_coresize =
>   	__ATTR(coresize, 0444, show_coresize, NULL);
>   
> @@ -936,7 +955,14 @@ static struct module_attribute modinfo_coresize =
>   static ssize_t show_datasize(struct module_attribute *mattr,
>   			     struct module_kobject *mk, char *buffer)
>   {
> -	return sprintf(buffer, "%u\n", mk->mod->data_layout.size);
> +	unsigned int size = 0;
> +
> +	for_class_mod_mem_type(type, core) {
> +		if (type == MOD_TEXT)
> +			continue;
> +		size += mk->mod->mem[type].size;
> +	}
> +	return sprintf(buffer, "%u\n", size);
>   }
>   
>   static struct module_attribute modinfo_datasize =
> @@ -946,7 +972,11 @@ static struct module_attribute modinfo_datasize =
>   static ssize_t show_initsize(struct module_attribute *mattr,
>   			     struct module_kobject *mk, char *buffer)
>   {
> -	return sprintf(buffer, "%u\n", mk->mod->init_layout.size);
> +	unsigned int size = 0;
> +
> +	for_class_mod_mem_type(type, init)
> +		size += mk->mod->mem[type].size;
> +	return sprintf(buffer, "%u\n", size);
>   }
>   
>   static struct module_attribute modinfo_initsize =
> @@ -1143,6 +1173,63 @@ void __weak module_arch_freeing_init(struct module *mod)
>   {
>   }
>   
> +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> +static bool mod_mem_use_vmalloc(enum mod_mem_type type)
> +{
> +	return mod_mem_is_core_data(type);
> +}
> +#else
> +static bool mod_mem_use_vmalloc(enum mod_mem_type type)
> +{
> +	return false;
> +}
> +#endif
> +
> +static void *module_memory_alloc(unsigned int size, enum mod_mem_type type)
> +{
> +	if (mod_mem_use_vmalloc(type))
> +		return vzalloc(size);
> +	return module_alloc(size);
> +}
> +
> +static void module_memory_free(void *ptr, enum mod_mem_type type)
> +{
> +	if (mod_mem_use_vmalloc(type))
> +		vfree(ptr);
> +	else
> +		module_memfree(ptr);
> +}
> +
> +static void free_mod_mem(struct module *mod)
> +{
> +	/* free the memory in the right order to avoid use-after-free */
> +	static enum mod_mem_type mod_mem_free_order[MOD_MEM_NUM_TYPES] = {
> +		/* first free init sections */
> +		MOD_INIT_TEXT,
> +		MOD_INIT_DATA,
> +		MOD_INIT_RODATA,
> +
> +		/* then core sections, except rw data */
> +		MOD_TEXT,
> +		MOD_RODATA,
> +		MOD_RO_AFTER_INIT,
> +
> +		/* last, rw data */
> +		MOD_DATA,
> +	};
> +	int i;
> +
> +	for (i = 0; i < MOD_MEM_NUM_TYPES; i++) {
> +		enum mod_mem_type type = mod_mem_free_order[i];
> +		struct module_memory *mod_mem = &mod->mem[type];
> +
> +		/* Free lock-classes; relies on the preceding sync_rcu(). */
> +		lockdep_free_key_range(mod_mem->base, mod_mem->size);
> +		if (mod_mem->size)
> +			module_memory_free(mod_mem->base, type);
> +	}
> +}
> +
>   /* Free a module, remove from lists, etc. */
>   static void free_module(struct module *mod)
>   {
> @@ -1189,18 +1276,10 @@ static void free_module(struct module *mod)
>   
>   	/* This may be empty, but that's OK */
>   	module_arch_freeing_init(mod);
> -	module_memfree(mod->init_layout.base);
>   	kfree(mod->args);
>   	percpu_modfree(mod);
>   
> -	/* Free lock-classes; relies on the preceding sync_rcu(). */
> -	lockdep_free_key_range(mod->data_layout.base, mod->data_layout.size);
> -
> -	/* Finally, free the core (containing the module structure) */
> -	module_memfree(mod->core_layout.base);
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -	vfree(mod->data_layout.base);
> -#endif
> +	free_mod_mem(mod);
>   }
>   
>   void *__symbol_get(const char *symbol)
> @@ -1387,16 +1466,18 @@ unsigned int __weak arch_mod_section_prepend(struct module *mod,
>   	return 0;
>   }
>   
> -/* Update size with this section: return offset. */
> -long module_get_offset(struct module *mod, unsigned int *size,
> +long module_get_offset_and_type(struct module *mod, enum mod_mem_type type,
>   				Elf_Shdr *sechdr, unsigned int section)
>   {
> -	long ret;
> +	long offset;
> +	long mask = ((unsigned long)(type) & SH_ENTSIZE_TYPE_MASK) << SH_ENTSIZE_TYPE_SHIFT;
>   
> -	*size += arch_mod_section_prepend(mod, section);
> -	ret = ALIGN(*size, sechdr->sh_addralign ?: 1);
> -	*size = ret + sechdr->sh_size;
> -	return ret;
> +	mod->mem[type].size += arch_mod_section_prepend(mod, section);
> +	offset = ALIGN(mod->mem[type].size, sechdr->sh_addralign ?: 1);
> +	mod->mem[type].size = offset + sechdr->sh_size;
> +
> +	WARN_ON_ONCE(offset & mask);
> +	return offset | mask;
>   }
>   
>   static bool module_init_layout_section(const char *sname)
> @@ -1428,6 +1509,20 @@ static void layout_sections(struct module *mod, struct load_info *info)
>   		{ SHF_WRITE | SHF_ALLOC, ARCH_SHF_SMALL },
>   		{ ARCH_SHF_SMALL | SHF_ALLOC, 0 }
>   	};
> +	static int core_m_to_mem_type[] = {
> +		MOD_TEXT,
> +		MOD_RODATA,
> +		MOD_RO_AFTER_INIT,
> +		MOD_DATA,
> +		MOD_INVALID,
> +	};
> +	static int init_m_to_mem_type[] = {
> +		MOD_INIT_TEXT,
> +		MOD_INIT_RODATA,
> +		MOD_INVALID,
> +		MOD_INIT_DATA,
> +		MOD_INVALID,
> +	};
>   	unsigned int m, i;
>   
>   	for (i = 0; i < info->hdr->e_shnum; i++)
> @@ -1435,41 +1530,30 @@ static void layout_sections(struct module *mod, struct load_info *info)
>   
>   	pr_debug("Core section allocation order:\n");
>   	for (m = 0; m < ARRAY_SIZE(masks); ++m) {
> +		enum mod_mem_type type = core_m_to_mem_type[m];
> +
>   		for (i = 0; i < info->hdr->e_shnum; ++i) {
>   			Elf_Shdr *s = &info->sechdrs[i];
>   			const char *sname = info->secstrings + s->sh_name;
> -			unsigned int *sizep;
>   
>   			if ((s->sh_flags & masks[m][0]) != masks[m][0]
>   			    || (s->sh_flags & masks[m][1])
>   			    || s->sh_entsize != ~0UL
>   			    || module_init_layout_section(sname))
>   				continue;
> -			sizep = m ? &mod->data_layout.size : &mod->core_layout.size;
> -			s->sh_entsize = module_get_offset(mod, sizep, s, i);
> +
> +			if (WARN_ON_ONCE(type == MOD_INVALID))
> +				continue;
> +
> +			s->sh_entsize = module_get_offset_and_type(mod, type, s, i);
>   			pr_debug("\t%s\n", sname);
>   		}
> -		switch (m) {
> -		case 0: /* executable */
> -			mod->core_layout.size = strict_align(mod->core_layout.size);
> -			mod->core_layout.text_size = mod->core_layout.size;
> -			break;
> -		case 1: /* RO: text and ro-data */
> -			mod->data_layout.size = strict_align(mod->data_layout.size);
> -			mod->data_layout.ro_size = mod->data_layout.size;
> -			break;
> -		case 2: /* RO after init */
> -			mod->data_layout.size = strict_align(mod->data_layout.size);
> -			mod->data_layout.ro_after_init_size = mod->data_layout.size;
> -			break;
> -		case 4: /* whole core */
> -			mod->data_layout.size = strict_align(mod->data_layout.size);
> -			break;
> -		}
>   	}
>   
>   	pr_debug("Init section allocation order:\n");
>   	for (m = 0; m < ARRAY_SIZE(masks); ++m) {
> +		enum mod_mem_type type = init_m_to_mem_type[m];
> +
>   		for (i = 0; i < info->hdr->e_shnum; ++i) {
>   			Elf_Shdr *s = &info->sechdrs[i];
>   			const char *sname = info->secstrings + s->sh_name;
> @@ -1479,30 +1563,13 @@ static void layout_sections(struct module *mod, struct load_info *info)
>   			    || s->sh_entsize != ~0UL
>   			    || !module_init_layout_section(sname))
>   				continue;
> -			s->sh_entsize = (module_get_offset(mod, &mod->init_layout.size, s, i)
> -					 | INIT_OFFSET_MASK);
> +
> +			if (WARN_ON_ONCE(type == MOD_INVALID))
> +				continue;
> +
> +			s->sh_entsize = module_get_offset_and_type(mod, type, s, i);
>   			pr_debug("\t%s\n", sname);
>   		}
> -		switch (m) {
> -		case 0: /* executable */
> -			mod->init_layout.size = strict_align(mod->init_layout.size);
> -			mod->init_layout.text_size = mod->init_layout.size;
> -			break;
> -		case 1: /* RO: text and ro-data */
> -			mod->init_layout.size = strict_align(mod->init_layout.size);
> -			mod->init_layout.ro_size = mod->init_layout.size;
> -			break;
> -		case 2:
> -			/*
> -			 * RO after init doesn't apply to init_layout (only
> -			 * core_layout), so it just takes the value of ro_size.
> -			 */
> -			mod->init_layout.ro_after_init_size = mod->init_layout.ro_size;
> -			break;
> -		case 4: /* whole init */
> -			mod->init_layout.size = strict_align(mod->init_layout.size);
> -			break;
> -		}
>   	}
>   }
>   
> @@ -2122,72 +2189,42 @@ static int move_module(struct module *mod, struct load_info *info)
>   {
>   	int i;
>   	void *ptr;
> +	enum mod_mem_type t;
>   
> -	/* Do the allocs. */
> -	ptr = module_alloc(mod->core_layout.size);
> -	/*
> -	 * The pointer to this block is stored in the module structure
> -	 * which is inside the block. Just mark it as not being a
> -	 * leak.
> -	 */
> -	kmemleak_not_leak(ptr);
> -	if (!ptr)
> -		return -ENOMEM;
> -
> -	memset(ptr, 0, mod->core_layout.size);
> -	mod->core_layout.base = ptr;
> -
> -	if (mod->init_layout.size) {
> -		ptr = module_alloc(mod->init_layout.size);
> -		/*
> -		 * The pointer to this block is stored in the module structure
> -		 * which is inside the block. This block doesn't need to be
> -		 * scanned as it contains data and code that will be freed
> -		 * after the module is initialized.
> -		 */
> -		kmemleak_ignore(ptr);
> -		if (!ptr) {
> -			module_memfree(mod->core_layout.base);
> -			return -ENOMEM;
> +	for_each_mod_mem_type(type) {
> +		if (!mod->mem[type].size) {
> +			mod->mem[type].base = NULL;
> +			continue;
>   		}
> -		memset(ptr, 0, mod->init_layout.size);
> -		mod->init_layout.base = ptr;
> -	} else
> -		mod->init_layout.base = NULL;
> +		mod->mem[type].size = PAGE_ALIGN(mod->mem[type].size);
> +		ptr = module_memory_alloc(mod->mem[type].size, type);
>   
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -	/* Do the allocs. */
> -	ptr = vzalloc(mod->data_layout.size);
>   		/*
>   		 * The pointer to this block is stored in the module structure
>   		 * which is inside the block. Just mark it as not being a
>   		 * leak.
>   		 */
> -	kmemleak_not_leak(ptr);
> +		kmemleak_ignore(ptr);
>   		if (!ptr) {
> -		module_memfree(mod->core_layout.base);
> -		module_memfree(mod->init_layout.base);
> -		return -ENOMEM;
> +			t = type;
> +			goto out_enomem;
> +		}
> +		memset(ptr, 0, mod->mem[type].size);
> +		mod->mem[type].base = ptr;
>   	}
>   
> -	mod->data_layout.base = ptr;
> -#endif
>   	/* Transfer each section which specifies SHF_ALLOC */
>   	pr_debug("final section addresses:\n");
>   	for (i = 0; i < info->hdr->e_shnum; i++) {
>   		void *dest;
>   		Elf_Shdr *shdr = &info->sechdrs[i];
> +		enum mod_mem_type type = shdr->sh_entsize >> SH_ENTSIZE_TYPE_SHIFT;
>   
>   		if (!(shdr->sh_flags & SHF_ALLOC))
>   			continue;
>   
> -		if (shdr->sh_entsize & INIT_OFFSET_MASK)
> -			dest = mod->init_layout.base
> -				+ (shdr->sh_entsize & ~INIT_OFFSET_MASK);
> -		else if (!(shdr->sh_flags & SHF_EXECINSTR))
> -			dest = mod->data_layout.base + shdr->sh_entsize;
> -		else
> -			dest = mod->core_layout.base + shdr->sh_entsize;
> +		dest = mod->mem[type].base +
> +			(shdr->sh_entsize & SH_ENTSIZE_OFFSET_MASK);
>   
>   		if (shdr->sh_type != SHT_NOBITS)
>   			memcpy(dest, (void *)shdr->sh_addr, shdr->sh_size);
> @@ -2198,6 +2235,10 @@ static int move_module(struct module *mod, struct load_info *info)
>   	}
>   
>   	return 0;
> +out_enomem:
> +	for (t--; t >= 0; t--)
> +		module_memory_free(mod->mem[t].base, t);
> +	return -ENOMEM;
>   }
>   
>   static int check_module_license_and_versions(struct module *mod)
> @@ -2242,12 +2283,13 @@ static void flush_module_icache(const struct module *mod)
>   	 * Do it before processing of module parameters, so the module
>   	 * can provide parameter accessor functions of its own.
>   	 */
> -	if (mod->init_layout.base)
> -		flush_icache_range((unsigned long)mod->init_layout.base,
> -				   (unsigned long)mod->init_layout.base
> -				   + mod->init_layout.size);
> -	flush_icache_range((unsigned long)mod->core_layout.base,
> -			   (unsigned long)mod->core_layout.base + mod->core_layout.size);
> +	for_each_mod_mem_type(type) {
> +		const struct module_memory *mod_mem = &mod->mem[type];
> +
> +		if (mod_mem->size)
> +			flush_icache_range((unsigned long)mod_mem->base,
> +					   (unsigned long)mod_mem->base + mod_mem->size);
> +	}
>   }
>   
>   bool __weak module_elf_check_arch(Elf_Ehdr *hdr)
> @@ -2350,11 +2392,8 @@ static void module_deallocate(struct module *mod, struct load_info *info)
>   {
>   	percpu_modfree(mod);
>   	module_arch_freeing_init(mod);
> -	module_memfree(mod->init_layout.base);
> -	module_memfree(mod->core_layout.base);
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -	vfree(mod->data_layout.base);
> -#endif
> +
> +	free_mod_mem(mod);
>   }
>   
>   int __weak module_finalize(const Elf_Ehdr *hdr,
> @@ -2415,7 +2454,9 @@ static void do_mod_ctors(struct module *mod)
>   /* For freeing module_init on success, in case kallsyms traversing */
>   struct mod_initfree {
>   	struct llist_node node;
> -	void *module_init;
> +	void *init_text;
> +	void *init_data;
> +	void *init_rodata;
>   };
>   
>   static void do_free_init(struct work_struct *w)
> @@ -2429,7 +2470,9 @@ static void do_free_init(struct work_struct *w)
>   
>   	llist_for_each_safe(pos, n, list) {
>   		initfree = container_of(pos, struct mod_initfree, node);
> -		module_memfree(initfree->module_init);
> +		module_memfree(initfree->init_text);
> +		module_memfree(initfree->init_data);
> +		module_memfree(initfree->init_rodata);
>   		kfree(initfree);
>   	}
>   }
> @@ -2456,7 +2499,9 @@ static noinline int do_init_module(struct module *mod)
>   		ret = -ENOMEM;
>   		goto fail;
>   	}
> -	freeinit->module_init = mod->init_layout.base;
> +	freeinit->init_text = mod->mem[MOD_INIT_TEXT].base;
> +	freeinit->init_data = mod->mem[MOD_INIT_DATA].base;
> +	freeinit->init_rodata = mod->mem[MOD_INIT_RODATA].base;
>   
>   	do_mod_ctors(mod);
>   	/* Start the module */
> @@ -2492,8 +2537,8 @@ static noinline int do_init_module(struct module *mod)
>   	if (!mod->async_probe_requested)
>   		async_synchronize_full();
>   
> -	ftrace_free_mem(mod, mod->init_layout.base, mod->init_layout.base +
> -			mod->init_layout.size);
> +	ftrace_free_mem(mod, mod->mem[MOD_INIT_TEXT].base,
> +			mod->mem[MOD_INIT_TEXT].base + mod->mem[MOD_INIT_TEXT].size);
>   	mutex_lock(&module_mutex);
>   	/* Drop initial reference. */
>   	module_put(mod);
> @@ -2505,11 +2550,10 @@ static noinline int do_init_module(struct module *mod)
>   	module_enable_ro(mod, true);
>   	mod_tree_remove_init(mod);
>   	module_arch_freeing_init(mod);
> -	mod->init_layout.base = NULL;
> -	mod->init_layout.size = 0;
> -	mod->init_layout.ro_size = 0;
> -	mod->init_layout.ro_after_init_size = 0;
> -	mod->init_layout.text_size = 0;
> +	for_class_mod_mem_type(type, init) {
> +		mod->mem[type].base = NULL;
> +		mod->mem[type].size = 0;
> +	}
>   #ifdef CONFIG_DEBUG_INFO_BTF_MODULES
>   	/* .BTF is not SHF_ALLOC and will get removed, so sanitize pointer */
>   	mod->btf_data = NULL;
> @@ -2628,9 +2672,6 @@ static int complete_formation(struct module *mod, struct load_info *info)
>   	module_bug_finalize(info->hdr, info->sechdrs, mod);
>   	module_cfi_finalize(info->hdr, info->sechdrs, mod);
>   
> -	if (module_check_misalignment(mod))
> -		goto out_misaligned;
> -
>   	module_enable_ro(mod, false);
>   	module_enable_nx(mod);
>   	module_enable_x(mod);
> @@ -2644,8 +2685,6 @@ static int complete_formation(struct module *mod, struct load_info *info)
>   
>   	return 0;
>   
> -out_misaligned:
> -	err = -EINVAL;
>   out:
>   	mutex_unlock(&module_mutex);
>   	return err;
> @@ -2909,7 +2948,10 @@ static int load_module(struct load_info *info, const char __user *uargs,
>   	mutex_unlock(&module_mutex);
>    free_module:
>   	/* Free lock-classes; relies on the preceding sync_rcu() */
> -	lockdep_free_key_range(mod->data_layout.base, mod->data_layout.size);
> +	for_class_mod_mem_type(type, core) {
> +		lockdep_free_key_range(mod->mem[type].base,
> +				       mod->mem[type].size);
> +	}
>   
>   	module_deallocate(mod, info);
>    free_copy:
> @@ -3060,20 +3102,21 @@ bool is_module_address(unsigned long addr)
>   struct module *__module_address(unsigned long addr)
>   {
>   	struct module *mod;
> -	struct mod_tree_root *tree;
>   
>   	if (addr >= mod_tree.addr_min && addr <= mod_tree.addr_max)
> -		tree = &mod_tree;
> +		goto lookup;
> +
>   #ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -	else if (addr >= mod_data_tree.addr_min && addr <= mod_data_tree.addr_max)
> -		tree = &mod_data_tree;
> +	if (addr >= mod_tree.data_addr_min && addr <= mod_tree.data_addr_max)
> +		goto lookup;
>   #endif
> -	else
> +
>   	return NULL;
>   
> +lookup:
>   	module_assert_mutex_or_preempt();
>   
> -	mod = mod_find(addr, tree);
> +	mod = mod_find(addr, &mod_tree);
>   	if (mod) {
>   		BUG_ON(!within_module(addr, mod));
>   		if (mod->state == MODULE_STATE_UNFORMED)
> @@ -3113,8 +3156,8 @@ struct module *__module_text_address(unsigned long addr)
>   	struct module *mod = __module_address(addr);
>   	if (mod) {
>   		/* Make sure it's within the text section. */
> -		if (!within(addr, mod->init_layout.base, mod->init_layout.text_size)
> -		    && !within(addr, mod->core_layout.base, mod->core_layout.text_size))
> +		if (!within_module_mem_type(addr, mod, MOD_TEXT) &&
> +		    !within_module_mem_type(addr, mod, MOD_INIT_TEXT))
>   			mod = NULL;
>   	}
>   	return mod;
> diff --git a/kernel/module/procfs.c b/kernel/module/procfs.c
> index cf5b9f1e6ec4..0a4841e88adb 100644
> --- a/kernel/module/procfs.c
> +++ b/kernel/module/procfs.c
> @@ -62,6 +62,15 @@ static void m_stop(struct seq_file *m, void *p)
>   	mutex_unlock(&module_mutex);
>   }
>   
> +static unsigned int module_total_size(struct module *mod)
> +{
> +	int size = 0;
> +
> +	for_each_mod_mem_type(type)
> +		size += mod->mem[type].size;
> +	return size;
> +}
> +
>   static int m_show(struct seq_file *m, void *p)
>   {
>   	struct module *mod = list_entry(p, struct module, list);
> @@ -73,10 +82,7 @@ static int m_show(struct seq_file *m, void *p)
>   	if (mod->state == MODULE_STATE_UNFORMED)
>   		return 0;
>   
> -	size = mod->init_layout.size + mod->core_layout.size;
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -	size += mod->data_layout.size;
> -#endif
> +	size = module_total_size(mod);
>   	seq_printf(m, "%s %u", mod->name, size);
>   	print_unload_info(m, mod);
>   
> @@ -86,7 +92,7 @@ static int m_show(struct seq_file *m, void *p)
>   		   mod->state == MODULE_STATE_COMING ? "Loading" :
>   		   "Live");
>   	/* Used by oprofile and other similar tools. */
> -	value = m->private ? NULL : mod->core_layout.base;
> +	value = m->private ? NULL : mod->mem[MOD_TEXT].base;
>   	seq_printf(m, " 0x%px", value);
>   
>   	/* Taints info */
> diff --git a/kernel/module/strict_rwx.c b/kernel/module/strict_rwx.c
> index 14fbea66f12f..2e79a77f40eb 100644
> --- a/kernel/module/strict_rwx.c
> +++ b/kernel/module/strict_rwx.c
> @@ -11,82 +11,26 @@
>   #include <linux/set_memory.h>
>   #include "internal.h"
>   
> -/*
> - * LKM RO/NX protection: protect module's text/ro-data
> - * from modification and any data from execution.
> - *
> - * General layout of module is:
> - *          [text] [read-only-data] [ro-after-init] [writable data]
> - * text_size -----^                ^               ^               ^
> - * ro_size ------------------------|               |               |
> - * ro_after_init_size -----------------------------|               |
> - * size -----------------------------------------------------------|
> - *
> - * These values are always page-aligned (as is base) when
> - * CONFIG_STRICT_MODULE_RWX is set.
> - */
> +static void module_set_memory(
> +	const struct module *mod, enum mod_mem_type type,
> +	int (*set_memory)(unsigned long start, int num_pages))
> +{
> +	const struct module_memory *mod_mem = &mod->mem[type];
> +
> +	set_vm_flush_reset_perms(mod_mem->base);
> +	set_memory((unsigned long)mod_mem->base, mod_mem->size >> PAGE_SHIFT);
> +}
>   
>   /*
>    * Since some arches are moving towards PAGE_KERNEL module allocations instead
> - * of PAGE_KERNEL_EXEC, keep frob_text() and module_enable_x() independent of
> + * of PAGE_KERNEL_EXEC, keep module_enable_x() independent of
>    * CONFIG_STRICT_MODULE_RWX because they are needed regardless of whether we
>    * are strict.
>    */
> -static void frob_text(const struct module_layout *layout,
> -		      int (*set_memory)(unsigned long start, int num_pages))
> -{
> -	set_memory((unsigned long)layout->base,
> -		   PAGE_ALIGN(layout->text_size) >> PAGE_SHIFT);
> -}
> -
> -static void frob_rodata(const struct module_layout *layout,
> -		 int (*set_memory)(unsigned long start, int num_pages))
> -{
> -	set_memory((unsigned long)layout->base + layout->text_size,
> -		   (layout->ro_size - layout->text_size) >> PAGE_SHIFT);
> -}
> -
> -static void frob_ro_after_init(const struct module_layout *layout,
> -			int (*set_memory)(unsigned long start, int num_pages))
> -{
> -	set_memory((unsigned long)layout->base + layout->ro_size,
> -		   (layout->ro_after_init_size - layout->ro_size) >> PAGE_SHIFT);
> -}
> -
> -static void frob_writable_data(const struct module_layout *layout,
> -			int (*set_memory)(unsigned long start, int num_pages))
> -{
> -	set_memory((unsigned long)layout->base + layout->ro_after_init_size,
> -		   (layout->size - layout->ro_after_init_size) >> PAGE_SHIFT);
> -}
> -
> -static bool layout_check_misalignment(const struct module_layout *layout)
> -{
> -	return WARN_ON(!PAGE_ALIGNED(layout->base)) ||
> -	       WARN_ON(!PAGE_ALIGNED(layout->text_size)) ||
> -	       WARN_ON(!PAGE_ALIGNED(layout->ro_size)) ||
> -	       WARN_ON(!PAGE_ALIGNED(layout->ro_after_init_size)) ||
> -	       WARN_ON(!PAGE_ALIGNED(layout->size));
> -}
> -
> -bool module_check_misalignment(const struct module *mod)
> -{
> -	if (!IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
> -		return false;
> -
> -	return layout_check_misalignment(&mod->core_layout) ||
> -	       layout_check_misalignment(&mod->data_layout) ||
> -	       layout_check_misalignment(&mod->init_layout);
> -}
> -
>   void module_enable_x(const struct module *mod)
>   {
> -	if (!PAGE_ALIGNED(mod->core_layout.base) ||
> -	    !PAGE_ALIGNED(mod->init_layout.base))
> -		return;
> -
> -	frob_text(&mod->core_layout, set_memory_x);
> -	frob_text(&mod->init_layout, set_memory_x);
> +	for_class_mod_mem_type(type, text)
> +		module_set_memory(mod, type, set_memory_x);
>   }
>   
>   void module_enable_ro(const struct module *mod, bool after_init)
> @@ -98,16 +42,13 @@ void module_enable_ro(const struct module *mod, bool after_init)
>   		return;
>   #endif
>   
> -	set_vm_flush_reset_perms(mod->core_layout.base);
> -	set_vm_flush_reset_perms(mod->init_layout.base);
> -	frob_text(&mod->core_layout, set_memory_ro);
> -
> -	frob_rodata(&mod->data_layout, set_memory_ro);
> -	frob_text(&mod->init_layout, set_memory_ro);
> -	frob_rodata(&mod->init_layout, set_memory_ro);
> +	module_set_memory(mod, MOD_TEXT, set_memory_ro);
> +	module_set_memory(mod, MOD_INIT_TEXT, set_memory_ro);
> +	module_set_memory(mod, MOD_RODATA, set_memory_ro);
> +	module_set_memory(mod, MOD_INIT_RODATA, set_memory_ro);
>   
>   	if (after_init)
> -		frob_ro_after_init(&mod->data_layout, set_memory_ro);
> +		module_set_memory(mod, MOD_RO_AFTER_INIT, set_memory_ro);
>   }
>   
>   void module_enable_nx(const struct module *mod)
> @@ -115,11 +56,8 @@ void module_enable_nx(const struct module *mod)
>   	if (!IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
>   		return;
>   
> -	frob_rodata(&mod->data_layout, set_memory_nx);
> -	frob_ro_after_init(&mod->data_layout, set_memory_nx);
> -	frob_writable_data(&mod->data_layout, set_memory_nx);
> -	frob_rodata(&mod->init_layout, set_memory_nx);
> -	frob_writable_data(&mod->init_layout, set_memory_nx);
> +	for_class_mod_mem_type(type, data)
> +		module_set_memory(mod, type, set_memory_nx);
>   }
>   
>   int module_enforce_rwx_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
> diff --git a/kernel/module/tree_lookup.c b/kernel/module/tree_lookup.c
> index 8ec5cfd60496..277197977d43 100644
> --- a/kernel/module/tree_lookup.c
> +++ b/kernel/module/tree_lookup.c
> @@ -21,16 +21,16 @@
>   
>   static __always_inline unsigned long __mod_tree_val(struct latch_tree_node *n)
>   {
> -	struct module_layout *layout = container_of(n, struct module_layout, mtn.node);
> +	struct module_memory *mod_mem = container_of(n, struct module_memory, mtn.node);
>   
> -	return (unsigned long)layout->base;
> +	return (unsigned long)mod_mem->base;
>   }
>   
>   static __always_inline unsigned long __mod_tree_size(struct latch_tree_node *n)
>   {
> -	struct module_layout *layout = container_of(n, struct module_layout, mtn.node);
> +	struct module_memory *mod_mem = container_of(n, struct module_memory, mtn.node);
>   
> -	return (unsigned long)layout->size;
> +	return (unsigned long)mod_mem->size;
>   }
>   
>   static __always_inline bool
> @@ -77,32 +77,27 @@ static void __mod_tree_remove(struct mod_tree_node *node, struct mod_tree_root *
>    */
>   void mod_tree_insert(struct module *mod)
>   {
> -	mod->core_layout.mtn.mod = mod;
> -	mod->init_layout.mtn.mod = mod;
> -
> -	__mod_tree_insert(&mod->core_layout.mtn, &mod_tree);
> -	if (mod->init_layout.size)
> -		__mod_tree_insert(&mod->init_layout.mtn, &mod_tree);
> -
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -	mod->data_layout.mtn.mod = mod;
> -	__mod_tree_insert(&mod->data_layout.mtn, &mod_data_tree);
> -#endif
> +	for_each_mod_mem_type(type) {
> +		mod->mem[type].mtn.mod = mod;
> +		if (mod->mem[type].size)
> +			__mod_tree_insert(&mod->mem[type].mtn, &mod_tree);
> +	}
>   }
>   
>   void mod_tree_remove_init(struct module *mod)
>   {
> -	if (mod->init_layout.size)
> -		__mod_tree_remove(&mod->init_layout.mtn, &mod_tree);
> +	for_class_mod_mem_type(type, init) {
> +		if (mod->mem[type].size)
> +			__mod_tree_remove(&mod->mem[type].mtn, &mod_tree);
> +	}
>   }
>   
>   void mod_tree_remove(struct module *mod)
>   {
> -	__mod_tree_remove(&mod->core_layout.mtn, &mod_tree);
> -	mod_tree_remove_init(mod);
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -	__mod_tree_remove(&mod->data_layout.mtn, &mod_data_tree);
> -#endif
> +	for_each_mod_mem_type(type) {
> +		if (mod->mem[type].size)
> +			__mod_tree_remove(&mod->mem[type].mtn, &mod_tree);
> +	}
>   }
>   
>   struct module *mod_find(unsigned long addr, struct mod_tree_root *tree)
Song Liu Feb. 1, 2023, 6:41 p.m. UTC | #2
On Wed, Feb 1, 2023 at 10:35 AM Christophe Leroy
<christophe.leroy@csgroup.eu> wrote:
>
> Hi,
>
> On what tree / branch does this patch applies ?
>
> I have tried with modules/modules-next, linus/master, with v6.1. It
> conflicts on all of them.
>
> 'git am -3' even says that 'It does not apply to blobs recorded in its
> index.'

Hmmm... I was using "git format-patch -b" to skip space changes. I
guess I should use --ignore-space-at-eol instead.

Resending it.

Thanks,
Song

PS: it was based on modules-next.

>
> Christophe
>
> Le 01/02/2023 à 18:36, Song Liu a écrit :
> > module_layout manages different types of memory (text, data, rodata, etc.)
> > in one allocation, which is problematic for some reasons:
> >
> > 1. It is hard to enable CONFIG_STRICT_MODULE_RWX.
> > 2. It is hard to use huge pages in modules (and not break strict rwx).
> > 3. Many archs uses module_layout for arch-specific data, but it is not
> >     obvious how these data are used (are they RO, RX, or RW?)
> >
> > Improve the scenario by replacing 2 (or 3) module_layout per module with
> > up to 7 module_memory per module:
> >
> >          MOD_TEXT,
> >          MOD_DATA,
> >          MOD_RODATA,
> >          MOD_RO_AFTER_INIT,
> >          MOD_INIT_TEXT,
> >          MOD_INIT_DATA,
> >          MOD_INIT_RODATA,
> >
> > and allocating them separately. This adds slightly more entries to
> > mod_tree (from up to 3 entries per module, to up to 7 entries per
> > module). However, this at most adds a small constant overhead to
> > __module_address(), which is expected to be fast.
> >
> > Various archs use module_layout for different data. These data are put
> > into different module_memory based on their location in module_layout.
> > IOW, data that used to go with text is allocated with MOD_MEM_TYPE_TEXT;
> > data that used to go with data is allocated with MOD_MEM_TYPE_DATA, etc.
> >
> > module_memory simplifies quite some of the module code. For example,
> > ARCH_WANTS_MODULES_DATA_IN_VMALLOC is a lot cleaner, as it just uses a
> > different allocator for the data. kernel/module/strict_rwx.c is also
> > much cleaner with module_memory.
> >
> > Signed-off-by: Song Liu <song@kernel.org>
> > Cc: Luis Chamberlain <mcgrof@kernel.org>
> > Cc: Thomas Gleixner <tglx@linutronix.de>
> > Cc: Peter Zijlstra <peterz@infradead.org>
> > Cc: Guenter Roeck <linux@roeck-us.net>
> > Cc: Christophe Leroy <christophe.leroy@csgroup.eu>
> >
> > ---
> >
> > This is the preparation work for the type aware module_alloc() discussed
> > in [1]. While this work is not covered much in the discussion, it is a
> > critical step of the effort.
> >
> > As this part grows pretty big (~1000 lines, + and -), I would like get
> > some feedback on it, so that I know it is on the right track.
> >
> > Please share your comments. Thanks!
> >
> > Test coverage: Tested on x86_64.
> > Build tested by kernel test bot in [2]. The only regression in [2] was a
> > typo in parisc, which is also fixed.
> >
> > [1] https://lore.kernel.org/linux-mm/20221107223921.3451913-1-song@kernel.org/T/#u
> > [2] https://lore.kernel.org/linux-raid/63b8827e.clJQX2wg+I+tiX7m%25lkp@intel.com/T/#u
> >
> > Changes v1 => v2:
> > 1. Add data_addr_[min|max] for CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> >     case.
> >
> > Changes v2 => v3:
> > 1. Fix and remove the last use of INIT_OFFSET_MASK.
> > 2. Add more information in the commit log. (Luis Chamberlain)
> > 3. Rebase and fix issues in x86/calltrunks.
> > 4. Minor cleanup.
> >
> > Changes v3 => v4:
> > 1. Shorten enum/variable names, so that the code are easier to read.
> >     (Christophe Leroy)
> > 2. Remove an used variable. (Guenter Roeck, Christophe Leroy)
> >
> > Changes v4 => v5:
> > 1. Simplify some code some code. (Peter Zijlstra, Christophe Leroy)
> > 2. Remove module_check_misalignment(), which is not useful any more.
> >
> > Changes v5 => v6:
> > 1. Improve mod_mem_type_is_* and for_*mod_mem_type marcos.
> >     (Peter Zijlstra).
> > ---
> >   arch/alpha/kernel/module.c      |   2 +-
> >   arch/arc/kernel/unwind.c        |   9 +-
> >   arch/arm/kernel/module-plts.c   |   2 +-
> >   arch/arm64/kernel/module-plts.c |   2 +-
> >   arch/ia64/kernel/module.c       |  24 ++-
> >   arch/mips/kernel/vpe.c          |  11 +-
> >   arch/parisc/kernel/module.c     |  51 ++---
> >   arch/powerpc/kernel/module_32.c |   7 +-
> >   arch/s390/kernel/module.c       |  26 +--
> >   arch/x86/kernel/callthunks.c    |   4 +-
> >   arch/x86/kernel/module.c        |   4 +-
> >   include/linux/module.h          |  86 +++++---
> >   kernel/module/internal.h        |  40 ++--
> >   kernel/module/kallsyms.c        |  56 ++---
> >   kernel/module/kdb.c             |  17 +-
> >   kernel/module/main.c            | 355 ++++++++++++++++++--------------
> >   kernel/module/procfs.c          |  16 +-
> >   kernel/module/strict_rwx.c      | 100 ++-------
> >   kernel/module/tree_lookup.c     |  39 ++--
> >   19 files changed, 432 insertions(+), 419 deletions(-)
> >
> > diff --git a/arch/alpha/kernel/module.c b/arch/alpha/kernel/module.c
> > index 5b60c248de9e..9109213abc09 100644
> > --- a/arch/alpha/kernel/module.c
> > +++ b/arch/alpha/kernel/module.c
> > @@ -148,7 +148,7 @@ apply_relocate_add(Elf64_Shdr *sechdrs, const char *strtab,
> >
> >       /* The small sections were sorted to the end of the segment.
> >          The following should definitely cover them.  */
> > -     gp = (u64)me->core_layout.base + me->core_layout.size - 0x8000;
> > +     gp = (u64)me->mem[MOD_DATA].base + me->mem[MOD_DATA].size - 0x8000;
> >       got = sechdrs[me->arch.gotsecindex].sh_addr;
> >
> >       for (i = 0; i < n; i++) {
> > diff --git a/arch/arc/kernel/unwind.c b/arch/arc/kernel/unwind.c
> > index 200270a94558..933451f4494f 100644
> > --- a/arch/arc/kernel/unwind.c
> > +++ b/arch/arc/kernel/unwind.c
> > @@ -369,6 +369,8 @@ void *unwind_add_table(struct module *module, const void *table_start,
> >                      unsigned long table_size)
> >   {
> >       struct unwind_table *table;
> > +     struct module_memory *mod_mem_core_text;
> > +     struct module_memory *mod_mem_init_text;
> >
> >       if (table_size <= 0)
> >               return NULL;
> > @@ -377,9 +379,12 @@ void *unwind_add_table(struct module *module, const void *table_start,
> >       if (!table)
> >               return NULL;
> >
> > +     mod_mem_core_text = &module->mem[MOD_TEXT];
> > +     mod_mem_init_text = &module->mem[MOD_INIT_TEXT];
> > +
> >       init_unwind_table(table, module->name,
> > -                       module->core_layout.base, module->core_layout.size,
> > -                       module->init_layout.base, module->init_layout.size,
> > +                       mod_mem_core_text->base, mod_mem_core_text->size,
> > +                       mod_mem_init_text->base, mod_mem_init_text->size,
> >                         table_start, table_size,
> >                         NULL, 0);
> >
> > diff --git a/arch/arm/kernel/module-plts.c b/arch/arm/kernel/module-plts.c
> > index af7c322ebed6..9d4ecb6b1412 100644
> > --- a/arch/arm/kernel/module-plts.c
> > +++ b/arch/arm/kernel/module-plts.c
> > @@ -30,7 +30,7 @@ static const u32 fixed_plts[] = {
> >
> >   static bool in_init(const struct module *mod, unsigned long loc)
> >   {
> > -     return loc - (u32)mod->init_layout.base < mod->init_layout.size;
> > +     return within_module_init(loc, mod);
> >   }
> >
> >   static void prealloc_fixed(struct mod_plt_sec *pltsec, struct plt_entries *plt)
> > diff --git a/arch/arm64/kernel/module-plts.c b/arch/arm64/kernel/module-plts.c
> > index 5a0a8f552a61..fedffa83a283 100644
> > --- a/arch/arm64/kernel/module-plts.c
> > +++ b/arch/arm64/kernel/module-plts.c
> > @@ -67,7 +67,7 @@ static bool plt_entries_equal(const struct plt_entry *a,
> >
> >   static bool in_init(const struct module *mod, void *loc)
> >   {
> > -     return (u64)loc - (u64)mod->init_layout.base < mod->init_layout.size;
> > +     return within_module_init(loc, mod);
> >   }
> >
> >   u64 module_emit_plt_entry(struct module *mod, Elf64_Shdr *sechdrs,
> > diff --git a/arch/ia64/kernel/module.c b/arch/ia64/kernel/module.c
> > index 8f62cf97f691..3661135da9d9 100644
> > --- a/arch/ia64/kernel/module.c
> > +++ b/arch/ia64/kernel/module.c
> > @@ -485,19 +485,19 @@ module_frob_arch_sections (Elf_Ehdr *ehdr, Elf_Shdr *sechdrs, char *secstrings,
> >       return 0;
> >   }
> >
> > -static inline int
> > +static inline bool
> >   in_init (const struct module *mod, uint64_t addr)
> >   {
> > -     return addr - (uint64_t) mod->init_layout.base < mod->init_layout.size;
> > +     return within_module_init(addr, mod);
> >   }
> >
> > -static inline int
> > +static inline bool
> >   in_core (const struct module *mod, uint64_t addr)
> >   {
> > -     return addr - (uint64_t) mod->core_layout.base < mod->core_layout.size;
> > +     return within_module_core(addr, mod);
> >   }
> >
> > -static inline int
> > +static inline bool
> >   is_internal (const struct module *mod, uint64_t value)
> >   {
> >       return in_init(mod, value) || in_core(mod, value);
> > @@ -677,7 +677,8 @@ do_reloc (struct module *mod, uint8_t r_type, Elf64_Sym *sym, uint64_t addend,
> >               break;
> >
> >             case RV_BDREL:
> > -             val -= (uint64_t) (in_init(mod, val) ? mod->init_layout.base : mod->core_layout.base);
> > +             val -= (uint64_t) (in_init(mod, val) ? mod->mem[MOD_INIT_TEXT].base :
> > +                                mod->mem[MOD_TEXT].base);
> >               break;
> >
> >             case RV_LTV:
> > @@ -812,15 +813,18 @@ apply_relocate_add (Elf64_Shdr *sechdrs, const char *strtab, unsigned int symind
> >                *     addresses have been selected...
> >                */
> >               uint64_t gp;
> > -             if (mod->core_layout.size > MAX_LTOFF)
> > +             struct module_memory *mod_mem;
> > +
> > +             mod_mem = &mod->mem[MOD_DATA];
> > +             if (mod_mem->size > MAX_LTOFF)
> >                       /*
> >                        * This takes advantage of fact that SHF_ARCH_SMALL gets allocated
> >                        * at the end of the module.
> >                        */
> > -                     gp = mod->core_layout.size - MAX_LTOFF / 2;
> > +                     gp = mod_mem->size - MAX_LTOFF / 2;
> >               else
> > -                     gp = mod->core_layout.size / 2;
> > -             gp = (uint64_t) mod->core_layout.base + ((gp + 7) & -8);
> > +                     gp = mod_mem->size / 2;
> > +             gp = (uint64_t) mod_mem->base + ((gp + 7) & -8);
> >               mod->arch.gp = gp;
> >               DEBUGP("%s: placing gp at 0x%lx\n", __func__, gp);
> >       }
> > diff --git a/arch/mips/kernel/vpe.c b/arch/mips/kernel/vpe.c
> > index 13294972707b..ab114c686f9d 100644
> > --- a/arch/mips/kernel/vpe.c
> > +++ b/arch/mips/kernel/vpe.c
> > @@ -199,18 +199,17 @@ static void layout_sections(struct module *mod, const Elf_Ehdr *hdr,
> >       for (m = 0; m < ARRAY_SIZE(masks); ++m) {
> >               for (i = 0; i < hdr->e_shnum; ++i) {
> >                       Elf_Shdr *s = &sechdrs[i];
> > +                     struct module_memory *mod_mem;
> > +
> > +                     mod_mem = &mod->mem[MOD_TEXT];
> >
> >                       if ((s->sh_flags & masks[m][0]) != masks[m][0]
> >                           || (s->sh_flags & masks[m][1])
> >                           || s->sh_entsize != ~0UL)
> >                               continue;
> >                       s->sh_entsize =
> > -                             get_offset((unsigned long *)&mod->core_layout.size, s);
> > +                             get_offset((unsigned long *)&mod_mem->size, s);
> >               }
> > -
> > -             if (m == 0)
> > -                     mod->core_layout.text_size = mod->core_layout.size;
> > -
> >       }
> >   }
> >
> > @@ -641,7 +640,7 @@ static int vpe_elfload(struct vpe *v)
> >               layout_sections(&mod, hdr, sechdrs, secstrings);
> >       }
> >
> > -     v->load_addr = alloc_progmem(mod.core_layout.size);
> > +     v->load_addr = alloc_progmem(mod.mod_mem[MOD_TEXT].size);
> >       if (!v->load_addr)
> >               return -ENOMEM;
> >
> > diff --git a/arch/parisc/kernel/module.c b/arch/parisc/kernel/module.c
> > index 7df140545b22..f6e38c4d3904 100644
> > --- a/arch/parisc/kernel/module.c
> > +++ b/arch/parisc/kernel/module.c
> > @@ -27,9 +27,9 @@
> >    *      We are not doing SEGREL32 handling correctly. According to the ABI, we
> >    *      should do a value offset, like this:
> >    *                  if (in_init(me, (void *)val))
> > - *                           val -= (uint32_t)me->init_layout.base;
> > + *                           val -= (uint32_t)me->mem[MOD_INIT_TEXT].base;
> >    *                  else
> > - *                           val -= (uint32_t)me->core_layout.base;
> > + *                           val -= (uint32_t)me->mem[MOD_TEXT].base;
> >    *  However, SEGREL32 is used only for PARISC unwind entries, and we want
> >    *  those entries to have an absolute address, and not just an offset.
> >    *
> > @@ -76,25 +76,6 @@
> >    * allows us to allocate up to 4095 GOT entries. */
> >   #define MAX_GOTS    4095
> >
> > -/* three functions to determine where in the module core
> > - * or init pieces the location is */
> > -static inline int in_init(struct module *me, void *loc)
> > -{
> > -     return (loc >= me->init_layout.base &&
> > -             loc <= (me->init_layout.base + me->init_layout.size));
> > -}
> > -
> > -static inline int in_core(struct module *me, void *loc)
> > -{
> > -     return (loc >= me->core_layout.base &&
> > -             loc <= (me->core_layout.base + me->core_layout.size));
> > -}
> > -
> > -static inline int in_local(struct module *me, void *loc)
> > -{
> > -     return in_init(me, loc) || in_core(me, loc);
> > -}
> > -
> >   #ifndef CONFIG_64BIT
> >   struct got_entry {
> >       Elf32_Addr addr;
> > @@ -302,6 +283,7 @@ int module_frob_arch_sections(CONST Elf_Ehdr *hdr,
> >   {
> >       unsigned long gots = 0, fdescs = 0, len;
> >       unsigned int i;
> > +     struct module_memory *mod_mem;
> >
> >       len = hdr->e_shnum * sizeof(me->arch.section[0]);
> >       me->arch.section = kzalloc(len, GFP_KERNEL);
> > @@ -346,14 +328,15 @@ int module_frob_arch_sections(CONST Elf_Ehdr *hdr,
> >               me->arch.section[s].stub_entries += count;
> >       }
> >
> > +     mod_mem = &me->mem[MOD_TEXT];
> >       /* align things a bit */
> > -     me->core_layout.size = ALIGN(me->core_layout.size, 16);
> > -     me->arch.got_offset = me->core_layout.size;
> > -     me->core_layout.size += gots * sizeof(struct got_entry);
> > +     mod_mem->size = ALIGN(mod_mem->size, 16);
> > +     me->arch.got_offset = mod_mem->size;
> > +     mod_mem->size += gots * sizeof(struct got_entry);
> >
> > -     me->core_layout.size = ALIGN(me->core_layout.size, 16);
> > -     me->arch.fdesc_offset = me->core_layout.size;
> > -     me->core_layout.size += fdescs * sizeof(Elf_Fdesc);
> > +     mod_mem->size = ALIGN(mod_mem->size, 16);
> > +     me->arch.fdesc_offset = mod_mem->size;
> > +     mod_mem->size += fdescs * sizeof(Elf_Fdesc);
> >
> >       me->arch.got_max = gots;
> >       me->arch.fdesc_max = fdescs;
> > @@ -371,7 +354,7 @@ static Elf64_Word get_got(struct module *me, unsigned long value, long addend)
> >
> >       BUG_ON(value == 0);
> >
> > -     got = me->core_layout.base + me->arch.got_offset;
> > +     got = me->mem[MOD_TEXT].base + me->arch.got_offset;
> >       for (i = 0; got[i].addr; i++)
> >               if (got[i].addr == value)
> >                       goto out;
> > @@ -389,7 +372,7 @@ static Elf64_Word get_got(struct module *me, unsigned long value, long addend)
> >   #ifdef CONFIG_64BIT
> >   static Elf_Addr get_fdesc(struct module *me, unsigned long value)
> >   {
> > -     Elf_Fdesc *fdesc = me->core_layout.base + me->arch.fdesc_offset;
> > +     Elf_Fdesc *fdesc = me->mem[MOD_TEXT].base + me->arch.fdesc_offset;
> >
> >       if (!value) {
> >               printk(KERN_ERR "%s: zero OPD requested!\n", me->name);
> > @@ -407,7 +390,7 @@ static Elf_Addr get_fdesc(struct module *me, unsigned long value)
> >
> >       /* Create new one */
> >       fdesc->addr = value;
> > -     fdesc->gp = (Elf_Addr)me->core_layout.base + me->arch.got_offset;
> > +     fdesc->gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
> >       return (Elf_Addr)fdesc;
> >   }
> >   #endif /* CONFIG_64BIT */
> > @@ -742,7 +725,7 @@ int apply_relocate_add(Elf_Shdr *sechdrs,
> >                              loc, val);
> >                       val += addend;
> >                       /* can we reach it locally? */
> > -                     if (in_local(me, (void *)val)) {
> > +                     if (within_module(val, me)) {
> >                               /* this is the case where the symbol is local
> >                                * to the module, but in a different section,
> >                                * so stub the jump in case it's more than 22
> > @@ -801,7 +784,7 @@ int apply_relocate_add(Elf_Shdr *sechdrs,
> >                       break;
> >               case R_PARISC_FPTR64:
> >                       /* 64-bit function address */
> > -                     if(in_local(me, (void *)(val + addend))) {
> > +                     if (within_module(val + addend, me)) {
> >                               *loc64 = get_fdesc(me, val+addend);
> >                               pr_debug("FDESC for %s at %llx points to %llx\n",
> >                                      strtab + sym->st_name, *loc64,
> > @@ -839,7 +822,7 @@ register_unwind_table(struct module *me,
> >
> >       table = (unsigned char *)sechdrs[me->arch.unwind_section].sh_addr;
> >       end = table + sechdrs[me->arch.unwind_section].sh_size;
> > -     gp = (Elf_Addr)me->core_layout.base + me->arch.got_offset;
> > +     gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
> >
> >       pr_debug("register_unwind_table(), sect = %d at 0x%p - 0x%p (gp=0x%lx)\n",
> >              me->arch.unwind_section, table, end, gp);
> > @@ -977,7 +960,7 @@ void module_arch_cleanup(struct module *mod)
> >   #ifdef CONFIG_64BIT
> >   void *dereference_module_function_descriptor(struct module *mod, void *ptr)
> >   {
> > -     unsigned long start_opd = (Elf64_Addr)mod->core_layout.base +
> > +     unsigned long start_opd = (Elf64_Addr)mod->mem[MOD_TEXT].base +
> >                                  mod->arch.fdesc_offset;
> >       unsigned long end_opd = start_opd +
> >                               mod->arch.fdesc_count * sizeof(Elf64_Fdesc);
> > diff --git a/arch/powerpc/kernel/module_32.c b/arch/powerpc/kernel/module_32.c
> > index ea6536171778..2729179e9dc0 100644
> > --- a/arch/powerpc/kernel/module_32.c
> > +++ b/arch/powerpc/kernel/module_32.c
> > @@ -163,8 +163,7 @@ static uint32_t do_plt_call(void *location,
> >
> >       pr_debug("Doing plt for call to 0x%x at 0x%x\n", val, (unsigned int)location);
> >       /* Init, or core PLT? */
> > -     if (location >= mod->core_layout.base
> > -         && location < mod->core_layout.base + mod->core_layout.size)
> > +     if (within_module_init(location, mod))
> >               entry = (void *)sechdrs[mod->arch.core_plt_section].sh_addr;
> >       else
> >               entry = (void *)sechdrs[mod->arch.init_plt_section].sh_addr;
> > @@ -322,14 +321,14 @@ notrace int module_trampoline_target(struct module *mod, unsigned long addr,
> >
> >   int module_finalize_ftrace(struct module *module, const Elf_Shdr *sechdrs)
> >   {
> > -     module->arch.tramp = do_plt_call(module->core_layout.base,
> > +     module->arch.tramp = do_plt_call(module->mem[MOD_TEXT].base,
> >                                        (unsigned long)ftrace_caller,
> >                                        sechdrs, module);
> >       if (!module->arch.tramp)
> >               return -ENOENT;
> >
> >   #ifdef CONFIG_DYNAMIC_FTRACE_WITH_REGS
> > -     module->arch.tramp_regs = do_plt_call(module->core_layout.base,
> > +     module->arch.tramp_regs = do_plt_call(module->mem[MOD_TEXT].base,
> >                                             (unsigned long)ftrace_regs_caller,
> >                                             sechdrs, module);
> >       if (!module->arch.tramp_regs)
> > diff --git a/arch/s390/kernel/module.c b/arch/s390/kernel/module.c
> > index 2d159b32885b..3ed455b785ad 100644
> > --- a/arch/s390/kernel/module.c
> > +++ b/arch/s390/kernel/module.c
> > @@ -126,6 +126,7 @@ int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
> >       Elf_Rela *rela;
> >       char *strings;
> >       int nrela, i, j;
> > +     struct module_memory *mod_mem;
> >
> >       /* Find symbol table and string table. */
> >       symtab = NULL;
> > @@ -173,14 +174,15 @@ int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
> >
> >       /* Increase core size by size of got & plt and set start
> >          offsets for got and plt. */
> > -     me->core_layout.size = ALIGN(me->core_layout.size, 4);
> > -     me->arch.got_offset = me->core_layout.size;
> > -     me->core_layout.size += me->arch.got_size;
> > -     me->arch.plt_offset = me->core_layout.size;
> > +     mod_mem = &me->mem[MOD_TEXT];
> > +     mod_mem->size = ALIGN(mod_mem->size, 4);
> > +     me->arch.got_offset = mod_mem->size;
> > +     mod_mem->size += me->arch.got_size;
> > +     me->arch.plt_offset = mod_mem->size;
> >       if (me->arch.plt_size) {
> >               if (IS_ENABLED(CONFIG_EXPOLINE) && !nospec_disable)
> >                       me->arch.plt_size += PLT_ENTRY_SIZE;
> > -             me->core_layout.size += me->arch.plt_size;
> > +             mod_mem->size += me->arch.plt_size;
> >       }
> >       return 0;
> >   }
> > @@ -304,7 +306,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> >       case R_390_GOTPLT64:    /* 64 bit offset to jump slot.  */
> >       case R_390_GOTPLTENT:   /* 32 bit rel. offset to jump slot >> 1. */
> >               if (info->got_initialized == 0) {
> > -                     Elf_Addr *gotent = me->core_layout.base +
> > +                     Elf_Addr *gotent = me->mem[MOD_TEXT].base +
> >                                          me->arch.got_offset +
> >                                          info->got_offset;
> >
> > @@ -329,7 +331,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> >                       rc = apply_rela_bits(loc, val, 0, 64, 0, write);
> >               else if (r_type == R_390_GOTENT ||
> >                        r_type == R_390_GOTPLTENT) {
> > -                     val += (Elf_Addr) me->core_layout.base - loc;
> > +                     val += (Elf_Addr) me->mem[MOD_TEXT].base - loc;
> >                       rc = apply_rela_bits(loc, val, 1, 32, 1, write);
> >               }
> >               break;
> > @@ -345,7 +347,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> >                       char *plt_base;
> >                       char *ip;
> >
> > -                     plt_base = me->core_layout.base + me->arch.plt_offset;
> > +                     plt_base = me->mem[MOD_TEXT].base + me->arch.plt_offset;
> >                       ip = plt_base + info->plt_offset;
> >                       *(int *)insn = 0x0d10e310;      /* basr 1,0  */
> >                       *(int *)&insn[4] = 0x100c0004;  /* lg   1,12(1) */
> > @@ -375,7 +377,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> >                              val - loc + 0xffffUL < 0x1ffffeUL) ||
> >                             (r_type == R_390_PLT32DBL &&
> >                              val - loc + 0xffffffffULL < 0x1fffffffeULL)))
> > -                             val = (Elf_Addr) me->core_layout.base +
> > +                             val = (Elf_Addr) me->mem[MOD_TEXT].base +
> >                                       me->arch.plt_offset +
> >                                       info->plt_offset;
> >                       val += rela->r_addend - loc;
> > @@ -397,7 +399,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> >       case R_390_GOTOFF32:    /* 32 bit offset to GOT.  */
> >       case R_390_GOTOFF64:    /* 64 bit offset to GOT. */
> >               val = val + rela->r_addend -
> > -                     ((Elf_Addr) me->core_layout.base + me->arch.got_offset);
> > +                     ((Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset);
> >               if (r_type == R_390_GOTOFF16)
> >                       rc = apply_rela_bits(loc, val, 0, 16, 0, write);
> >               else if (r_type == R_390_GOTOFF32)
> > @@ -407,7 +409,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> >               break;
> >       case R_390_GOTPC:       /* 32 bit PC relative offset to GOT. */
> >       case R_390_GOTPCDBL:    /* 32 bit PC rel. off. to GOT shifted by 1. */
> > -             val = (Elf_Addr) me->core_layout.base + me->arch.got_offset +
> > +             val = (Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset +
> >                       rela->r_addend - loc;
> >               if (r_type == R_390_GOTPC)
> >                       rc = apply_rela_bits(loc, val, 1, 32, 0, write);
> > @@ -515,7 +517,7 @@ int module_finalize(const Elf_Ehdr *hdr,
> >           !nospec_disable && me->arch.plt_size) {
> >               unsigned int *ij;
> >
> > -             ij = me->core_layout.base + me->arch.plt_offset +
> > +             ij = me->mem[MOD_TEXT].base + me->arch.plt_offset +
> >                       me->arch.plt_size - PLT_ENTRY_SIZE;
> >               ij[0] = 0xc6000000;     /* exrl %r0,.+10        */
> >               ij[1] = 0x0005a7f4;     /* j    .               */
> > diff --git a/arch/x86/kernel/callthunks.c b/arch/x86/kernel/callthunks.c
> > index ffea98f9064b..22ab13966427 100644
> > --- a/arch/x86/kernel/callthunks.c
> > +++ b/arch/x86/kernel/callthunks.c
> > @@ -330,8 +330,8 @@ void noinline callthunks_patch_module_calls(struct callthunk_sites *cs,
> >                                           struct module *mod)
> >   {
> >       struct core_text ct = {
> > -             .base = (unsigned long)mod->core_layout.base,
> > -             .end  = (unsigned long)mod->core_layout.base + mod->core_layout.size,
> > +             .base = (unsigned long)mod->mem[MOD_TEXT].base,
> > +             .end  = (unsigned long)mod->mem[MOD_TEXT].base + mod->mem[MOD_TEXT].size,
> >               .name = mod->name,
> >       };
> >
> > diff --git a/arch/x86/kernel/module.c b/arch/x86/kernel/module.c
> > index 705fb2a41d7d..dbc8e226c464 100644
> > --- a/arch/x86/kernel/module.c
> > +++ b/arch/x86/kernel/module.c
> > @@ -339,8 +339,8 @@ int module_finalize(const Elf_Ehdr *hdr,
> >       }
> >       if (locks) {
> >               void *lseg = (void *)locks->sh_addr;
> > -             void *text = me->core_layout.base;
> > -             void *text_end = text + me->core_layout.text_size;
> > +             void *text = me->mem[MOD_TEXT].base;
> > +             void *text_end = text + me->mem[MOD_TEXT].size;
> >               alternatives_smp_module_add(me, me->name,
> >                                           lseg, lseg + locks->sh_size,
> >                                           text, text_end);
> > diff --git a/include/linux/module.h b/include/linux/module.h
> > index 8c5909c0076c..ccc9de8a93db 100644
> > --- a/include/linux/module.h
> > +++ b/include/linux/module.h
> > @@ -320,17 +320,47 @@ struct mod_tree_node {
> >       struct latch_tree_node node;
> >   };
> >
> > -struct module_layout {
> > -     /* The actual code + data. */
> > +enum mod_mem_type {
> > +     MOD_TEXT = 0,
> > +     MOD_DATA,
> > +     MOD_RODATA,
> > +     MOD_RO_AFTER_INIT,
> > +     MOD_INIT_TEXT,
> > +     MOD_INIT_DATA,
> > +     MOD_INIT_RODATA,
> > +
> > +     MOD_MEM_NUM_TYPES,
> > +     MOD_INVALID = -1,
> > +};
> > +
> > +#define mod_mem_type_is_init(type)   \
> > +     ((type) == MOD_INIT_TEXT ||     \
> > +      (type) == MOD_INIT_DATA ||     \
> > +      (type) == MOD_INIT_RODATA)
> > +
> > +#define mod_mem_type_is_core(type) (!mod_mem_type_is_init(type))
> > +
> > +#define mod_mem_type_is_text(type)   \
> > +      ((type) == MOD_TEXT ||         \
> > +       (type) == MOD_INIT_TEXT)
> > +
> > +#define mod_mem_type_is_data(type) (!mod_mem_type_is_text(type))
> > +
> > +#define mod_mem_type_is_core_data(type)      \
> > +     (mod_mem_type_is_core(type) &&  \
> > +      mod_mem_type_is_data(type))
> > +
> > +#define for_each_mod_mem_type(type)                  \
> > +     for (enum mod_mem_type (type) = 0;              \
> > +          (type) < MOD_MEM_NUM_TYPES; (type)++)
> > +
> > +#define for_class_mod_mem_type(type, class)          \
> > +     for_each_mod_mem_type(type)                     \
> > +             if (mod_mem_type_is_##class(type))
> > +
> > +struct module_memory {
> >       void *base;
> > -     /* Total size. */
> >       unsigned int size;
> > -     /* The size of the executable code.  */
> > -     unsigned int text_size;
> > -     /* Size of RO section of the module (text+rodata) */
> > -     unsigned int ro_size;
> > -     /* Size of RO after init section */
> > -     unsigned int ro_after_init_size;
> >
> >   #ifdef CONFIG_MODULES_TREE_LOOKUP
> >       struct mod_tree_node mtn;
> > @@ -339,9 +369,9 @@ struct module_layout {
> >
> >   #ifdef CONFIG_MODULES_TREE_LOOKUP
> >   /* Only touch one cacheline for common rbtree-for-core-layout case. */
> > -#define __module_layout_align ____cacheline_aligned
> > +#define __module_memory_align ____cacheline_aligned
> >   #else
> > -#define __module_layout_align
> > +#define __module_memory_align
> >   #endif
> >
> >   struct mod_kallsyms {
> > @@ -418,12 +448,8 @@ struct module {
> >       /* Startup function. */
> >       int (*init)(void);
> >
> > -     /* Core layout: rbtree is accessed frequently, so keep together. */
> > -     struct module_layout core_layout __module_layout_align;
> > -     struct module_layout init_layout;
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -     struct module_layout data_layout;
> > -#endif
> > +     /* rbtree is accessed frequently, so keep together. */
> > +     struct module_memory mem[MOD_MEM_NUM_TYPES] __module_memory_align;
> >
> >       /* Arch-specific module values */
> >       struct mod_arch_specific arch;
> > @@ -573,23 +599,33 @@ bool __is_module_percpu_address(unsigned long addr, unsigned long *can_addr);
> >   bool is_module_percpu_address(unsigned long addr);
> >   bool is_module_text_address(unsigned long addr);
> >
> > +static inline bool within_module_mem_type(unsigned long addr,
> > +                                       const struct module *mod,
> > +                                       enum mod_mem_type type)
> > +{
> > +     unsigned long base, size;
> > +
> > +     base = (unsigned long)mod->mem[type].base;
> > +     size = mod->mem[type].size;
> > +     return addr - base < size;
> > +}
> > +
> >   static inline bool within_module_core(unsigned long addr,
> >                                     const struct module *mod)
> >   {
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -     if ((unsigned long)mod->data_layout.base <= addr &&
> > -         addr < (unsigned long)mod->data_layout.base + mod->data_layout.size)
> > +     for_class_mod_mem_type(type, core)
> > +             if (within_module_mem_type(addr, mod, type))
> >                       return true;
> > -#endif
> > -     return (unsigned long)mod->core_layout.base <= addr &&
> > -            addr < (unsigned long)mod->core_layout.base + mod->core_layout.size;
> > +     return false;
> >   }
> >
> >   static inline bool within_module_init(unsigned long addr,
> >                                     const struct module *mod)
> >   {
> > -     return (unsigned long)mod->init_layout.base <= addr &&
> > -            addr < (unsigned long)mod->init_layout.base + mod->init_layout.size;
> > +     for_class_mod_mem_type(type, init)
> > +             if (within_module_mem_type(addr, mod, type))
> > +                     return true;
> > +     return false;
> >   }
> >
> >   static inline bool within_module(unsigned long addr, const struct module *mod)
> > diff --git a/kernel/module/internal.h b/kernel/module/internal.h
> > index 2e2bf236f558..c6c44ceca07a 100644
> > --- a/kernel/module/internal.h
> > +++ b/kernel/module/internal.h
> > @@ -17,27 +17,19 @@
> >   #define ARCH_SHF_SMALL 0
> >   #endif
> >
> > -/* If this is set, the section belongs in the init part of the module */
> > -#define INIT_OFFSET_MASK (1UL << (BITS_PER_LONG - 1))
> > -/* Maximum number of characters written by module_flags() */
> > -#define MODULE_FLAGS_BUF_SIZE (TAINT_FLAGS_COUNT + 4)
> > -
> > -#ifndef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -#define      data_layout core_layout
> > -#endif
> > -
> >   /*
> > - * Modules' sections will be aligned on page boundaries
> > - * to ensure complete separation of code and data, but
> > - * only when CONFIG_STRICT_MODULE_RWX=y
> > + * Use highest 4 bits of sh_entsize to store the mod_mem_type of this
> > + * section. This leaves 28 bits for offset on 32-bit systems, which is
> > + * about 256 MiB (WARN_ON_ONCE if we exceed that).
> >    */
> > -static inline unsigned int strict_align(unsigned int size)
> > -{
> > -     if (IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
> > -             return PAGE_ALIGN(size);
> > -     else
> > -             return size;
> > -}
> > +
> > +#define SH_ENTSIZE_TYPE_BITS 4
> > +#define SH_ENTSIZE_TYPE_SHIFT        (BITS_PER_LONG - SH_ENTSIZE_TYPE_BITS)
> > +#define SH_ENTSIZE_TYPE_MASK ((1UL << SH_ENTSIZE_TYPE_BITS) - 1)
> > +#define SH_ENTSIZE_OFFSET_MASK       ((1UL << (BITS_PER_LONG - SH_ENTSIZE_TYPE_BITS)) - 1)
> > +
> > +/* Maximum number of characters written by module_flags() */
> > +#define MODULE_FLAGS_BUF_SIZE (TAINT_FLAGS_COUNT + 4)
> >
> >   extern struct mutex module_mutex;
> >   extern struct list_head modules;
> > @@ -101,8 +93,8 @@ int try_to_force_load(struct module *mod, const char *reason);
> >   bool find_symbol(struct find_symbol_arg *fsa);
> >   struct module *find_module_all(const char *name, size_t len, bool even_unformed);
> >   int cmp_name(const void *name, const void *sym);
> > -long module_get_offset(struct module *mod, unsigned int *size, Elf_Shdr *sechdr,
> > -                    unsigned int section);
> > +long module_get_offset_and_type(struct module *mod, enum mod_mem_type type,
> > +                             Elf_Shdr *sechdr, unsigned int section);
> >   char *module_flags(struct module *mod, char *buf, bool show_state);
> >   size_t module_flags_taint(unsigned long taints, char *buf);
> >
> > @@ -190,10 +182,13 @@ struct mod_tree_root {
> >   #endif
> >       unsigned long addr_min;
> >       unsigned long addr_max;
> > +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > +     unsigned long data_addr_min;
> > +     unsigned long data_addr_max;
> > +#endif
> >   };
> >
> >   extern struct mod_tree_root mod_tree;
> > -extern struct mod_tree_root mod_data_tree;
> >
> >   #ifdef CONFIG_MODULES_TREE_LOOKUP
> >   void mod_tree_insert(struct module *mod);
> > @@ -224,7 +219,6 @@ void module_enable_nx(const struct module *mod);
> >   void module_enable_x(const struct module *mod);
> >   int module_enforce_rwx_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
> >                               char *secstrings, struct module *mod);
> > -bool module_check_misalignment(const struct module *mod);
> >
> >   #ifdef CONFIG_MODULE_SIG
> >   int module_sig_check(struct load_info *info, int flags);
> > diff --git a/kernel/module/kallsyms.c b/kernel/module/kallsyms.c
> > index 4523f99b0358..acdca7692961 100644
> > --- a/kernel/module/kallsyms.c
> > +++ b/kernel/module/kallsyms.c
> > @@ -78,6 +78,7 @@ static bool is_core_symbol(const Elf_Sym *src, const Elf_Shdr *sechdrs,
> >                          unsigned int shnum, unsigned int pcpundx)
> >   {
> >       const Elf_Shdr *sec;
> > +     enum mod_mem_type type;
> >
> >       if (src->st_shndx == SHN_UNDEF ||
> >           src->st_shndx >= shnum ||
> > @@ -90,11 +91,12 @@ static bool is_core_symbol(const Elf_Sym *src, const Elf_Shdr *sechdrs,
> >   #endif
> >
> >       sec = sechdrs + src->st_shndx;
> > +     type = sec->sh_entsize >> SH_ENTSIZE_TYPE_SHIFT;
> >       if (!(sec->sh_flags & SHF_ALLOC)
> >   #ifndef CONFIG_KALLSYMS_ALL
> >           || !(sec->sh_flags & SHF_EXECINSTR)
> >   #endif
> > -         || (sec->sh_entsize & INIT_OFFSET_MASK))
> > +         || mod_mem_type_is_init(type))
> >               return false;
> >
> >       return true;
> > @@ -113,11 +115,13 @@ void layout_symtab(struct module *mod, struct load_info *info)
> >       Elf_Shdr *strsect = info->sechdrs + info->index.str;
> >       const Elf_Sym *src;
> >       unsigned int i, nsrc, ndst, strtab_size = 0;
> > +     struct module_memory *mod_mem_data = &mod->mem[MOD_DATA];
> > +     struct module_memory *mod_mem_init_data = &mod->mem[MOD_INIT_DATA];
> >
> >       /* Put symbol section at end of init part of module. */
> >       symsect->sh_flags |= SHF_ALLOC;
> > -     symsect->sh_entsize = module_get_offset(mod, &mod->init_layout.size, symsect,
> > -                                             info->index.sym) | INIT_OFFSET_MASK;
> > +     symsect->sh_entsize = module_get_offset_and_type(mod, MOD_INIT_DATA,
> > +                                                      symsect, info->index.sym);
> >       pr_debug("\t%s\n", info->secstrings + symsect->sh_name);
> >
> >       src = (void *)info->hdr + symsect->sh_offset;
> > @@ -134,28 +138,27 @@ void layout_symtab(struct module *mod, struct load_info *info)
> >       }
> >
> >       /* Append room for core symbols at end of core part. */
> > -     info->symoffs = ALIGN(mod->data_layout.size, symsect->sh_addralign ?: 1);
> > -     info->stroffs = mod->data_layout.size = info->symoffs + ndst * sizeof(Elf_Sym);
> > -     mod->data_layout.size += strtab_size;
> > +     info->symoffs = ALIGN(mod_mem_data->size, symsect->sh_addralign ?: 1);
> > +     info->stroffs = mod_mem_data->size = info->symoffs + ndst * sizeof(Elf_Sym);
> > +     mod_mem_data->size += strtab_size;
> >       /* Note add_kallsyms() computes strtab_size as core_typeoffs - stroffs */
> > -     info->core_typeoffs = mod->data_layout.size;
> > -     mod->data_layout.size += ndst * sizeof(char);
> > -     mod->data_layout.size = strict_align(mod->data_layout.size);
> > +     info->core_typeoffs = mod_mem_data->size;
> > +     mod_mem_data->size += ndst * sizeof(char);
> >
> >       /* Put string table section at end of init part of module. */
> >       strsect->sh_flags |= SHF_ALLOC;
> > -     strsect->sh_entsize = module_get_offset(mod, &mod->init_layout.size, strsect,
> > -                                             info->index.str) | INIT_OFFSET_MASK;
> > +     strsect->sh_entsize = module_get_offset_and_type(mod, MOD_INIT_DATA,
> > +                                                      strsect, info->index.str);
> >       pr_debug("\t%s\n", info->secstrings + strsect->sh_name);
> >
> >       /* We'll tack temporary mod_kallsyms on the end. */
> > -     mod->init_layout.size = ALIGN(mod->init_layout.size,
> > +     mod_mem_init_data->size = ALIGN(mod_mem_init_data->size,
> >                                       __alignof__(struct mod_kallsyms));
> > -     info->mod_kallsyms_init_off = mod->init_layout.size;
> > -     mod->init_layout.size += sizeof(struct mod_kallsyms);
> > -     info->init_typeoffs = mod->init_layout.size;
> > -     mod->init_layout.size += nsrc * sizeof(char);
> > -     mod->init_layout.size = strict_align(mod->init_layout.size);
> > +     info->mod_kallsyms_init_off = mod_mem_init_data->size;
> > +
> > +     mod_mem_init_data->size += sizeof(struct mod_kallsyms);
> > +     info->init_typeoffs = mod_mem_init_data->size;
> > +     mod_mem_init_data->size += nsrc * sizeof(char);
> >   }
> >
> >   /*
> > @@ -171,9 +174,11 @@ void add_kallsyms(struct module *mod, const struct load_info *info)
> >       char *s;
> >       Elf_Shdr *symsec = &info->sechdrs[info->index.sym];
> >       unsigned long strtab_size;
> > +     void *data_base = mod->mem[MOD_DATA].base;
> > +     void *init_data_base = mod->mem[MOD_INIT_DATA].base;
> >
> >       /* Set up to point into init section. */
> > -     mod->kallsyms = (void __rcu *)mod->init_layout.base +
> > +     mod->kallsyms = (void __rcu *)init_data_base +
> >               info->mod_kallsyms_init_off;
> >
> >       rcu_read_lock();
> > @@ -183,15 +188,15 @@ void add_kallsyms(struct module *mod, const struct load_info *info)
> >       /* Make sure we get permanent strtab: don't use info->strtab. */
> >       rcu_dereference(mod->kallsyms)->strtab =
> >               (void *)info->sechdrs[info->index.str].sh_addr;
> > -     rcu_dereference(mod->kallsyms)->typetab = mod->init_layout.base + info->init_typeoffs;
> > +     rcu_dereference(mod->kallsyms)->typetab = init_data_base + info->init_typeoffs;
> >
> >       /*
> >        * Now populate the cut down core kallsyms for after init
> >        * and set types up while we still have access to sections.
> >        */
> > -     mod->core_kallsyms.symtab = dst = mod->data_layout.base + info->symoffs;
> > -     mod->core_kallsyms.strtab = s = mod->data_layout.base + info->stroffs;
> > -     mod->core_kallsyms.typetab = mod->data_layout.base + info->core_typeoffs;
> > +     mod->core_kallsyms.symtab = dst = data_base + info->symoffs;
> > +     mod->core_kallsyms.strtab = s = data_base + info->stroffs;
> > +     mod->core_kallsyms.typetab = data_base + info->core_typeoffs;
> >       strtab_size = info->core_typeoffs - info->stroffs;
> >       src = rcu_dereference(mod->kallsyms)->symtab;
> >       for (ndst = i = 0; i < rcu_dereference(mod->kallsyms)->num_symtab; i++) {
> > @@ -267,12 +272,15 @@ static const char *find_kallsyms_symbol(struct module *mod,
> >       unsigned int i, best = 0;
> >       unsigned long nextval, bestval;
> >       struct mod_kallsyms *kallsyms = rcu_dereference_sched(mod->kallsyms);
> > +     struct module_memory *mod_mem;
> >
> >       /* At worse, next value is at end of module */
> >       if (within_module_init(addr, mod))
> > -             nextval = (unsigned long)mod->init_layout.base + mod->init_layout.text_size;
> > +             mod_mem = &mod->mem[MOD_INIT_TEXT];
> >       else
> > -             nextval = (unsigned long)mod->core_layout.base + mod->core_layout.text_size;
> > +             mod_mem = &mod->mem[MOD_TEXT];
> > +
> > +     nextval = (unsigned long)mod_mem->base + mod_mem->size;
> >
> >       bestval = kallsyms_symbol_value(&kallsyms->symtab[best]);
> >
> > diff --git a/kernel/module/kdb.c b/kernel/module/kdb.c
> > index f4317f92e189..995c32d3698f 100644
> > --- a/kernel/module/kdb.c
> > +++ b/kernel/module/kdb.c
> > @@ -26,10 +26,11 @@ int kdb_lsmod(int argc, const char **argv)
> >               if (mod->state == MODULE_STATE_UNFORMED)
> >                       continue;
> >
> > -             kdb_printf("%-20s%8u", mod->name, mod->core_layout.size);
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -             kdb_printf("/%8u", mod->data_layout.size);
> > -#endif
> > +             kdb_printf("%-20s%8u", mod->name, mod->mem[MOD_TEXT].size);
> > +             kdb_printf("/%8u", mod->mem[MOD_RODATA].size);
> > +             kdb_printf("/%8u", mod->mem[MOD_RO_AFTER_INIT].size);
> > +             kdb_printf("/%8u", mod->mem[MOD_DATA].size);
> > +
> >               kdb_printf("  0x%px ", (void *)mod);
> >   #ifdef CONFIG_MODULE_UNLOAD
> >               kdb_printf("%4d ", module_refcount(mod));
> > @@ -40,10 +41,10 @@ int kdb_lsmod(int argc, const char **argv)
> >                       kdb_printf(" (Loading)");
> >               else
> >                       kdb_printf(" (Live)");
> > -             kdb_printf(" 0x%px", mod->core_layout.base);
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -             kdb_printf("/0x%px", mod->data_layout.base);
> > -#endif
> > +             kdb_printf(" 0x%px", mod->mem[MOD_TEXT].base);
> > +             kdb_printf("/0x%px", mod->mem[MOD_RODATA].base);
> > +             kdb_printf("/0x%px", mod->mem[MOD_RO_AFTER_INIT].base);
> > +             kdb_printf("/0x%px", mod->mem[MOD_DATA].base);
> >
> >   #ifdef CONFIG_MODULE_UNLOAD
> >               {
> > diff --git a/kernel/module/main.c b/kernel/module/main.c
> > index d3be89de706d..9f890d14072c 100644
> > --- a/kernel/module/main.c
> > +++ b/kernel/module/main.c
> > @@ -80,12 +80,6 @@ struct mod_tree_root mod_tree __cacheline_aligned = {
> >       .addr_min = -1UL,
> >   };
> >
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -struct mod_tree_root mod_data_tree __cacheline_aligned = {
> > -     .addr_min = -1UL,
> > -};
> > -#endif
> > -
> >   struct symsearch {
> >       const struct kernel_symbol *start, *stop;
> >       const s32 *crcs;
> > @@ -93,14 +87,24 @@ struct symsearch {
> >   };
> >
> >   /*
> > - * Bounds of module text, for speeding up __module_address.
> > + * Bounds of module memory, for speeding up __module_address.
> >    * Protected by module_mutex.
> >    */
> > -static void __mod_update_bounds(void *base, unsigned int size, struct mod_tree_root *tree)
> > +static void __mod_update_bounds(enum mod_mem_type type __maybe_unused, void *base,
> > +                             unsigned int size, struct mod_tree_root *tree)
> >   {
> >       unsigned long min = (unsigned long)base;
> >       unsigned long max = min + size;
> >
> > +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > +     if (mod_mem_type_is_core_data(type)) {
> > +             if (min < tree->data_addr_min)
> > +                     tree->data_addr_min = min;
> > +             if (max > tree->data_addr_max)
> > +                     tree->data_addr_max = max;
> > +             return;
> > +     }
> > +#endif
> >       if (min < tree->addr_min)
> >               tree->addr_min = min;
> >       if (max > tree->addr_max)
> > @@ -109,12 +113,12 @@ static void __mod_update_bounds(void *base, unsigned int size, struct mod_tree_r
> >
> >   static void mod_update_bounds(struct module *mod)
> >   {
> > -     __mod_update_bounds(mod->core_layout.base, mod->core_layout.size, &mod_tree);
> > -     if (mod->init_layout.size)
> > -             __mod_update_bounds(mod->init_layout.base, mod->init_layout.size, &mod_tree);
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -     __mod_update_bounds(mod->data_layout.base, mod->data_layout.size, &mod_data_tree);
> > -#endif
> > +     for_each_mod_mem_type(type) {
> > +             struct module_memory *mod_mem = &mod->mem[type];
> > +
> > +             if (mod_mem->size)
> > +                     __mod_update_bounds(type, mod_mem->base, mod_mem->size, &mod_tree);
> > +     }
> >   }
> >
> >   /* Block module loading/unloading? */
> > @@ -923,12 +927,27 @@ static ssize_t store_uevent(struct module_attribute *mattr,
> >   struct module_attribute module_uevent =
> >       __ATTR(uevent, 0200, NULL, store_uevent);
> >
> > +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > +
> >   static ssize_t show_coresize(struct module_attribute *mattr,
> >                            struct module_kobject *mk, char *buffer)
> >   {
> > -     return sprintf(buffer, "%u\n", mk->mod->core_layout.size);
> > +     return sprintf(buffer, "%u\n", mk->mod->mem[MOD_TEXT].size);
> >   }
> >
> > +#else
> > +
> > +static ssize_t show_coresize(struct module_attribute *mattr,
> > +                          struct module_kobject *mk, char *buffer)
> > +{
> > +     unsigned int size = 0;
> > +
> > +     for_class_mod_mem_type(type, core)
> > +             size += mk->mod->mem[type].size;
> > +     return sprintf(buffer, "%u\n", size);
> > +}
> > +#endif
> > +
> >   static struct module_attribute modinfo_coresize =
> >       __ATTR(coresize, 0444, show_coresize, NULL);
> >
> > @@ -936,7 +955,14 @@ static struct module_attribute modinfo_coresize =
> >   static ssize_t show_datasize(struct module_attribute *mattr,
> >                            struct module_kobject *mk, char *buffer)
> >   {
> > -     return sprintf(buffer, "%u\n", mk->mod->data_layout.size);
> > +     unsigned int size = 0;
> > +
> > +     for_class_mod_mem_type(type, core) {
> > +             if (type == MOD_TEXT)
> > +                     continue;
> > +             size += mk->mod->mem[type].size;
> > +     }
> > +     return sprintf(buffer, "%u\n", size);
> >   }
> >
> >   static struct module_attribute modinfo_datasize =
> > @@ -946,7 +972,11 @@ static struct module_attribute modinfo_datasize =
> >   static ssize_t show_initsize(struct module_attribute *mattr,
> >                            struct module_kobject *mk, char *buffer)
> >   {
> > -     return sprintf(buffer, "%u\n", mk->mod->init_layout.size);
> > +     unsigned int size = 0;
> > +
> > +     for_class_mod_mem_type(type, init)
> > +             size += mk->mod->mem[type].size;
> > +     return sprintf(buffer, "%u\n", size);
> >   }
> >
> >   static struct module_attribute modinfo_initsize =
> > @@ -1143,6 +1173,63 @@ void __weak module_arch_freeing_init(struct module *mod)
> >   {
> >   }
> >
> > +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > +static bool mod_mem_use_vmalloc(enum mod_mem_type type)
> > +{
> > +     return mod_mem_is_core_data(type);
> > +}
> > +#else
> > +static bool mod_mem_use_vmalloc(enum mod_mem_type type)
> > +{
> > +     return false;
> > +}
> > +#endif
> > +
> > +static void *module_memory_alloc(unsigned int size, enum mod_mem_type type)
> > +{
> > +     if (mod_mem_use_vmalloc(type))
> > +             return vzalloc(size);
> > +     return module_alloc(size);
> > +}
> > +
> > +static void module_memory_free(void *ptr, enum mod_mem_type type)
> > +{
> > +     if (mod_mem_use_vmalloc(type))
> > +             vfree(ptr);
> > +     else
> > +             module_memfree(ptr);
> > +}
> > +
> > +static void free_mod_mem(struct module *mod)
> > +{
> > +     /* free the memory in the right order to avoid use-after-free */
> > +     static enum mod_mem_type mod_mem_free_order[MOD_MEM_NUM_TYPES] = {
> > +             /* first free init sections */
> > +             MOD_INIT_TEXT,
> > +             MOD_INIT_DATA,
> > +             MOD_INIT_RODATA,
> > +
> > +             /* then core sections, except rw data */
> > +             MOD_TEXT,
> > +             MOD_RODATA,
> > +             MOD_RO_AFTER_INIT,
> > +
> > +             /* last, rw data */
> > +             MOD_DATA,
> > +     };
> > +     int i;
> > +
> > +     for (i = 0; i < MOD_MEM_NUM_TYPES; i++) {
> > +             enum mod_mem_type type = mod_mem_free_order[i];
> > +             struct module_memory *mod_mem = &mod->mem[type];
> > +
> > +             /* Free lock-classes; relies on the preceding sync_rcu(). */
> > +             lockdep_free_key_range(mod_mem->base, mod_mem->size);
> > +             if (mod_mem->size)
> > +                     module_memory_free(mod_mem->base, type);
> > +     }
> > +}
> > +
> >   /* Free a module, remove from lists, etc. */
> >   static void free_module(struct module *mod)
> >   {
> > @@ -1189,18 +1276,10 @@ static void free_module(struct module *mod)
> >
> >       /* This may be empty, but that's OK */
> >       module_arch_freeing_init(mod);
> > -     module_memfree(mod->init_layout.base);
> >       kfree(mod->args);
> >       percpu_modfree(mod);
> >
> > -     /* Free lock-classes; relies on the preceding sync_rcu(). */
> > -     lockdep_free_key_range(mod->data_layout.base, mod->data_layout.size);
> > -
> > -     /* Finally, free the core (containing the module structure) */
> > -     module_memfree(mod->core_layout.base);
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -     vfree(mod->data_layout.base);
> > -#endif
> > +     free_mod_mem(mod);
> >   }
> >
> >   void *__symbol_get(const char *symbol)
> > @@ -1387,16 +1466,18 @@ unsigned int __weak arch_mod_section_prepend(struct module *mod,
> >       return 0;
> >   }
> >
> > -/* Update size with this section: return offset. */
> > -long module_get_offset(struct module *mod, unsigned int *size,
> > +long module_get_offset_and_type(struct module *mod, enum mod_mem_type type,
> >                               Elf_Shdr *sechdr, unsigned int section)
> >   {
> > -     long ret;
> > +     long offset;
> > +     long mask = ((unsigned long)(type) & SH_ENTSIZE_TYPE_MASK) << SH_ENTSIZE_TYPE_SHIFT;
> >
> > -     *size += arch_mod_section_prepend(mod, section);
> > -     ret = ALIGN(*size, sechdr->sh_addralign ?: 1);
> > -     *size = ret + sechdr->sh_size;
> > -     return ret;
> > +     mod->mem[type].size += arch_mod_section_prepend(mod, section);
> > +     offset = ALIGN(mod->mem[type].size, sechdr->sh_addralign ?: 1);
> > +     mod->mem[type].size = offset + sechdr->sh_size;
> > +
> > +     WARN_ON_ONCE(offset & mask);
> > +     return offset | mask;
> >   }
> >
> >   static bool module_init_layout_section(const char *sname)
> > @@ -1428,6 +1509,20 @@ static void layout_sections(struct module *mod, struct load_info *info)
> >               { SHF_WRITE | SHF_ALLOC, ARCH_SHF_SMALL },
> >               { ARCH_SHF_SMALL | SHF_ALLOC, 0 }
> >       };
> > +     static int core_m_to_mem_type[] = {
> > +             MOD_TEXT,
> > +             MOD_RODATA,
> > +             MOD_RO_AFTER_INIT,
> > +             MOD_DATA,
> > +             MOD_INVALID,
> > +     };
> > +     static int init_m_to_mem_type[] = {
> > +             MOD_INIT_TEXT,
> > +             MOD_INIT_RODATA,
> > +             MOD_INVALID,
> > +             MOD_INIT_DATA,
> > +             MOD_INVALID,
> > +     };
> >       unsigned int m, i;
> >
> >       for (i = 0; i < info->hdr->e_shnum; i++)
> > @@ -1435,41 +1530,30 @@ static void layout_sections(struct module *mod, struct load_info *info)
> >
> >       pr_debug("Core section allocation order:\n");
> >       for (m = 0; m < ARRAY_SIZE(masks); ++m) {
> > +             enum mod_mem_type type = core_m_to_mem_type[m];
> > +
> >               for (i = 0; i < info->hdr->e_shnum; ++i) {
> >                       Elf_Shdr *s = &info->sechdrs[i];
> >                       const char *sname = info->secstrings + s->sh_name;
> > -                     unsigned int *sizep;
> >
> >                       if ((s->sh_flags & masks[m][0]) != masks[m][0]
> >                           || (s->sh_flags & masks[m][1])
> >                           || s->sh_entsize != ~0UL
> >                           || module_init_layout_section(sname))
> >                               continue;
> > -                     sizep = m ? &mod->data_layout.size : &mod->core_layout.size;
> > -                     s->sh_entsize = module_get_offset(mod, sizep, s, i);
> > +
> > +                     if (WARN_ON_ONCE(type == MOD_INVALID))
> > +                             continue;
> > +
> > +                     s->sh_entsize = module_get_offset_and_type(mod, type, s, i);
> >                       pr_debug("\t%s\n", sname);
> >               }
> > -             switch (m) {
> > -             case 0: /* executable */
> > -                     mod->core_layout.size = strict_align(mod->core_layout.size);
> > -                     mod->core_layout.text_size = mod->core_layout.size;
> > -                     break;
> > -             case 1: /* RO: text and ro-data */
> > -                     mod->data_layout.size = strict_align(mod->data_layout.size);
> > -                     mod->data_layout.ro_size = mod->data_layout.size;
> > -                     break;
> > -             case 2: /* RO after init */
> > -                     mod->data_layout.size = strict_align(mod->data_layout.size);
> > -                     mod->data_layout.ro_after_init_size = mod->data_layout.size;
> > -                     break;
> > -             case 4: /* whole core */
> > -                     mod->data_layout.size = strict_align(mod->data_layout.size);
> > -                     break;
> > -             }
> >       }
> >
> >       pr_debug("Init section allocation order:\n");
> >       for (m = 0; m < ARRAY_SIZE(masks); ++m) {
> > +             enum mod_mem_type type = init_m_to_mem_type[m];
> > +
> >               for (i = 0; i < info->hdr->e_shnum; ++i) {
> >                       Elf_Shdr *s = &info->sechdrs[i];
> >                       const char *sname = info->secstrings + s->sh_name;
> > @@ -1479,30 +1563,13 @@ static void layout_sections(struct module *mod, struct load_info *info)
> >                           || s->sh_entsize != ~0UL
> >                           || !module_init_layout_section(sname))
> >                               continue;
> > -                     s->sh_entsize = (module_get_offset(mod, &mod->init_layout.size, s, i)
> > -                                      | INIT_OFFSET_MASK);
> > +
> > +                     if (WARN_ON_ONCE(type == MOD_INVALID))
> > +                             continue;
> > +
> > +                     s->sh_entsize = module_get_offset_and_type(mod, type, s, i);
> >                       pr_debug("\t%s\n", sname);
> >               }
> > -             switch (m) {
> > -             case 0: /* executable */
> > -                     mod->init_layout.size = strict_align(mod->init_layout.size);
> > -                     mod->init_layout.text_size = mod->init_layout.size;
> > -                     break;
> > -             case 1: /* RO: text and ro-data */
> > -                     mod->init_layout.size = strict_align(mod->init_layout.size);
> > -                     mod->init_layout.ro_size = mod->init_layout.size;
> > -                     break;
> > -             case 2:
> > -                     /*
> > -                      * RO after init doesn't apply to init_layout (only
> > -                      * core_layout), so it just takes the value of ro_size.
> > -                      */
> > -                     mod->init_layout.ro_after_init_size = mod->init_layout.ro_size;
> > -                     break;
> > -             case 4: /* whole init */
> > -                     mod->init_layout.size = strict_align(mod->init_layout.size);
> > -                     break;
> > -             }
> >       }
> >   }
> >
> > @@ -2122,72 +2189,42 @@ static int move_module(struct module *mod, struct load_info *info)
> >   {
> >       int i;
> >       void *ptr;
> > +     enum mod_mem_type t;
> >
> > -     /* Do the allocs. */
> > -     ptr = module_alloc(mod->core_layout.size);
> > -     /*
> > -      * The pointer to this block is stored in the module structure
> > -      * which is inside the block. Just mark it as not being a
> > -      * leak.
> > -      */
> > -     kmemleak_not_leak(ptr);
> > -     if (!ptr)
> > -             return -ENOMEM;
> > -
> > -     memset(ptr, 0, mod->core_layout.size);
> > -     mod->core_layout.base = ptr;
> > -
> > -     if (mod->init_layout.size) {
> > -             ptr = module_alloc(mod->init_layout.size);
> > -             /*
> > -              * The pointer to this block is stored in the module structure
> > -              * which is inside the block. This block doesn't need to be
> > -              * scanned as it contains data and code that will be freed
> > -              * after the module is initialized.
> > -              */
> > -             kmemleak_ignore(ptr);
> > -             if (!ptr) {
> > -                     module_memfree(mod->core_layout.base);
> > -                     return -ENOMEM;
> > +     for_each_mod_mem_type(type) {
> > +             if (!mod->mem[type].size) {
> > +                     mod->mem[type].base = NULL;
> > +                     continue;
> >               }
> > -             memset(ptr, 0, mod->init_layout.size);
> > -             mod->init_layout.base = ptr;
> > -     } else
> > -             mod->init_layout.base = NULL;
> > +             mod->mem[type].size = PAGE_ALIGN(mod->mem[type].size);
> > +             ptr = module_memory_alloc(mod->mem[type].size, type);
> >
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -     /* Do the allocs. */
> > -     ptr = vzalloc(mod->data_layout.size);
> >               /*
> >                * The pointer to this block is stored in the module structure
> >                * which is inside the block. Just mark it as not being a
> >                * leak.
> >                */
> > -     kmemleak_not_leak(ptr);
> > +             kmemleak_ignore(ptr);
> >               if (!ptr) {
> > -             module_memfree(mod->core_layout.base);
> > -             module_memfree(mod->init_layout.base);
> > -             return -ENOMEM;
> > +                     t = type;
> > +                     goto out_enomem;
> > +             }
> > +             memset(ptr, 0, mod->mem[type].size);
> > +             mod->mem[type].base = ptr;
> >       }
> >
> > -     mod->data_layout.base = ptr;
> > -#endif
> >       /* Transfer each section which specifies SHF_ALLOC */
> >       pr_debug("final section addresses:\n");
> >       for (i = 0; i < info->hdr->e_shnum; i++) {
> >               void *dest;
> >               Elf_Shdr *shdr = &info->sechdrs[i];
> > +             enum mod_mem_type type = shdr->sh_entsize >> SH_ENTSIZE_TYPE_SHIFT;
> >
> >               if (!(shdr->sh_flags & SHF_ALLOC))
> >                       continue;
> >
> > -             if (shdr->sh_entsize & INIT_OFFSET_MASK)
> > -                     dest = mod->init_layout.base
> > -                             + (shdr->sh_entsize & ~INIT_OFFSET_MASK);
> > -             else if (!(shdr->sh_flags & SHF_EXECINSTR))
> > -                     dest = mod->data_layout.base + shdr->sh_entsize;
> > -             else
> > -                     dest = mod->core_layout.base + shdr->sh_entsize;
> > +             dest = mod->mem[type].base +
> > +                     (shdr->sh_entsize & SH_ENTSIZE_OFFSET_MASK);
> >
> >               if (shdr->sh_type != SHT_NOBITS)
> >                       memcpy(dest, (void *)shdr->sh_addr, shdr->sh_size);
> > @@ -2198,6 +2235,10 @@ static int move_module(struct module *mod, struct load_info *info)
> >       }
> >
> >       return 0;
> > +out_enomem:
> > +     for (t--; t >= 0; t--)
> > +             module_memory_free(mod->mem[t].base, t);
> > +     return -ENOMEM;
> >   }
> >
> >   static int check_module_license_and_versions(struct module *mod)
> > @@ -2242,12 +2283,13 @@ static void flush_module_icache(const struct module *mod)
> >        * Do it before processing of module parameters, so the module
> >        * can provide parameter accessor functions of its own.
> >        */
> > -     if (mod->init_layout.base)
> > -             flush_icache_range((unsigned long)mod->init_layout.base,
> > -                                (unsigned long)mod->init_layout.base
> > -                                + mod->init_layout.size);
> > -     flush_icache_range((unsigned long)mod->core_layout.base,
> > -                        (unsigned long)mod->core_layout.base + mod->core_layout.size);
> > +     for_each_mod_mem_type(type) {
> > +             const struct module_memory *mod_mem = &mod->mem[type];
> > +
> > +             if (mod_mem->size)
> > +                     flush_icache_range((unsigned long)mod_mem->base,
> > +                                        (unsigned long)mod_mem->base + mod_mem->size);
> > +     }
> >   }
> >
> >   bool __weak module_elf_check_arch(Elf_Ehdr *hdr)
> > @@ -2350,11 +2392,8 @@ static void module_deallocate(struct module *mod, struct load_info *info)
> >   {
> >       percpu_modfree(mod);
> >       module_arch_freeing_init(mod);
> > -     module_memfree(mod->init_layout.base);
> > -     module_memfree(mod->core_layout.base);
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -     vfree(mod->data_layout.base);
> > -#endif
> > +
> > +     free_mod_mem(mod);
> >   }
> >
> >   int __weak module_finalize(const Elf_Ehdr *hdr,
> > @@ -2415,7 +2454,9 @@ static void do_mod_ctors(struct module *mod)
> >   /* For freeing module_init on success, in case kallsyms traversing */
> >   struct mod_initfree {
> >       struct llist_node node;
> > -     void *module_init;
> > +     void *init_text;
> > +     void *init_data;
> > +     void *init_rodata;
> >   };
> >
> >   static void do_free_init(struct work_struct *w)
> > @@ -2429,7 +2470,9 @@ static void do_free_init(struct work_struct *w)
> >
> >       llist_for_each_safe(pos, n, list) {
> >               initfree = container_of(pos, struct mod_initfree, node);
> > -             module_memfree(initfree->module_init);
> > +             module_memfree(initfree->init_text);
> > +             module_memfree(initfree->init_data);
> > +             module_memfree(initfree->init_rodata);
> >               kfree(initfree);
> >       }
> >   }
> > @@ -2456,7 +2499,9 @@ static noinline int do_init_module(struct module *mod)
> >               ret = -ENOMEM;
> >               goto fail;
> >       }
> > -     freeinit->module_init = mod->init_layout.base;
> > +     freeinit->init_text = mod->mem[MOD_INIT_TEXT].base;
> > +     freeinit->init_data = mod->mem[MOD_INIT_DATA].base;
> > +     freeinit->init_rodata = mod->mem[MOD_INIT_RODATA].base;
> >
> >       do_mod_ctors(mod);
> >       /* Start the module */
> > @@ -2492,8 +2537,8 @@ static noinline int do_init_module(struct module *mod)
> >       if (!mod->async_probe_requested)
> >               async_synchronize_full();
> >
> > -     ftrace_free_mem(mod, mod->init_layout.base, mod->init_layout.base +
> > -                     mod->init_layout.size);
> > +     ftrace_free_mem(mod, mod->mem[MOD_INIT_TEXT].base,
> > +                     mod->mem[MOD_INIT_TEXT].base + mod->mem[MOD_INIT_TEXT].size);
> >       mutex_lock(&module_mutex);
> >       /* Drop initial reference. */
> >       module_put(mod);
> > @@ -2505,11 +2550,10 @@ static noinline int do_init_module(struct module *mod)
> >       module_enable_ro(mod, true);
> >       mod_tree_remove_init(mod);
> >       module_arch_freeing_init(mod);
> > -     mod->init_layout.base = NULL;
> > -     mod->init_layout.size = 0;
> > -     mod->init_layout.ro_size = 0;
> > -     mod->init_layout.ro_after_init_size = 0;
> > -     mod->init_layout.text_size = 0;
> > +     for_class_mod_mem_type(type, init) {
> > +             mod->mem[type].base = NULL;
> > +             mod->mem[type].size = 0;
> > +     }
> >   #ifdef CONFIG_DEBUG_INFO_BTF_MODULES
> >       /* .BTF is not SHF_ALLOC and will get removed, so sanitize pointer */
> >       mod->btf_data = NULL;
> > @@ -2628,9 +2672,6 @@ static int complete_formation(struct module *mod, struct load_info *info)
> >       module_bug_finalize(info->hdr, info->sechdrs, mod);
> >       module_cfi_finalize(info->hdr, info->sechdrs, mod);
> >
> > -     if (module_check_misalignment(mod))
> > -             goto out_misaligned;
> > -
> >       module_enable_ro(mod, false);
> >       module_enable_nx(mod);
> >       module_enable_x(mod);
> > @@ -2644,8 +2685,6 @@ static int complete_formation(struct module *mod, struct load_info *info)
> >
> >       return 0;
> >
> > -out_misaligned:
> > -     err = -EINVAL;
> >   out:
> >       mutex_unlock(&module_mutex);
> >       return err;
> > @@ -2909,7 +2948,10 @@ static int load_module(struct load_info *info, const char __user *uargs,
> >       mutex_unlock(&module_mutex);
> >    free_module:
> >       /* Free lock-classes; relies on the preceding sync_rcu() */
> > -     lockdep_free_key_range(mod->data_layout.base, mod->data_layout.size);
> > +     for_class_mod_mem_type(type, core) {
> > +             lockdep_free_key_range(mod->mem[type].base,
> > +                                    mod->mem[type].size);
> > +     }
> >
> >       module_deallocate(mod, info);
> >    free_copy:
> > @@ -3060,20 +3102,21 @@ bool is_module_address(unsigned long addr)
> >   struct module *__module_address(unsigned long addr)
> >   {
> >       struct module *mod;
> > -     struct mod_tree_root *tree;
> >
> >       if (addr >= mod_tree.addr_min && addr <= mod_tree.addr_max)
> > -             tree = &mod_tree;
> > +             goto lookup;
> > +
> >   #ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -     else if (addr >= mod_data_tree.addr_min && addr <= mod_data_tree.addr_max)
> > -             tree = &mod_data_tree;
> > +     if (addr >= mod_tree.data_addr_min && addr <= mod_tree.data_addr_max)
> > +             goto lookup;
> >   #endif
> > -     else
> > +
> >       return NULL;
> >
> > +lookup:
> >       module_assert_mutex_or_preempt();
> >
> > -     mod = mod_find(addr, tree);
> > +     mod = mod_find(addr, &mod_tree);
> >       if (mod) {
> >               BUG_ON(!within_module(addr, mod));
> >               if (mod->state == MODULE_STATE_UNFORMED)
> > @@ -3113,8 +3156,8 @@ struct module *__module_text_address(unsigned long addr)
> >       struct module *mod = __module_address(addr);
> >       if (mod) {
> >               /* Make sure it's within the text section. */
> > -             if (!within(addr, mod->init_layout.base, mod->init_layout.text_size)
> > -                 && !within(addr, mod->core_layout.base, mod->core_layout.text_size))
> > +             if (!within_module_mem_type(addr, mod, MOD_TEXT) &&
> > +                 !within_module_mem_type(addr, mod, MOD_INIT_TEXT))
> >                       mod = NULL;
> >       }
> >       return mod;
> > diff --git a/kernel/module/procfs.c b/kernel/module/procfs.c
> > index cf5b9f1e6ec4..0a4841e88adb 100644
> > --- a/kernel/module/procfs.c
> > +++ b/kernel/module/procfs.c
> > @@ -62,6 +62,15 @@ static void m_stop(struct seq_file *m, void *p)
> >       mutex_unlock(&module_mutex);
> >   }
> >
> > +static unsigned int module_total_size(struct module *mod)
> > +{
> > +     int size = 0;
> > +
> > +     for_each_mod_mem_type(type)
> > +             size += mod->mem[type].size;
> > +     return size;
> > +}
> > +
> >   static int m_show(struct seq_file *m, void *p)
> >   {
> >       struct module *mod = list_entry(p, struct module, list);
> > @@ -73,10 +82,7 @@ static int m_show(struct seq_file *m, void *p)
> >       if (mod->state == MODULE_STATE_UNFORMED)
> >               return 0;
> >
> > -     size = mod->init_layout.size + mod->core_layout.size;
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -     size += mod->data_layout.size;
> > -#endif
> > +     size = module_total_size(mod);
> >       seq_printf(m, "%s %u", mod->name, size);
> >       print_unload_info(m, mod);
> >
> > @@ -86,7 +92,7 @@ static int m_show(struct seq_file *m, void *p)
> >                  mod->state == MODULE_STATE_COMING ? "Loading" :
> >                  "Live");
> >       /* Used by oprofile and other similar tools. */
> > -     value = m->private ? NULL : mod->core_layout.base;
> > +     value = m->private ? NULL : mod->mem[MOD_TEXT].base;
> >       seq_printf(m, " 0x%px", value);
> >
> >       /* Taints info */
> > diff --git a/kernel/module/strict_rwx.c b/kernel/module/strict_rwx.c
> > index 14fbea66f12f..2e79a77f40eb 100644
> > --- a/kernel/module/strict_rwx.c
> > +++ b/kernel/module/strict_rwx.c
> > @@ -11,82 +11,26 @@
> >   #include <linux/set_memory.h>
> >   #include "internal.h"
> >
> > -/*
> > - * LKM RO/NX protection: protect module's text/ro-data
> > - * from modification and any data from execution.
> > - *
> > - * General layout of module is:
> > - *          [text] [read-only-data] [ro-after-init] [writable data]
> > - * text_size -----^                ^               ^               ^
> > - * ro_size ------------------------|               |               |
> > - * ro_after_init_size -----------------------------|               |
> > - * size -----------------------------------------------------------|
> > - *
> > - * These values are always page-aligned (as is base) when
> > - * CONFIG_STRICT_MODULE_RWX is set.
> > - */
> > +static void module_set_memory(
> > +     const struct module *mod, enum mod_mem_type type,
> > +     int (*set_memory)(unsigned long start, int num_pages))
> > +{
> > +     const struct module_memory *mod_mem = &mod->mem[type];
> > +
> > +     set_vm_flush_reset_perms(mod_mem->base);
> > +     set_memory((unsigned long)mod_mem->base, mod_mem->size >> PAGE_SHIFT);
> > +}
> >
> >   /*
> >    * Since some arches are moving towards PAGE_KERNEL module allocations instead
> > - * of PAGE_KERNEL_EXEC, keep frob_text() and module_enable_x() independent of
> > + * of PAGE_KERNEL_EXEC, keep module_enable_x() independent of
> >    * CONFIG_STRICT_MODULE_RWX because they are needed regardless of whether we
> >    * are strict.
> >    */
> > -static void frob_text(const struct module_layout *layout,
> > -                   int (*set_memory)(unsigned long start, int num_pages))
> > -{
> > -     set_memory((unsigned long)layout->base,
> > -                PAGE_ALIGN(layout->text_size) >> PAGE_SHIFT);
> > -}
> > -
> > -static void frob_rodata(const struct module_layout *layout,
> > -              int (*set_memory)(unsigned long start, int num_pages))
> > -{
> > -     set_memory((unsigned long)layout->base + layout->text_size,
> > -                (layout->ro_size - layout->text_size) >> PAGE_SHIFT);
> > -}
> > -
> > -static void frob_ro_after_init(const struct module_layout *layout,
> > -                     int (*set_memory)(unsigned long start, int num_pages))
> > -{
> > -     set_memory((unsigned long)layout->base + layout->ro_size,
> > -                (layout->ro_after_init_size - layout->ro_size) >> PAGE_SHIFT);
> > -}
> > -
> > -static void frob_writable_data(const struct module_layout *layout,
> > -                     int (*set_memory)(unsigned long start, int num_pages))
> > -{
> > -     set_memory((unsigned long)layout->base + layout->ro_after_init_size,
> > -                (layout->size - layout->ro_after_init_size) >> PAGE_SHIFT);
> > -}
> > -
> > -static bool layout_check_misalignment(const struct module_layout *layout)
> > -{
> > -     return WARN_ON(!PAGE_ALIGNED(layout->base)) ||
> > -            WARN_ON(!PAGE_ALIGNED(layout->text_size)) ||
> > -            WARN_ON(!PAGE_ALIGNED(layout->ro_size)) ||
> > -            WARN_ON(!PAGE_ALIGNED(layout->ro_after_init_size)) ||
> > -            WARN_ON(!PAGE_ALIGNED(layout->size));
> > -}
> > -
> > -bool module_check_misalignment(const struct module *mod)
> > -{
> > -     if (!IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
> > -             return false;
> > -
> > -     return layout_check_misalignment(&mod->core_layout) ||
> > -            layout_check_misalignment(&mod->data_layout) ||
> > -            layout_check_misalignment(&mod->init_layout);
> > -}
> > -
> >   void module_enable_x(const struct module *mod)
> >   {
> > -     if (!PAGE_ALIGNED(mod->core_layout.base) ||
> > -         !PAGE_ALIGNED(mod->init_layout.base))
> > -             return;
> > -
> > -     frob_text(&mod->core_layout, set_memory_x);
> > -     frob_text(&mod->init_layout, set_memory_x);
> > +     for_class_mod_mem_type(type, text)
> > +             module_set_memory(mod, type, set_memory_x);
> >   }
> >
> >   void module_enable_ro(const struct module *mod, bool after_init)
> > @@ -98,16 +42,13 @@ void module_enable_ro(const struct module *mod, bool after_init)
> >               return;
> >   #endif
> >
> > -     set_vm_flush_reset_perms(mod->core_layout.base);
> > -     set_vm_flush_reset_perms(mod->init_layout.base);
> > -     frob_text(&mod->core_layout, set_memory_ro);
> > -
> > -     frob_rodata(&mod->data_layout, set_memory_ro);
> > -     frob_text(&mod->init_layout, set_memory_ro);
> > -     frob_rodata(&mod->init_layout, set_memory_ro);
> > +     module_set_memory(mod, MOD_TEXT, set_memory_ro);
> > +     module_set_memory(mod, MOD_INIT_TEXT, set_memory_ro);
> > +     module_set_memory(mod, MOD_RODATA, set_memory_ro);
> > +     module_set_memory(mod, MOD_INIT_RODATA, set_memory_ro);
> >
> >       if (after_init)
> > -             frob_ro_after_init(&mod->data_layout, set_memory_ro);
> > +             module_set_memory(mod, MOD_RO_AFTER_INIT, set_memory_ro);
> >   }
> >
> >   void module_enable_nx(const struct module *mod)
> > @@ -115,11 +56,8 @@ void module_enable_nx(const struct module *mod)
> >       if (!IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
> >               return;
> >
> > -     frob_rodata(&mod->data_layout, set_memory_nx);
> > -     frob_ro_after_init(&mod->data_layout, set_memory_nx);
> > -     frob_writable_data(&mod->data_layout, set_memory_nx);
> > -     frob_rodata(&mod->init_layout, set_memory_nx);
> > -     frob_writable_data(&mod->init_layout, set_memory_nx);
> > +     for_class_mod_mem_type(type, data)
> > +             module_set_memory(mod, type, set_memory_nx);
> >   }
> >
> >   int module_enforce_rwx_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
> > diff --git a/kernel/module/tree_lookup.c b/kernel/module/tree_lookup.c
> > index 8ec5cfd60496..277197977d43 100644
> > --- a/kernel/module/tree_lookup.c
> > +++ b/kernel/module/tree_lookup.c
> > @@ -21,16 +21,16 @@
> >
> >   static __always_inline unsigned long __mod_tree_val(struct latch_tree_node *n)
> >   {
> > -     struct module_layout *layout = container_of(n, struct module_layout, mtn.node);
> > +     struct module_memory *mod_mem = container_of(n, struct module_memory, mtn.node);
> >
> > -     return (unsigned long)layout->base;
> > +     return (unsigned long)mod_mem->base;
> >   }
> >
> >   static __always_inline unsigned long __mod_tree_size(struct latch_tree_node *n)
> >   {
> > -     struct module_layout *layout = container_of(n, struct module_layout, mtn.node);
> > +     struct module_memory *mod_mem = container_of(n, struct module_memory, mtn.node);
> >
> > -     return (unsigned long)layout->size;
> > +     return (unsigned long)mod_mem->size;
> >   }
> >
> >   static __always_inline bool
> > @@ -77,32 +77,27 @@ static void __mod_tree_remove(struct mod_tree_node *node, struct mod_tree_root *
> >    */
> >   void mod_tree_insert(struct module *mod)
> >   {
> > -     mod->core_layout.mtn.mod = mod;
> > -     mod->init_layout.mtn.mod = mod;
> > -
> > -     __mod_tree_insert(&mod->core_layout.mtn, &mod_tree);
> > -     if (mod->init_layout.size)
> > -             __mod_tree_insert(&mod->init_layout.mtn, &mod_tree);
> > -
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -     mod->data_layout.mtn.mod = mod;
> > -     __mod_tree_insert(&mod->data_layout.mtn, &mod_data_tree);
> > -#endif
> > +     for_each_mod_mem_type(type) {
> > +             mod->mem[type].mtn.mod = mod;
> > +             if (mod->mem[type].size)
> > +                     __mod_tree_insert(&mod->mem[type].mtn, &mod_tree);
> > +     }
> >   }
> >
> >   void mod_tree_remove_init(struct module *mod)
> >   {
> > -     if (mod->init_layout.size)
> > -             __mod_tree_remove(&mod->init_layout.mtn, &mod_tree);
> > +     for_class_mod_mem_type(type, init) {
> > +             if (mod->mem[type].size)
> > +                     __mod_tree_remove(&mod->mem[type].mtn, &mod_tree);
> > +     }
> >   }
> >
> >   void mod_tree_remove(struct module *mod)
> >   {
> > -     __mod_tree_remove(&mod->core_layout.mtn, &mod_tree);
> > -     mod_tree_remove_init(mod);
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -     __mod_tree_remove(&mod->data_layout.mtn, &mod_data_tree);
> > -#endif
> > +     for_each_mod_mem_type(type) {
> > +             if (mod->mem[type].size)
> > +                     __mod_tree_remove(&mod->mem[type].mtn, &mod_tree);
> > +     }
> >   }
> >
> >   struct module *mod_find(unsigned long addr, struct mod_tree_root *tree)
diff mbox series

Patch

diff --git a/arch/alpha/kernel/module.c b/arch/alpha/kernel/module.c
index 5b60c248de9e..9109213abc09 100644
--- a/arch/alpha/kernel/module.c
+++ b/arch/alpha/kernel/module.c
@@ -148,7 +148,7 @@  apply_relocate_add(Elf64_Shdr *sechdrs, const char *strtab,
 
 	/* The small sections were sorted to the end of the segment.
 	   The following should definitely cover them.  */
-	gp = (u64)me->core_layout.base + me->core_layout.size - 0x8000;
+	gp = (u64)me->mem[MOD_DATA].base + me->mem[MOD_DATA].size - 0x8000;
 	got = sechdrs[me->arch.gotsecindex].sh_addr;
 
 	for (i = 0; i < n; i++) {
diff --git a/arch/arc/kernel/unwind.c b/arch/arc/kernel/unwind.c
index 200270a94558..933451f4494f 100644
--- a/arch/arc/kernel/unwind.c
+++ b/arch/arc/kernel/unwind.c
@@ -369,6 +369,8 @@  void *unwind_add_table(struct module *module, const void *table_start,
 		       unsigned long table_size)
 {
 	struct unwind_table *table;
+	struct module_memory *mod_mem_core_text;
+	struct module_memory *mod_mem_init_text;
 
 	if (table_size <= 0)
 		return NULL;
@@ -377,9 +379,12 @@  void *unwind_add_table(struct module *module, const void *table_start,
 	if (!table)
 		return NULL;
 
+	mod_mem_core_text = &module->mem[MOD_TEXT];
+	mod_mem_init_text = &module->mem[MOD_INIT_TEXT];
+
 	init_unwind_table(table, module->name,
-			  module->core_layout.base, module->core_layout.size,
-			  module->init_layout.base, module->init_layout.size,
+			  mod_mem_core_text->base, mod_mem_core_text->size,
+			  mod_mem_init_text->base, mod_mem_init_text->size,
 			  table_start, table_size,
 			  NULL, 0);
 
diff --git a/arch/arm/kernel/module-plts.c b/arch/arm/kernel/module-plts.c
index af7c322ebed6..9d4ecb6b1412 100644
--- a/arch/arm/kernel/module-plts.c
+++ b/arch/arm/kernel/module-plts.c
@@ -30,7 +30,7 @@  static const u32 fixed_plts[] = {
 
 static bool in_init(const struct module *mod, unsigned long loc)
 {
-	return loc - (u32)mod->init_layout.base < mod->init_layout.size;
+	return within_module_init(loc, mod);
 }
 
 static void prealloc_fixed(struct mod_plt_sec *pltsec, struct plt_entries *plt)
diff --git a/arch/arm64/kernel/module-plts.c b/arch/arm64/kernel/module-plts.c
index 5a0a8f552a61..fedffa83a283 100644
--- a/arch/arm64/kernel/module-plts.c
+++ b/arch/arm64/kernel/module-plts.c
@@ -67,7 +67,7 @@  static bool plt_entries_equal(const struct plt_entry *a,
 
 static bool in_init(const struct module *mod, void *loc)
 {
-	return (u64)loc - (u64)mod->init_layout.base < mod->init_layout.size;
+	return within_module_init(loc, mod);
 }
 
 u64 module_emit_plt_entry(struct module *mod, Elf64_Shdr *sechdrs,
diff --git a/arch/ia64/kernel/module.c b/arch/ia64/kernel/module.c
index 8f62cf97f691..3661135da9d9 100644
--- a/arch/ia64/kernel/module.c
+++ b/arch/ia64/kernel/module.c
@@ -485,19 +485,19 @@  module_frob_arch_sections (Elf_Ehdr *ehdr, Elf_Shdr *sechdrs, char *secstrings,
 	return 0;
 }
 
-static inline int
+static inline bool
 in_init (const struct module *mod, uint64_t addr)
 {
-	return addr - (uint64_t) mod->init_layout.base < mod->init_layout.size;
+	return within_module_init(addr, mod);
 }
 
-static inline int
+static inline bool
 in_core (const struct module *mod, uint64_t addr)
 {
-	return addr - (uint64_t) mod->core_layout.base < mod->core_layout.size;
+	return within_module_core(addr, mod);
 }
 
-static inline int
+static inline bool
 is_internal (const struct module *mod, uint64_t value)
 {
 	return in_init(mod, value) || in_core(mod, value);
@@ -677,7 +677,8 @@  do_reloc (struct module *mod, uint8_t r_type, Elf64_Sym *sym, uint64_t addend,
 		break;
 
 	      case RV_BDREL:
-		val -= (uint64_t) (in_init(mod, val) ? mod->init_layout.base : mod->core_layout.base);
+		val -= (uint64_t) (in_init(mod, val) ? mod->mem[MOD_INIT_TEXT].base :
+				   mod->mem[MOD_TEXT].base);
 		break;
 
 	      case RV_LTV:
@@ -812,15 +813,18 @@  apply_relocate_add (Elf64_Shdr *sechdrs, const char *strtab, unsigned int symind
 		 *     addresses have been selected...
 		 */
 		uint64_t gp;
-		if (mod->core_layout.size > MAX_LTOFF)
+		struct module_memory *mod_mem;
+
+		mod_mem = &mod->mem[MOD_DATA];
+		if (mod_mem->size > MAX_LTOFF)
 			/*
 			 * This takes advantage of fact that SHF_ARCH_SMALL gets allocated
 			 * at the end of the module.
 			 */
-			gp = mod->core_layout.size - MAX_LTOFF / 2;
+			gp = mod_mem->size - MAX_LTOFF / 2;
 		else
-			gp = mod->core_layout.size / 2;
-		gp = (uint64_t) mod->core_layout.base + ((gp + 7) & -8);
+			gp = mod_mem->size / 2;
+		gp = (uint64_t) mod_mem->base + ((gp + 7) & -8);
 		mod->arch.gp = gp;
 		DEBUGP("%s: placing gp at 0x%lx\n", __func__, gp);
 	}
diff --git a/arch/mips/kernel/vpe.c b/arch/mips/kernel/vpe.c
index 13294972707b..ab114c686f9d 100644
--- a/arch/mips/kernel/vpe.c
+++ b/arch/mips/kernel/vpe.c
@@ -199,18 +199,17 @@  static void layout_sections(struct module *mod, const Elf_Ehdr *hdr,
 	for (m = 0; m < ARRAY_SIZE(masks); ++m) {
 		for (i = 0; i < hdr->e_shnum; ++i) {
 			Elf_Shdr *s = &sechdrs[i];
+			struct module_memory *mod_mem;
+
+			mod_mem = &mod->mem[MOD_TEXT];
 
 			if ((s->sh_flags & masks[m][0]) != masks[m][0]
 			    || (s->sh_flags & masks[m][1])
 			    || s->sh_entsize != ~0UL)
 				continue;
 			s->sh_entsize =
-				get_offset((unsigned long *)&mod->core_layout.size, s);
+				get_offset((unsigned long *)&mod_mem->size, s);
 		}
-
-		if (m == 0)
-			mod->core_layout.text_size = mod->core_layout.size;
-
 	}
 }
 
@@ -641,7 +640,7 @@  static int vpe_elfload(struct vpe *v)
 		layout_sections(&mod, hdr, sechdrs, secstrings);
 	}
 
-	v->load_addr = alloc_progmem(mod.core_layout.size);
+	v->load_addr = alloc_progmem(mod.mod_mem[MOD_TEXT].size);
 	if (!v->load_addr)
 		return -ENOMEM;
 
diff --git a/arch/parisc/kernel/module.c b/arch/parisc/kernel/module.c
index 7df140545b22..f6e38c4d3904 100644
--- a/arch/parisc/kernel/module.c
+++ b/arch/parisc/kernel/module.c
@@ -27,9 +27,9 @@ 
  *      We are not doing SEGREL32 handling correctly. According to the ABI, we
  *      should do a value offset, like this:
  *			if (in_init(me, (void *)val))
- *				val -= (uint32_t)me->init_layout.base;
+ *				val -= (uint32_t)me->mem[MOD_INIT_TEXT].base;
  *			else
- *				val -= (uint32_t)me->core_layout.base;
+ *				val -= (uint32_t)me->mem[MOD_TEXT].base;
  *	However, SEGREL32 is used only for PARISC unwind entries, and we want
  *	those entries to have an absolute address, and not just an offset.
  *
@@ -76,25 +76,6 @@ 
  * allows us to allocate up to 4095 GOT entries. */
 #define MAX_GOTS	4095
 
-/* three functions to determine where in the module core
- * or init pieces the location is */
-static inline int in_init(struct module *me, void *loc)
-{
-	return (loc >= me->init_layout.base &&
-		loc <= (me->init_layout.base + me->init_layout.size));
-}
-
-static inline int in_core(struct module *me, void *loc)
-{
-	return (loc >= me->core_layout.base &&
-		loc <= (me->core_layout.base + me->core_layout.size));
-}
-
-static inline int in_local(struct module *me, void *loc)
-{
-	return in_init(me, loc) || in_core(me, loc);
-}
-
 #ifndef CONFIG_64BIT
 struct got_entry {
 	Elf32_Addr addr;
@@ -302,6 +283,7 @@  int module_frob_arch_sections(CONST Elf_Ehdr *hdr,
 {
 	unsigned long gots = 0, fdescs = 0, len;
 	unsigned int i;
+	struct module_memory *mod_mem;
 
 	len = hdr->e_shnum * sizeof(me->arch.section[0]);
 	me->arch.section = kzalloc(len, GFP_KERNEL);
@@ -346,14 +328,15 @@  int module_frob_arch_sections(CONST Elf_Ehdr *hdr,
 		me->arch.section[s].stub_entries += count;
 	}
 
+	mod_mem = &me->mem[MOD_TEXT];
 	/* align things a bit */
-	me->core_layout.size = ALIGN(me->core_layout.size, 16);
-	me->arch.got_offset = me->core_layout.size;
-	me->core_layout.size += gots * sizeof(struct got_entry);
+	mod_mem->size = ALIGN(mod_mem->size, 16);
+	me->arch.got_offset = mod_mem->size;
+	mod_mem->size += gots * sizeof(struct got_entry);
 
-	me->core_layout.size = ALIGN(me->core_layout.size, 16);
-	me->arch.fdesc_offset = me->core_layout.size;
-	me->core_layout.size += fdescs * sizeof(Elf_Fdesc);
+	mod_mem->size = ALIGN(mod_mem->size, 16);
+	me->arch.fdesc_offset = mod_mem->size;
+	mod_mem->size += fdescs * sizeof(Elf_Fdesc);
 
 	me->arch.got_max = gots;
 	me->arch.fdesc_max = fdescs;
@@ -371,7 +354,7 @@  static Elf64_Word get_got(struct module *me, unsigned long value, long addend)
 
 	BUG_ON(value == 0);
 
-	got = me->core_layout.base + me->arch.got_offset;
+	got = me->mem[MOD_TEXT].base + me->arch.got_offset;
 	for (i = 0; got[i].addr; i++)
 		if (got[i].addr == value)
 			goto out;
@@ -389,7 +372,7 @@  static Elf64_Word get_got(struct module *me, unsigned long value, long addend)
 #ifdef CONFIG_64BIT
 static Elf_Addr get_fdesc(struct module *me, unsigned long value)
 {
-	Elf_Fdesc *fdesc = me->core_layout.base + me->arch.fdesc_offset;
+	Elf_Fdesc *fdesc = me->mem[MOD_TEXT].base + me->arch.fdesc_offset;
 
 	if (!value) {
 		printk(KERN_ERR "%s: zero OPD requested!\n", me->name);
@@ -407,7 +390,7 @@  static Elf_Addr get_fdesc(struct module *me, unsigned long value)
 
 	/* Create new one */
 	fdesc->addr = value;
-	fdesc->gp = (Elf_Addr)me->core_layout.base + me->arch.got_offset;
+	fdesc->gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
 	return (Elf_Addr)fdesc;
 }
 #endif /* CONFIG_64BIT */
@@ -742,7 +725,7 @@  int apply_relocate_add(Elf_Shdr *sechdrs,
 			       loc, val);
 			val += addend;
 			/* can we reach it locally? */
-			if (in_local(me, (void *)val)) {
+			if (within_module(val, me)) {
 				/* this is the case where the symbol is local
 				 * to the module, but in a different section,
 				 * so stub the jump in case it's more than 22
@@ -801,7 +784,7 @@  int apply_relocate_add(Elf_Shdr *sechdrs,
 			break;
 		case R_PARISC_FPTR64:
 			/* 64-bit function address */
-			if(in_local(me, (void *)(val + addend))) {
+			if (within_module(val + addend, me)) {
 				*loc64 = get_fdesc(me, val+addend);
 				pr_debug("FDESC for %s at %llx points to %llx\n",
 				       strtab + sym->st_name, *loc64,
@@ -839,7 +822,7 @@  register_unwind_table(struct module *me,
 
 	table = (unsigned char *)sechdrs[me->arch.unwind_section].sh_addr;
 	end = table + sechdrs[me->arch.unwind_section].sh_size;
-	gp = (Elf_Addr)me->core_layout.base + me->arch.got_offset;
+	gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
 
 	pr_debug("register_unwind_table(), sect = %d at 0x%p - 0x%p (gp=0x%lx)\n",
 	       me->arch.unwind_section, table, end, gp);
@@ -977,7 +960,7 @@  void module_arch_cleanup(struct module *mod)
 #ifdef CONFIG_64BIT
 void *dereference_module_function_descriptor(struct module *mod, void *ptr)
 {
-	unsigned long start_opd = (Elf64_Addr)mod->core_layout.base +
+	unsigned long start_opd = (Elf64_Addr)mod->mem[MOD_TEXT].base +
 				   mod->arch.fdesc_offset;
 	unsigned long end_opd = start_opd +
 				mod->arch.fdesc_count * sizeof(Elf64_Fdesc);
diff --git a/arch/powerpc/kernel/module_32.c b/arch/powerpc/kernel/module_32.c
index ea6536171778..2729179e9dc0 100644
--- a/arch/powerpc/kernel/module_32.c
+++ b/arch/powerpc/kernel/module_32.c
@@ -163,8 +163,7 @@  static uint32_t do_plt_call(void *location,
 
 	pr_debug("Doing plt for call to 0x%x at 0x%x\n", val, (unsigned int)location);
 	/* Init, or core PLT? */
-	if (location >= mod->core_layout.base
-	    && location < mod->core_layout.base + mod->core_layout.size)
+	if (within_module_init(location, mod))
 		entry = (void *)sechdrs[mod->arch.core_plt_section].sh_addr;
 	else
 		entry = (void *)sechdrs[mod->arch.init_plt_section].sh_addr;
@@ -322,14 +321,14 @@  notrace int module_trampoline_target(struct module *mod, unsigned long addr,
 
 int module_finalize_ftrace(struct module *module, const Elf_Shdr *sechdrs)
 {
-	module->arch.tramp = do_plt_call(module->core_layout.base,
+	module->arch.tramp = do_plt_call(module->mem[MOD_TEXT].base,
 					 (unsigned long)ftrace_caller,
 					 sechdrs, module);
 	if (!module->arch.tramp)
 		return -ENOENT;
 
 #ifdef CONFIG_DYNAMIC_FTRACE_WITH_REGS
-	module->arch.tramp_regs = do_plt_call(module->core_layout.base,
+	module->arch.tramp_regs = do_plt_call(module->mem[MOD_TEXT].base,
 					      (unsigned long)ftrace_regs_caller,
 					      sechdrs, module);
 	if (!module->arch.tramp_regs)
diff --git a/arch/s390/kernel/module.c b/arch/s390/kernel/module.c
index 2d159b32885b..3ed455b785ad 100644
--- a/arch/s390/kernel/module.c
+++ b/arch/s390/kernel/module.c
@@ -126,6 +126,7 @@  int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
 	Elf_Rela *rela;
 	char *strings;
 	int nrela, i, j;
+	struct module_memory *mod_mem;
 
 	/* Find symbol table and string table. */
 	symtab = NULL;
@@ -173,14 +174,15 @@  int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
 
 	/* Increase core size by size of got & plt and set start
 	   offsets for got and plt. */
-	me->core_layout.size = ALIGN(me->core_layout.size, 4);
-	me->arch.got_offset = me->core_layout.size;
-	me->core_layout.size += me->arch.got_size;
-	me->arch.plt_offset = me->core_layout.size;
+	mod_mem = &me->mem[MOD_TEXT];
+	mod_mem->size = ALIGN(mod_mem->size, 4);
+	me->arch.got_offset = mod_mem->size;
+	mod_mem->size += me->arch.got_size;
+	me->arch.plt_offset = mod_mem->size;
 	if (me->arch.plt_size) {
 		if (IS_ENABLED(CONFIG_EXPOLINE) && !nospec_disable)
 			me->arch.plt_size += PLT_ENTRY_SIZE;
-		me->core_layout.size += me->arch.plt_size;
+		mod_mem->size += me->arch.plt_size;
 	}
 	return 0;
 }
@@ -304,7 +306,7 @@  static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
 	case R_390_GOTPLT64:	/* 64 bit offset to jump slot.	*/
 	case R_390_GOTPLTENT:	/* 32 bit rel. offset to jump slot >> 1. */
 		if (info->got_initialized == 0) {
-			Elf_Addr *gotent = me->core_layout.base +
+			Elf_Addr *gotent = me->mem[MOD_TEXT].base +
 					   me->arch.got_offset +
 					   info->got_offset;
 
@@ -329,7 +331,7 @@  static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
 			rc = apply_rela_bits(loc, val, 0, 64, 0, write);
 		else if (r_type == R_390_GOTENT ||
 			 r_type == R_390_GOTPLTENT) {
-			val += (Elf_Addr) me->core_layout.base - loc;
+			val += (Elf_Addr) me->mem[MOD_TEXT].base - loc;
 			rc = apply_rela_bits(loc, val, 1, 32, 1, write);
 		}
 		break;
@@ -345,7 +347,7 @@  static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
 			char *plt_base;
 			char *ip;
 
-			plt_base = me->core_layout.base + me->arch.plt_offset;
+			plt_base = me->mem[MOD_TEXT].base + me->arch.plt_offset;
 			ip = plt_base + info->plt_offset;
 			*(int *)insn = 0x0d10e310;	/* basr 1,0  */
 			*(int *)&insn[4] = 0x100c0004;	/* lg	1,12(1) */
@@ -375,7 +377,7 @@  static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
 			       val - loc + 0xffffUL < 0x1ffffeUL) ||
 			      (r_type == R_390_PLT32DBL &&
 			       val - loc + 0xffffffffULL < 0x1fffffffeULL)))
-				val = (Elf_Addr) me->core_layout.base +
+				val = (Elf_Addr) me->mem[MOD_TEXT].base +
 					me->arch.plt_offset +
 					info->plt_offset;
 			val += rela->r_addend - loc;
@@ -397,7 +399,7 @@  static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
 	case R_390_GOTOFF32:	/* 32 bit offset to GOT.  */
 	case R_390_GOTOFF64:	/* 64 bit offset to GOT. */
 		val = val + rela->r_addend -
-			((Elf_Addr) me->core_layout.base + me->arch.got_offset);
+			((Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset);
 		if (r_type == R_390_GOTOFF16)
 			rc = apply_rela_bits(loc, val, 0, 16, 0, write);
 		else if (r_type == R_390_GOTOFF32)
@@ -407,7 +409,7 @@  static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
 		break;
 	case R_390_GOTPC:	/* 32 bit PC relative offset to GOT. */
 	case R_390_GOTPCDBL:	/* 32 bit PC rel. off. to GOT shifted by 1. */
-		val = (Elf_Addr) me->core_layout.base + me->arch.got_offset +
+		val = (Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset +
 			rela->r_addend - loc;
 		if (r_type == R_390_GOTPC)
 			rc = apply_rela_bits(loc, val, 1, 32, 0, write);
@@ -515,7 +517,7 @@  int module_finalize(const Elf_Ehdr *hdr,
 	    !nospec_disable && me->arch.plt_size) {
 		unsigned int *ij;
 
-		ij = me->core_layout.base + me->arch.plt_offset +
+		ij = me->mem[MOD_TEXT].base + me->arch.plt_offset +
 			me->arch.plt_size - PLT_ENTRY_SIZE;
 		ij[0] = 0xc6000000;	/* exrl	%r0,.+10	*/
 		ij[1] = 0x0005a7f4;	/* j	.		*/
diff --git a/arch/x86/kernel/callthunks.c b/arch/x86/kernel/callthunks.c
index ffea98f9064b..22ab13966427 100644
--- a/arch/x86/kernel/callthunks.c
+++ b/arch/x86/kernel/callthunks.c
@@ -330,8 +330,8 @@  void noinline callthunks_patch_module_calls(struct callthunk_sites *cs,
 					    struct module *mod)
 {
 	struct core_text ct = {
-		.base = (unsigned long)mod->core_layout.base,
-		.end  = (unsigned long)mod->core_layout.base + mod->core_layout.size,
+		.base = (unsigned long)mod->mem[MOD_TEXT].base,
+		.end  = (unsigned long)mod->mem[MOD_TEXT].base + mod->mem[MOD_TEXT].size,
 		.name = mod->name,
 	};
 
diff --git a/arch/x86/kernel/module.c b/arch/x86/kernel/module.c
index 705fb2a41d7d..dbc8e226c464 100644
--- a/arch/x86/kernel/module.c
+++ b/arch/x86/kernel/module.c
@@ -339,8 +339,8 @@  int module_finalize(const Elf_Ehdr *hdr,
 	}
 	if (locks) {
 		void *lseg = (void *)locks->sh_addr;
-		void *text = me->core_layout.base;
-		void *text_end = text + me->core_layout.text_size;
+		void *text = me->mem[MOD_TEXT].base;
+		void *text_end = text + me->mem[MOD_TEXT].size;
 		alternatives_smp_module_add(me, me->name,
 					    lseg, lseg + locks->sh_size,
 					    text, text_end);
diff --git a/include/linux/module.h b/include/linux/module.h
index 8c5909c0076c..ccc9de8a93db 100644
--- a/include/linux/module.h
+++ b/include/linux/module.h
@@ -320,17 +320,47 @@  struct mod_tree_node {
 	struct latch_tree_node node;
 };
 
-struct module_layout {
-	/* The actual code + data. */
+enum mod_mem_type {
+	MOD_TEXT = 0,
+	MOD_DATA,
+	MOD_RODATA,
+	MOD_RO_AFTER_INIT,
+	MOD_INIT_TEXT,
+	MOD_INIT_DATA,
+	MOD_INIT_RODATA,
+
+	MOD_MEM_NUM_TYPES,
+	MOD_INVALID = -1,
+};
+
+#define mod_mem_type_is_init(type)	\
+	((type) == MOD_INIT_TEXT ||	\
+	 (type) == MOD_INIT_DATA ||	\
+	 (type) == MOD_INIT_RODATA)
+
+#define mod_mem_type_is_core(type) (!mod_mem_type_is_init(type))
+
+#define mod_mem_type_is_text(type)	\
+	 ((type) == MOD_TEXT ||		\
+	  (type) == MOD_INIT_TEXT)
+
+#define mod_mem_type_is_data(type) (!mod_mem_type_is_text(type))
+
+#define mod_mem_type_is_core_data(type)	\
+	(mod_mem_type_is_core(type) &&	\
+	 mod_mem_type_is_data(type))
+
+#define for_each_mod_mem_type(type)			\
+	for (enum mod_mem_type (type) = 0;		\
+	     (type) < MOD_MEM_NUM_TYPES; (type)++)
+
+#define for_class_mod_mem_type(type, class)		\
+	for_each_mod_mem_type(type)			\
+		if (mod_mem_type_is_##class(type))
+
+struct module_memory {
 	void *base;
-	/* Total size. */
 	unsigned int size;
-	/* The size of the executable code.  */
-	unsigned int text_size;
-	/* Size of RO section of the module (text+rodata) */
-	unsigned int ro_size;
-	/* Size of RO after init section */
-	unsigned int ro_after_init_size;
 
 #ifdef CONFIG_MODULES_TREE_LOOKUP
 	struct mod_tree_node mtn;
@@ -339,9 +369,9 @@  struct module_layout {
 
 #ifdef CONFIG_MODULES_TREE_LOOKUP
 /* Only touch one cacheline for common rbtree-for-core-layout case. */
-#define __module_layout_align ____cacheline_aligned
+#define __module_memory_align ____cacheline_aligned
 #else
-#define __module_layout_align
+#define __module_memory_align
 #endif
 
 struct mod_kallsyms {
@@ -418,12 +448,8 @@  struct module {
 	/* Startup function. */
 	int (*init)(void);
 
-	/* Core layout: rbtree is accessed frequently, so keep together. */
-	struct module_layout core_layout __module_layout_align;
-	struct module_layout init_layout;
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-	struct module_layout data_layout;
-#endif
+	/* rbtree is accessed frequently, so keep together. */
+	struct module_memory mem[MOD_MEM_NUM_TYPES] __module_memory_align;
 
 	/* Arch-specific module values */
 	struct mod_arch_specific arch;
@@ -573,23 +599,33 @@  bool __is_module_percpu_address(unsigned long addr, unsigned long *can_addr);
 bool is_module_percpu_address(unsigned long addr);
 bool is_module_text_address(unsigned long addr);
 
+static inline bool within_module_mem_type(unsigned long addr,
+					  const struct module *mod,
+					  enum mod_mem_type type)
+{
+	unsigned long base, size;
+
+	base = (unsigned long)mod->mem[type].base;
+	size = mod->mem[type].size;
+	return addr - base < size;
+}
+
 static inline bool within_module_core(unsigned long addr,
 				      const struct module *mod)
 {
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-	if ((unsigned long)mod->data_layout.base <= addr &&
-	    addr < (unsigned long)mod->data_layout.base + mod->data_layout.size)
+	for_class_mod_mem_type(type, core)
+		if (within_module_mem_type(addr, mod, type))
 			return true;
-#endif
-	return (unsigned long)mod->core_layout.base <= addr &&
-	       addr < (unsigned long)mod->core_layout.base + mod->core_layout.size;
+	return false;
 }
 
 static inline bool within_module_init(unsigned long addr,
 				      const struct module *mod)
 {
-	return (unsigned long)mod->init_layout.base <= addr &&
-	       addr < (unsigned long)mod->init_layout.base + mod->init_layout.size;
+	for_class_mod_mem_type(type, init)
+		if (within_module_mem_type(addr, mod, type))
+			return true;
+	return false;
 }
 
 static inline bool within_module(unsigned long addr, const struct module *mod)
diff --git a/kernel/module/internal.h b/kernel/module/internal.h
index 2e2bf236f558..c6c44ceca07a 100644
--- a/kernel/module/internal.h
+++ b/kernel/module/internal.h
@@ -17,27 +17,19 @@ 
 #define ARCH_SHF_SMALL 0
 #endif
 
-/* If this is set, the section belongs in the init part of the module */
-#define INIT_OFFSET_MASK (1UL << (BITS_PER_LONG - 1))
-/* Maximum number of characters written by module_flags() */
-#define MODULE_FLAGS_BUF_SIZE (TAINT_FLAGS_COUNT + 4)
-
-#ifndef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-#define	data_layout core_layout
-#endif
-
 /*
- * Modules' sections will be aligned on page boundaries
- * to ensure complete separation of code and data, but
- * only when CONFIG_STRICT_MODULE_RWX=y
+ * Use highest 4 bits of sh_entsize to store the mod_mem_type of this
+ * section. This leaves 28 bits for offset on 32-bit systems, which is
+ * about 256 MiB (WARN_ON_ONCE if we exceed that).
  */
-static inline unsigned int strict_align(unsigned int size)
-{
-	if (IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
-		return PAGE_ALIGN(size);
-	else
-		return size;
-}
+
+#define SH_ENTSIZE_TYPE_BITS	4
+#define SH_ENTSIZE_TYPE_SHIFT	(BITS_PER_LONG - SH_ENTSIZE_TYPE_BITS)
+#define SH_ENTSIZE_TYPE_MASK	((1UL << SH_ENTSIZE_TYPE_BITS) - 1)
+#define SH_ENTSIZE_OFFSET_MASK	((1UL << (BITS_PER_LONG - SH_ENTSIZE_TYPE_BITS)) - 1)
+
+/* Maximum number of characters written by module_flags() */
+#define MODULE_FLAGS_BUF_SIZE (TAINT_FLAGS_COUNT + 4)
 
 extern struct mutex module_mutex;
 extern struct list_head modules;
@@ -101,8 +93,8 @@  int try_to_force_load(struct module *mod, const char *reason);
 bool find_symbol(struct find_symbol_arg *fsa);
 struct module *find_module_all(const char *name, size_t len, bool even_unformed);
 int cmp_name(const void *name, const void *sym);
-long module_get_offset(struct module *mod, unsigned int *size, Elf_Shdr *sechdr,
-		       unsigned int section);
+long module_get_offset_and_type(struct module *mod, enum mod_mem_type type,
+				Elf_Shdr *sechdr, unsigned int section);
 char *module_flags(struct module *mod, char *buf, bool show_state);
 size_t module_flags_taint(unsigned long taints, char *buf);
 
@@ -190,10 +182,13 @@  struct mod_tree_root {
 #endif
 	unsigned long addr_min;
 	unsigned long addr_max;
+#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
+	unsigned long data_addr_min;
+	unsigned long data_addr_max;
+#endif
 };
 
 extern struct mod_tree_root mod_tree;
-extern struct mod_tree_root mod_data_tree;
 
 #ifdef CONFIG_MODULES_TREE_LOOKUP
 void mod_tree_insert(struct module *mod);
@@ -224,7 +219,6 @@  void module_enable_nx(const struct module *mod);
 void module_enable_x(const struct module *mod);
 int module_enforce_rwx_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
 				char *secstrings, struct module *mod);
-bool module_check_misalignment(const struct module *mod);
 
 #ifdef CONFIG_MODULE_SIG
 int module_sig_check(struct load_info *info, int flags);
diff --git a/kernel/module/kallsyms.c b/kernel/module/kallsyms.c
index 4523f99b0358..acdca7692961 100644
--- a/kernel/module/kallsyms.c
+++ b/kernel/module/kallsyms.c
@@ -78,6 +78,7 @@  static bool is_core_symbol(const Elf_Sym *src, const Elf_Shdr *sechdrs,
 			   unsigned int shnum, unsigned int pcpundx)
 {
 	const Elf_Shdr *sec;
+	enum mod_mem_type type;
 
 	if (src->st_shndx == SHN_UNDEF ||
 	    src->st_shndx >= shnum ||
@@ -90,11 +91,12 @@  static bool is_core_symbol(const Elf_Sym *src, const Elf_Shdr *sechdrs,
 #endif
 
 	sec = sechdrs + src->st_shndx;
+	type = sec->sh_entsize >> SH_ENTSIZE_TYPE_SHIFT;
 	if (!(sec->sh_flags & SHF_ALLOC)
 #ifndef CONFIG_KALLSYMS_ALL
 	    || !(sec->sh_flags & SHF_EXECINSTR)
 #endif
-	    || (sec->sh_entsize & INIT_OFFSET_MASK))
+	    || mod_mem_type_is_init(type))
 		return false;
 
 	return true;
@@ -113,11 +115,13 @@  void layout_symtab(struct module *mod, struct load_info *info)
 	Elf_Shdr *strsect = info->sechdrs + info->index.str;
 	const Elf_Sym *src;
 	unsigned int i, nsrc, ndst, strtab_size = 0;
+	struct module_memory *mod_mem_data = &mod->mem[MOD_DATA];
+	struct module_memory *mod_mem_init_data = &mod->mem[MOD_INIT_DATA];
 
 	/* Put symbol section at end of init part of module. */
 	symsect->sh_flags |= SHF_ALLOC;
-	symsect->sh_entsize = module_get_offset(mod, &mod->init_layout.size, symsect,
-						info->index.sym) | INIT_OFFSET_MASK;
+	symsect->sh_entsize = module_get_offset_and_type(mod, MOD_INIT_DATA,
+							 symsect, info->index.sym);
 	pr_debug("\t%s\n", info->secstrings + symsect->sh_name);
 
 	src = (void *)info->hdr + symsect->sh_offset;
@@ -134,28 +138,27 @@  void layout_symtab(struct module *mod, struct load_info *info)
 	}
 
 	/* Append room for core symbols at end of core part. */
-	info->symoffs = ALIGN(mod->data_layout.size, symsect->sh_addralign ?: 1);
-	info->stroffs = mod->data_layout.size = info->symoffs + ndst * sizeof(Elf_Sym);
-	mod->data_layout.size += strtab_size;
+	info->symoffs = ALIGN(mod_mem_data->size, symsect->sh_addralign ?: 1);
+	info->stroffs = mod_mem_data->size = info->symoffs + ndst * sizeof(Elf_Sym);
+	mod_mem_data->size += strtab_size;
 	/* Note add_kallsyms() computes strtab_size as core_typeoffs - stroffs */
-	info->core_typeoffs = mod->data_layout.size;
-	mod->data_layout.size += ndst * sizeof(char);
-	mod->data_layout.size = strict_align(mod->data_layout.size);
+	info->core_typeoffs = mod_mem_data->size;
+	mod_mem_data->size += ndst * sizeof(char);
 
 	/* Put string table section at end of init part of module. */
 	strsect->sh_flags |= SHF_ALLOC;
-	strsect->sh_entsize = module_get_offset(mod, &mod->init_layout.size, strsect,
-						info->index.str) | INIT_OFFSET_MASK;
+	strsect->sh_entsize = module_get_offset_and_type(mod, MOD_INIT_DATA,
+							 strsect, info->index.str);
 	pr_debug("\t%s\n", info->secstrings + strsect->sh_name);
 
 	/* We'll tack temporary mod_kallsyms on the end. */
-	mod->init_layout.size = ALIGN(mod->init_layout.size,
+	mod_mem_init_data->size = ALIGN(mod_mem_init_data->size,
 					__alignof__(struct mod_kallsyms));
-	info->mod_kallsyms_init_off = mod->init_layout.size;
-	mod->init_layout.size += sizeof(struct mod_kallsyms);
-	info->init_typeoffs = mod->init_layout.size;
-	mod->init_layout.size += nsrc * sizeof(char);
-	mod->init_layout.size = strict_align(mod->init_layout.size);
+	info->mod_kallsyms_init_off = mod_mem_init_data->size;
+
+	mod_mem_init_data->size += sizeof(struct mod_kallsyms);
+	info->init_typeoffs = mod_mem_init_data->size;
+	mod_mem_init_data->size += nsrc * sizeof(char);
 }
 
 /*
@@ -171,9 +174,11 @@  void add_kallsyms(struct module *mod, const struct load_info *info)
 	char *s;
 	Elf_Shdr *symsec = &info->sechdrs[info->index.sym];
 	unsigned long strtab_size;
+	void *data_base = mod->mem[MOD_DATA].base;
+	void *init_data_base = mod->mem[MOD_INIT_DATA].base;
 
 	/* Set up to point into init section. */
-	mod->kallsyms = (void __rcu *)mod->init_layout.base +
+	mod->kallsyms = (void __rcu *)init_data_base +
 		info->mod_kallsyms_init_off;
 
 	rcu_read_lock();
@@ -183,15 +188,15 @@  void add_kallsyms(struct module *mod, const struct load_info *info)
 	/* Make sure we get permanent strtab: don't use info->strtab. */
 	rcu_dereference(mod->kallsyms)->strtab =
 		(void *)info->sechdrs[info->index.str].sh_addr;
-	rcu_dereference(mod->kallsyms)->typetab = mod->init_layout.base + info->init_typeoffs;
+	rcu_dereference(mod->kallsyms)->typetab = init_data_base + info->init_typeoffs;
 
 	/*
 	 * Now populate the cut down core kallsyms for after init
 	 * and set types up while we still have access to sections.
 	 */
-	mod->core_kallsyms.symtab = dst = mod->data_layout.base + info->symoffs;
-	mod->core_kallsyms.strtab = s = mod->data_layout.base + info->stroffs;
-	mod->core_kallsyms.typetab = mod->data_layout.base + info->core_typeoffs;
+	mod->core_kallsyms.symtab = dst = data_base + info->symoffs;
+	mod->core_kallsyms.strtab = s = data_base + info->stroffs;
+	mod->core_kallsyms.typetab = data_base + info->core_typeoffs;
 	strtab_size = info->core_typeoffs - info->stroffs;
 	src = rcu_dereference(mod->kallsyms)->symtab;
 	for (ndst = i = 0; i < rcu_dereference(mod->kallsyms)->num_symtab; i++) {
@@ -267,12 +272,15 @@  static const char *find_kallsyms_symbol(struct module *mod,
 	unsigned int i, best = 0;
 	unsigned long nextval, bestval;
 	struct mod_kallsyms *kallsyms = rcu_dereference_sched(mod->kallsyms);
+	struct module_memory *mod_mem;
 
 	/* At worse, next value is at end of module */
 	if (within_module_init(addr, mod))
-		nextval = (unsigned long)mod->init_layout.base + mod->init_layout.text_size;
+		mod_mem = &mod->mem[MOD_INIT_TEXT];
 	else
-		nextval = (unsigned long)mod->core_layout.base + mod->core_layout.text_size;
+		mod_mem = &mod->mem[MOD_TEXT];
+
+	nextval = (unsigned long)mod_mem->base + mod_mem->size;
 
 	bestval = kallsyms_symbol_value(&kallsyms->symtab[best]);
 
diff --git a/kernel/module/kdb.c b/kernel/module/kdb.c
index f4317f92e189..995c32d3698f 100644
--- a/kernel/module/kdb.c
+++ b/kernel/module/kdb.c
@@ -26,10 +26,11 @@  int kdb_lsmod(int argc, const char **argv)
 		if (mod->state == MODULE_STATE_UNFORMED)
 			continue;
 
-		kdb_printf("%-20s%8u", mod->name, mod->core_layout.size);
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-		kdb_printf("/%8u", mod->data_layout.size);
-#endif
+		kdb_printf("%-20s%8u", mod->name, mod->mem[MOD_TEXT].size);
+		kdb_printf("/%8u", mod->mem[MOD_RODATA].size);
+		kdb_printf("/%8u", mod->mem[MOD_RO_AFTER_INIT].size);
+		kdb_printf("/%8u", mod->mem[MOD_DATA].size);
+
 		kdb_printf("  0x%px ", (void *)mod);
 #ifdef CONFIG_MODULE_UNLOAD
 		kdb_printf("%4d ", module_refcount(mod));
@@ -40,10 +41,10 @@  int kdb_lsmod(int argc, const char **argv)
 			kdb_printf(" (Loading)");
 		else
 			kdb_printf(" (Live)");
-		kdb_printf(" 0x%px", mod->core_layout.base);
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-		kdb_printf("/0x%px", mod->data_layout.base);
-#endif
+		kdb_printf(" 0x%px", mod->mem[MOD_TEXT].base);
+		kdb_printf("/0x%px", mod->mem[MOD_RODATA].base);
+		kdb_printf("/0x%px", mod->mem[MOD_RO_AFTER_INIT].base);
+		kdb_printf("/0x%px", mod->mem[MOD_DATA].base);
 
 #ifdef CONFIG_MODULE_UNLOAD
 		{
diff --git a/kernel/module/main.c b/kernel/module/main.c
index d3be89de706d..9f890d14072c 100644
--- a/kernel/module/main.c
+++ b/kernel/module/main.c
@@ -80,12 +80,6 @@  struct mod_tree_root mod_tree __cacheline_aligned = {
 	.addr_min = -1UL,
 };
 
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-struct mod_tree_root mod_data_tree __cacheline_aligned = {
-	.addr_min = -1UL,
-};
-#endif
-
 struct symsearch {
 	const struct kernel_symbol *start, *stop;
 	const s32 *crcs;
@@ -93,14 +87,24 @@  struct symsearch {
 };
 
 /*
- * Bounds of module text, for speeding up __module_address.
+ * Bounds of module memory, for speeding up __module_address.
  * Protected by module_mutex.
  */
-static void __mod_update_bounds(void *base, unsigned int size, struct mod_tree_root *tree)
+static void __mod_update_bounds(enum mod_mem_type type __maybe_unused, void *base,
+				unsigned int size, struct mod_tree_root *tree)
 {
 	unsigned long min = (unsigned long)base;
 	unsigned long max = min + size;
 
+#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
+	if (mod_mem_type_is_core_data(type)) {
+		if (min < tree->data_addr_min)
+			tree->data_addr_min = min;
+		if (max > tree->data_addr_max)
+			tree->data_addr_max = max;
+		return;
+	}
+#endif
 	if (min < tree->addr_min)
 		tree->addr_min = min;
 	if (max > tree->addr_max)
@@ -109,12 +113,12 @@  static void __mod_update_bounds(void *base, unsigned int size, struct mod_tree_r
 
 static void mod_update_bounds(struct module *mod)
 {
-	__mod_update_bounds(mod->core_layout.base, mod->core_layout.size, &mod_tree);
-	if (mod->init_layout.size)
-		__mod_update_bounds(mod->init_layout.base, mod->init_layout.size, &mod_tree);
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-	__mod_update_bounds(mod->data_layout.base, mod->data_layout.size, &mod_data_tree);
-#endif
+	for_each_mod_mem_type(type) {
+		struct module_memory *mod_mem = &mod->mem[type];
+
+		if (mod_mem->size)
+			__mod_update_bounds(type, mod_mem->base, mod_mem->size, &mod_tree);
+	}
 }
 
 /* Block module loading/unloading? */
@@ -923,12 +927,27 @@  static ssize_t store_uevent(struct module_attribute *mattr,
 struct module_attribute module_uevent =
 	__ATTR(uevent, 0200, NULL, store_uevent);
 
+#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
+
 static ssize_t show_coresize(struct module_attribute *mattr,
 			     struct module_kobject *mk, char *buffer)
 {
-	return sprintf(buffer, "%u\n", mk->mod->core_layout.size);
+	return sprintf(buffer, "%u\n", mk->mod->mem[MOD_TEXT].size);
 }
 
+#else
+
+static ssize_t show_coresize(struct module_attribute *mattr,
+			     struct module_kobject *mk, char *buffer)
+{
+	unsigned int size = 0;
+
+	for_class_mod_mem_type(type, core)
+		size += mk->mod->mem[type].size;
+	return sprintf(buffer, "%u\n", size);
+}
+#endif
+
 static struct module_attribute modinfo_coresize =
 	__ATTR(coresize, 0444, show_coresize, NULL);
 
@@ -936,7 +955,14 @@  static struct module_attribute modinfo_coresize =
 static ssize_t show_datasize(struct module_attribute *mattr,
 			     struct module_kobject *mk, char *buffer)
 {
-	return sprintf(buffer, "%u\n", mk->mod->data_layout.size);
+	unsigned int size = 0;
+
+	for_class_mod_mem_type(type, core) {
+		if (type == MOD_TEXT)
+			continue;
+		size += mk->mod->mem[type].size;
+	}
+	return sprintf(buffer, "%u\n", size);
 }
 
 static struct module_attribute modinfo_datasize =
@@ -946,7 +972,11 @@  static struct module_attribute modinfo_datasize =
 static ssize_t show_initsize(struct module_attribute *mattr,
 			     struct module_kobject *mk, char *buffer)
 {
-	return sprintf(buffer, "%u\n", mk->mod->init_layout.size);
+	unsigned int size = 0;
+
+	for_class_mod_mem_type(type, init)
+		size += mk->mod->mem[type].size;
+	return sprintf(buffer, "%u\n", size);
 }
 
 static struct module_attribute modinfo_initsize =
@@ -1143,6 +1173,63 @@  void __weak module_arch_freeing_init(struct module *mod)
 {
 }
 
+#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
+static bool mod_mem_use_vmalloc(enum mod_mem_type type)
+{
+	return mod_mem_is_core_data(type);
+}
+#else
+static bool mod_mem_use_vmalloc(enum mod_mem_type type)
+{
+	return false;
+}
+#endif
+
+static void *module_memory_alloc(unsigned int size, enum mod_mem_type type)
+{
+	if (mod_mem_use_vmalloc(type))
+		return vzalloc(size);
+	return module_alloc(size);
+}
+
+static void module_memory_free(void *ptr, enum mod_mem_type type)
+{
+	if (mod_mem_use_vmalloc(type))
+		vfree(ptr);
+	else
+		module_memfree(ptr);
+}
+
+static void free_mod_mem(struct module *mod)
+{
+	/* free the memory in the right order to avoid use-after-free */
+	static enum mod_mem_type mod_mem_free_order[MOD_MEM_NUM_TYPES] = {
+		/* first free init sections */
+		MOD_INIT_TEXT,
+		MOD_INIT_DATA,
+		MOD_INIT_RODATA,
+
+		/* then core sections, except rw data */
+		MOD_TEXT,
+		MOD_RODATA,
+		MOD_RO_AFTER_INIT,
+
+		/* last, rw data */
+		MOD_DATA,
+	};
+	int i;
+
+	for (i = 0; i < MOD_MEM_NUM_TYPES; i++) {
+		enum mod_mem_type type = mod_mem_free_order[i];
+		struct module_memory *mod_mem = &mod->mem[type];
+
+		/* Free lock-classes; relies on the preceding sync_rcu(). */
+		lockdep_free_key_range(mod_mem->base, mod_mem->size);
+		if (mod_mem->size)
+			module_memory_free(mod_mem->base, type);
+	}
+}
+
 /* Free a module, remove from lists, etc. */
 static void free_module(struct module *mod)
 {
@@ -1189,18 +1276,10 @@  static void free_module(struct module *mod)
 
 	/* This may be empty, but that's OK */
 	module_arch_freeing_init(mod);
-	module_memfree(mod->init_layout.base);
 	kfree(mod->args);
 	percpu_modfree(mod);
 
-	/* Free lock-classes; relies on the preceding sync_rcu(). */
-	lockdep_free_key_range(mod->data_layout.base, mod->data_layout.size);
-
-	/* Finally, free the core (containing the module structure) */
-	module_memfree(mod->core_layout.base);
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-	vfree(mod->data_layout.base);
-#endif
+	free_mod_mem(mod);
 }
 
 void *__symbol_get(const char *symbol)
@@ -1387,16 +1466,18 @@  unsigned int __weak arch_mod_section_prepend(struct module *mod,
 	return 0;
 }
 
-/* Update size with this section: return offset. */
-long module_get_offset(struct module *mod, unsigned int *size,
+long module_get_offset_and_type(struct module *mod, enum mod_mem_type type,
 				Elf_Shdr *sechdr, unsigned int section)
 {
-	long ret;
+	long offset;
+	long mask = ((unsigned long)(type) & SH_ENTSIZE_TYPE_MASK) << SH_ENTSIZE_TYPE_SHIFT;
 
-	*size += arch_mod_section_prepend(mod, section);
-	ret = ALIGN(*size, sechdr->sh_addralign ?: 1);
-	*size = ret + sechdr->sh_size;
-	return ret;
+	mod->mem[type].size += arch_mod_section_prepend(mod, section);
+	offset = ALIGN(mod->mem[type].size, sechdr->sh_addralign ?: 1);
+	mod->mem[type].size = offset + sechdr->sh_size;
+
+	WARN_ON_ONCE(offset & mask);
+	return offset | mask;
 }
 
 static bool module_init_layout_section(const char *sname)
@@ -1428,6 +1509,20 @@  static void layout_sections(struct module *mod, struct load_info *info)
 		{ SHF_WRITE | SHF_ALLOC, ARCH_SHF_SMALL },
 		{ ARCH_SHF_SMALL | SHF_ALLOC, 0 }
 	};
+	static int core_m_to_mem_type[] = {
+		MOD_TEXT,
+		MOD_RODATA,
+		MOD_RO_AFTER_INIT,
+		MOD_DATA,
+		MOD_INVALID,
+	};
+	static int init_m_to_mem_type[] = {
+		MOD_INIT_TEXT,
+		MOD_INIT_RODATA,
+		MOD_INVALID,
+		MOD_INIT_DATA,
+		MOD_INVALID,
+	};
 	unsigned int m, i;
 
 	for (i = 0; i < info->hdr->e_shnum; i++)
@@ -1435,41 +1530,30 @@  static void layout_sections(struct module *mod, struct load_info *info)
 
 	pr_debug("Core section allocation order:\n");
 	for (m = 0; m < ARRAY_SIZE(masks); ++m) {
+		enum mod_mem_type type = core_m_to_mem_type[m];
+
 		for (i = 0; i < info->hdr->e_shnum; ++i) {
 			Elf_Shdr *s = &info->sechdrs[i];
 			const char *sname = info->secstrings + s->sh_name;
-			unsigned int *sizep;
 
 			if ((s->sh_flags & masks[m][0]) != masks[m][0]
 			    || (s->sh_flags & masks[m][1])
 			    || s->sh_entsize != ~0UL
 			    || module_init_layout_section(sname))
 				continue;
-			sizep = m ? &mod->data_layout.size : &mod->core_layout.size;
-			s->sh_entsize = module_get_offset(mod, sizep, s, i);
+
+			if (WARN_ON_ONCE(type == MOD_INVALID))
+				continue;
+
+			s->sh_entsize = module_get_offset_and_type(mod, type, s, i);
 			pr_debug("\t%s\n", sname);
 		}
-		switch (m) {
-		case 0: /* executable */
-			mod->core_layout.size = strict_align(mod->core_layout.size);
-			mod->core_layout.text_size = mod->core_layout.size;
-			break;
-		case 1: /* RO: text and ro-data */
-			mod->data_layout.size = strict_align(mod->data_layout.size);
-			mod->data_layout.ro_size = mod->data_layout.size;
-			break;
-		case 2: /* RO after init */
-			mod->data_layout.size = strict_align(mod->data_layout.size);
-			mod->data_layout.ro_after_init_size = mod->data_layout.size;
-			break;
-		case 4: /* whole core */
-			mod->data_layout.size = strict_align(mod->data_layout.size);
-			break;
-		}
 	}
 
 	pr_debug("Init section allocation order:\n");
 	for (m = 0; m < ARRAY_SIZE(masks); ++m) {
+		enum mod_mem_type type = init_m_to_mem_type[m];
+
 		for (i = 0; i < info->hdr->e_shnum; ++i) {
 			Elf_Shdr *s = &info->sechdrs[i];
 			const char *sname = info->secstrings + s->sh_name;
@@ -1479,30 +1563,13 @@  static void layout_sections(struct module *mod, struct load_info *info)
 			    || s->sh_entsize != ~0UL
 			    || !module_init_layout_section(sname))
 				continue;
-			s->sh_entsize = (module_get_offset(mod, &mod->init_layout.size, s, i)
-					 | INIT_OFFSET_MASK);
+
+			if (WARN_ON_ONCE(type == MOD_INVALID))
+				continue;
+
+			s->sh_entsize = module_get_offset_and_type(mod, type, s, i);
 			pr_debug("\t%s\n", sname);
 		}
-		switch (m) {
-		case 0: /* executable */
-			mod->init_layout.size = strict_align(mod->init_layout.size);
-			mod->init_layout.text_size = mod->init_layout.size;
-			break;
-		case 1: /* RO: text and ro-data */
-			mod->init_layout.size = strict_align(mod->init_layout.size);
-			mod->init_layout.ro_size = mod->init_layout.size;
-			break;
-		case 2:
-			/*
-			 * RO after init doesn't apply to init_layout (only
-			 * core_layout), so it just takes the value of ro_size.
-			 */
-			mod->init_layout.ro_after_init_size = mod->init_layout.ro_size;
-			break;
-		case 4: /* whole init */
-			mod->init_layout.size = strict_align(mod->init_layout.size);
-			break;
-		}
 	}
 }
 
@@ -2122,72 +2189,42 @@  static int move_module(struct module *mod, struct load_info *info)
 {
 	int i;
 	void *ptr;
+	enum mod_mem_type t;
 
-	/* Do the allocs. */
-	ptr = module_alloc(mod->core_layout.size);
-	/*
-	 * The pointer to this block is stored in the module structure
-	 * which is inside the block. Just mark it as not being a
-	 * leak.
-	 */
-	kmemleak_not_leak(ptr);
-	if (!ptr)
-		return -ENOMEM;
-
-	memset(ptr, 0, mod->core_layout.size);
-	mod->core_layout.base = ptr;
-
-	if (mod->init_layout.size) {
-		ptr = module_alloc(mod->init_layout.size);
-		/*
-		 * The pointer to this block is stored in the module structure
-		 * which is inside the block. This block doesn't need to be
-		 * scanned as it contains data and code that will be freed
-		 * after the module is initialized.
-		 */
-		kmemleak_ignore(ptr);
-		if (!ptr) {
-			module_memfree(mod->core_layout.base);
-			return -ENOMEM;
+	for_each_mod_mem_type(type) {
+		if (!mod->mem[type].size) {
+			mod->mem[type].base = NULL;
+			continue;
 		}
-		memset(ptr, 0, mod->init_layout.size);
-		mod->init_layout.base = ptr;
-	} else
-		mod->init_layout.base = NULL;
+		mod->mem[type].size = PAGE_ALIGN(mod->mem[type].size);
+		ptr = module_memory_alloc(mod->mem[type].size, type);
 
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-	/* Do the allocs. */
-	ptr = vzalloc(mod->data_layout.size);
 		/*
 		 * The pointer to this block is stored in the module structure
 		 * which is inside the block. Just mark it as not being a
 		 * leak.
 		 */
-	kmemleak_not_leak(ptr);
+		kmemleak_ignore(ptr);
 		if (!ptr) {
-		module_memfree(mod->core_layout.base);
-		module_memfree(mod->init_layout.base);
-		return -ENOMEM;
+			t = type;
+			goto out_enomem;
+		}
+		memset(ptr, 0, mod->mem[type].size);
+		mod->mem[type].base = ptr;
 	}
 
-	mod->data_layout.base = ptr;
-#endif
 	/* Transfer each section which specifies SHF_ALLOC */
 	pr_debug("final section addresses:\n");
 	for (i = 0; i < info->hdr->e_shnum; i++) {
 		void *dest;
 		Elf_Shdr *shdr = &info->sechdrs[i];
+		enum mod_mem_type type = shdr->sh_entsize >> SH_ENTSIZE_TYPE_SHIFT;
 
 		if (!(shdr->sh_flags & SHF_ALLOC))
 			continue;
 
-		if (shdr->sh_entsize & INIT_OFFSET_MASK)
-			dest = mod->init_layout.base
-				+ (shdr->sh_entsize & ~INIT_OFFSET_MASK);
-		else if (!(shdr->sh_flags & SHF_EXECINSTR))
-			dest = mod->data_layout.base + shdr->sh_entsize;
-		else
-			dest = mod->core_layout.base + shdr->sh_entsize;
+		dest = mod->mem[type].base +
+			(shdr->sh_entsize & SH_ENTSIZE_OFFSET_MASK);
 
 		if (shdr->sh_type != SHT_NOBITS)
 			memcpy(dest, (void *)shdr->sh_addr, shdr->sh_size);
@@ -2198,6 +2235,10 @@  static int move_module(struct module *mod, struct load_info *info)
 	}
 
 	return 0;
+out_enomem:
+	for (t--; t >= 0; t--)
+		module_memory_free(mod->mem[t].base, t);
+	return -ENOMEM;
 }
 
 static int check_module_license_and_versions(struct module *mod)
@@ -2242,12 +2283,13 @@  static void flush_module_icache(const struct module *mod)
 	 * Do it before processing of module parameters, so the module
 	 * can provide parameter accessor functions of its own.
 	 */
-	if (mod->init_layout.base)
-		flush_icache_range((unsigned long)mod->init_layout.base,
-				   (unsigned long)mod->init_layout.base
-				   + mod->init_layout.size);
-	flush_icache_range((unsigned long)mod->core_layout.base,
-			   (unsigned long)mod->core_layout.base + mod->core_layout.size);
+	for_each_mod_mem_type(type) {
+		const struct module_memory *mod_mem = &mod->mem[type];
+
+		if (mod_mem->size)
+			flush_icache_range((unsigned long)mod_mem->base,
+					   (unsigned long)mod_mem->base + mod_mem->size);
+	}
 }
 
 bool __weak module_elf_check_arch(Elf_Ehdr *hdr)
@@ -2350,11 +2392,8 @@  static void module_deallocate(struct module *mod, struct load_info *info)
 {
 	percpu_modfree(mod);
 	module_arch_freeing_init(mod);
-	module_memfree(mod->init_layout.base);
-	module_memfree(mod->core_layout.base);
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-	vfree(mod->data_layout.base);
-#endif
+
+	free_mod_mem(mod);
 }
 
 int __weak module_finalize(const Elf_Ehdr *hdr,
@@ -2415,7 +2454,9 @@  static void do_mod_ctors(struct module *mod)
 /* For freeing module_init on success, in case kallsyms traversing */
 struct mod_initfree {
 	struct llist_node node;
-	void *module_init;
+	void *init_text;
+	void *init_data;
+	void *init_rodata;
 };
 
 static void do_free_init(struct work_struct *w)
@@ -2429,7 +2470,9 @@  static void do_free_init(struct work_struct *w)
 
 	llist_for_each_safe(pos, n, list) {
 		initfree = container_of(pos, struct mod_initfree, node);
-		module_memfree(initfree->module_init);
+		module_memfree(initfree->init_text);
+		module_memfree(initfree->init_data);
+		module_memfree(initfree->init_rodata);
 		kfree(initfree);
 	}
 }
@@ -2456,7 +2499,9 @@  static noinline int do_init_module(struct module *mod)
 		ret = -ENOMEM;
 		goto fail;
 	}
-	freeinit->module_init = mod->init_layout.base;
+	freeinit->init_text = mod->mem[MOD_INIT_TEXT].base;
+	freeinit->init_data = mod->mem[MOD_INIT_DATA].base;
+	freeinit->init_rodata = mod->mem[MOD_INIT_RODATA].base;
 
 	do_mod_ctors(mod);
 	/* Start the module */
@@ -2492,8 +2537,8 @@  static noinline int do_init_module(struct module *mod)
 	if (!mod->async_probe_requested)
 		async_synchronize_full();
 
-	ftrace_free_mem(mod, mod->init_layout.base, mod->init_layout.base +
-			mod->init_layout.size);
+	ftrace_free_mem(mod, mod->mem[MOD_INIT_TEXT].base,
+			mod->mem[MOD_INIT_TEXT].base + mod->mem[MOD_INIT_TEXT].size);
 	mutex_lock(&module_mutex);
 	/* Drop initial reference. */
 	module_put(mod);
@@ -2505,11 +2550,10 @@  static noinline int do_init_module(struct module *mod)
 	module_enable_ro(mod, true);
 	mod_tree_remove_init(mod);
 	module_arch_freeing_init(mod);
-	mod->init_layout.base = NULL;
-	mod->init_layout.size = 0;
-	mod->init_layout.ro_size = 0;
-	mod->init_layout.ro_after_init_size = 0;
-	mod->init_layout.text_size = 0;
+	for_class_mod_mem_type(type, init) {
+		mod->mem[type].base = NULL;
+		mod->mem[type].size = 0;
+	}
 #ifdef CONFIG_DEBUG_INFO_BTF_MODULES
 	/* .BTF is not SHF_ALLOC and will get removed, so sanitize pointer */
 	mod->btf_data = NULL;
@@ -2628,9 +2672,6 @@  static int complete_formation(struct module *mod, struct load_info *info)
 	module_bug_finalize(info->hdr, info->sechdrs, mod);
 	module_cfi_finalize(info->hdr, info->sechdrs, mod);
 
-	if (module_check_misalignment(mod))
-		goto out_misaligned;
-
 	module_enable_ro(mod, false);
 	module_enable_nx(mod);
 	module_enable_x(mod);
@@ -2644,8 +2685,6 @@  static int complete_formation(struct module *mod, struct load_info *info)
 
 	return 0;
 
-out_misaligned:
-	err = -EINVAL;
 out:
 	mutex_unlock(&module_mutex);
 	return err;
@@ -2909,7 +2948,10 @@  static int load_module(struct load_info *info, const char __user *uargs,
 	mutex_unlock(&module_mutex);
  free_module:
 	/* Free lock-classes; relies on the preceding sync_rcu() */
-	lockdep_free_key_range(mod->data_layout.base, mod->data_layout.size);
+	for_class_mod_mem_type(type, core) {
+		lockdep_free_key_range(mod->mem[type].base,
+				       mod->mem[type].size);
+	}
 
 	module_deallocate(mod, info);
  free_copy:
@@ -3060,20 +3102,21 @@  bool is_module_address(unsigned long addr)
 struct module *__module_address(unsigned long addr)
 {
 	struct module *mod;
-	struct mod_tree_root *tree;
 
 	if (addr >= mod_tree.addr_min && addr <= mod_tree.addr_max)
-		tree = &mod_tree;
+		goto lookup;
+
 #ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-	else if (addr >= mod_data_tree.addr_min && addr <= mod_data_tree.addr_max)
-		tree = &mod_data_tree;
+	if (addr >= mod_tree.data_addr_min && addr <= mod_tree.data_addr_max)
+		goto lookup;
 #endif
-	else
+
 	return NULL;
 
+lookup:
 	module_assert_mutex_or_preempt();
 
-	mod = mod_find(addr, tree);
+	mod = mod_find(addr, &mod_tree);
 	if (mod) {
 		BUG_ON(!within_module(addr, mod));
 		if (mod->state == MODULE_STATE_UNFORMED)
@@ -3113,8 +3156,8 @@  struct module *__module_text_address(unsigned long addr)
 	struct module *mod = __module_address(addr);
 	if (mod) {
 		/* Make sure it's within the text section. */
-		if (!within(addr, mod->init_layout.base, mod->init_layout.text_size)
-		    && !within(addr, mod->core_layout.base, mod->core_layout.text_size))
+		if (!within_module_mem_type(addr, mod, MOD_TEXT) &&
+		    !within_module_mem_type(addr, mod, MOD_INIT_TEXT))
 			mod = NULL;
 	}
 	return mod;
diff --git a/kernel/module/procfs.c b/kernel/module/procfs.c
index cf5b9f1e6ec4..0a4841e88adb 100644
--- a/kernel/module/procfs.c
+++ b/kernel/module/procfs.c
@@ -62,6 +62,15 @@  static void m_stop(struct seq_file *m, void *p)
 	mutex_unlock(&module_mutex);
 }
 
+static unsigned int module_total_size(struct module *mod)
+{
+	int size = 0;
+
+	for_each_mod_mem_type(type)
+		size += mod->mem[type].size;
+	return size;
+}
+
 static int m_show(struct seq_file *m, void *p)
 {
 	struct module *mod = list_entry(p, struct module, list);
@@ -73,10 +82,7 @@  static int m_show(struct seq_file *m, void *p)
 	if (mod->state == MODULE_STATE_UNFORMED)
 		return 0;
 
-	size = mod->init_layout.size + mod->core_layout.size;
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-	size += mod->data_layout.size;
-#endif
+	size = module_total_size(mod);
 	seq_printf(m, "%s %u", mod->name, size);
 	print_unload_info(m, mod);
 
@@ -86,7 +92,7 @@  static int m_show(struct seq_file *m, void *p)
 		   mod->state == MODULE_STATE_COMING ? "Loading" :
 		   "Live");
 	/* Used by oprofile and other similar tools. */
-	value = m->private ? NULL : mod->core_layout.base;
+	value = m->private ? NULL : mod->mem[MOD_TEXT].base;
 	seq_printf(m, " 0x%px", value);
 
 	/* Taints info */
diff --git a/kernel/module/strict_rwx.c b/kernel/module/strict_rwx.c
index 14fbea66f12f..2e79a77f40eb 100644
--- a/kernel/module/strict_rwx.c
+++ b/kernel/module/strict_rwx.c
@@ -11,82 +11,26 @@ 
 #include <linux/set_memory.h>
 #include "internal.h"
 
-/*
- * LKM RO/NX protection: protect module's text/ro-data
- * from modification and any data from execution.
- *
- * General layout of module is:
- *          [text] [read-only-data] [ro-after-init] [writable data]
- * text_size -----^                ^               ^               ^
- * ro_size ------------------------|               |               |
- * ro_after_init_size -----------------------------|               |
- * size -----------------------------------------------------------|
- *
- * These values are always page-aligned (as is base) when
- * CONFIG_STRICT_MODULE_RWX is set.
- */
+static void module_set_memory(
+	const struct module *mod, enum mod_mem_type type,
+	int (*set_memory)(unsigned long start, int num_pages))
+{
+	const struct module_memory *mod_mem = &mod->mem[type];
+
+	set_vm_flush_reset_perms(mod_mem->base);
+	set_memory((unsigned long)mod_mem->base, mod_mem->size >> PAGE_SHIFT);
+}
 
 /*
  * Since some arches are moving towards PAGE_KERNEL module allocations instead
- * of PAGE_KERNEL_EXEC, keep frob_text() and module_enable_x() independent of
+ * of PAGE_KERNEL_EXEC, keep module_enable_x() independent of
  * CONFIG_STRICT_MODULE_RWX because they are needed regardless of whether we
  * are strict.
  */
-static void frob_text(const struct module_layout *layout,
-		      int (*set_memory)(unsigned long start, int num_pages))
-{
-	set_memory((unsigned long)layout->base,
-		   PAGE_ALIGN(layout->text_size) >> PAGE_SHIFT);
-}
-
-static void frob_rodata(const struct module_layout *layout,
-		 int (*set_memory)(unsigned long start, int num_pages))
-{
-	set_memory((unsigned long)layout->base + layout->text_size,
-		   (layout->ro_size - layout->text_size) >> PAGE_SHIFT);
-}
-
-static void frob_ro_after_init(const struct module_layout *layout,
-			int (*set_memory)(unsigned long start, int num_pages))
-{
-	set_memory((unsigned long)layout->base + layout->ro_size,
-		   (layout->ro_after_init_size - layout->ro_size) >> PAGE_SHIFT);
-}
-
-static void frob_writable_data(const struct module_layout *layout,
-			int (*set_memory)(unsigned long start, int num_pages))
-{
-	set_memory((unsigned long)layout->base + layout->ro_after_init_size,
-		   (layout->size - layout->ro_after_init_size) >> PAGE_SHIFT);
-}
-
-static bool layout_check_misalignment(const struct module_layout *layout)
-{
-	return WARN_ON(!PAGE_ALIGNED(layout->base)) ||
-	       WARN_ON(!PAGE_ALIGNED(layout->text_size)) ||
-	       WARN_ON(!PAGE_ALIGNED(layout->ro_size)) ||
-	       WARN_ON(!PAGE_ALIGNED(layout->ro_after_init_size)) ||
-	       WARN_ON(!PAGE_ALIGNED(layout->size));
-}
-
-bool module_check_misalignment(const struct module *mod)
-{
-	if (!IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
-		return false;
-
-	return layout_check_misalignment(&mod->core_layout) ||
-	       layout_check_misalignment(&mod->data_layout) ||
-	       layout_check_misalignment(&mod->init_layout);
-}
-
 void module_enable_x(const struct module *mod)
 {
-	if (!PAGE_ALIGNED(mod->core_layout.base) ||
-	    !PAGE_ALIGNED(mod->init_layout.base))
-		return;
-
-	frob_text(&mod->core_layout, set_memory_x);
-	frob_text(&mod->init_layout, set_memory_x);
+	for_class_mod_mem_type(type, text)
+		module_set_memory(mod, type, set_memory_x);
 }
 
 void module_enable_ro(const struct module *mod, bool after_init)
@@ -98,16 +42,13 @@  void module_enable_ro(const struct module *mod, bool after_init)
 		return;
 #endif
 
-	set_vm_flush_reset_perms(mod->core_layout.base);
-	set_vm_flush_reset_perms(mod->init_layout.base);
-	frob_text(&mod->core_layout, set_memory_ro);
-
-	frob_rodata(&mod->data_layout, set_memory_ro);
-	frob_text(&mod->init_layout, set_memory_ro);
-	frob_rodata(&mod->init_layout, set_memory_ro);
+	module_set_memory(mod, MOD_TEXT, set_memory_ro);
+	module_set_memory(mod, MOD_INIT_TEXT, set_memory_ro);
+	module_set_memory(mod, MOD_RODATA, set_memory_ro);
+	module_set_memory(mod, MOD_INIT_RODATA, set_memory_ro);
 
 	if (after_init)
-		frob_ro_after_init(&mod->data_layout, set_memory_ro);
+		module_set_memory(mod, MOD_RO_AFTER_INIT, set_memory_ro);
 }
 
 void module_enable_nx(const struct module *mod)
@@ -115,11 +56,8 @@  void module_enable_nx(const struct module *mod)
 	if (!IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
 		return;
 
-	frob_rodata(&mod->data_layout, set_memory_nx);
-	frob_ro_after_init(&mod->data_layout, set_memory_nx);
-	frob_writable_data(&mod->data_layout, set_memory_nx);
-	frob_rodata(&mod->init_layout, set_memory_nx);
-	frob_writable_data(&mod->init_layout, set_memory_nx);
+	for_class_mod_mem_type(type, data)
+		module_set_memory(mod, type, set_memory_nx);
 }
 
 int module_enforce_rwx_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
diff --git a/kernel/module/tree_lookup.c b/kernel/module/tree_lookup.c
index 8ec5cfd60496..277197977d43 100644
--- a/kernel/module/tree_lookup.c
+++ b/kernel/module/tree_lookup.c
@@ -21,16 +21,16 @@ 
 
 static __always_inline unsigned long __mod_tree_val(struct latch_tree_node *n)
 {
-	struct module_layout *layout = container_of(n, struct module_layout, mtn.node);
+	struct module_memory *mod_mem = container_of(n, struct module_memory, mtn.node);
 
-	return (unsigned long)layout->base;
+	return (unsigned long)mod_mem->base;
 }
 
 static __always_inline unsigned long __mod_tree_size(struct latch_tree_node *n)
 {
-	struct module_layout *layout = container_of(n, struct module_layout, mtn.node);
+	struct module_memory *mod_mem = container_of(n, struct module_memory, mtn.node);
 
-	return (unsigned long)layout->size;
+	return (unsigned long)mod_mem->size;
 }
 
 static __always_inline bool
@@ -77,32 +77,27 @@  static void __mod_tree_remove(struct mod_tree_node *node, struct mod_tree_root *
  */
 void mod_tree_insert(struct module *mod)
 {
-	mod->core_layout.mtn.mod = mod;
-	mod->init_layout.mtn.mod = mod;
-
-	__mod_tree_insert(&mod->core_layout.mtn, &mod_tree);
-	if (mod->init_layout.size)
-		__mod_tree_insert(&mod->init_layout.mtn, &mod_tree);
-
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-	mod->data_layout.mtn.mod = mod;
-	__mod_tree_insert(&mod->data_layout.mtn, &mod_data_tree);
-#endif
+	for_each_mod_mem_type(type) {
+		mod->mem[type].mtn.mod = mod;
+		if (mod->mem[type].size)
+			__mod_tree_insert(&mod->mem[type].mtn, &mod_tree);
+	}
 }
 
 void mod_tree_remove_init(struct module *mod)
 {
-	if (mod->init_layout.size)
-		__mod_tree_remove(&mod->init_layout.mtn, &mod_tree);
+	for_class_mod_mem_type(type, init) {
+		if (mod->mem[type].size)
+			__mod_tree_remove(&mod->mem[type].mtn, &mod_tree);
+	}
 }
 
 void mod_tree_remove(struct module *mod)
 {
-	__mod_tree_remove(&mod->core_layout.mtn, &mod_tree);
-	mod_tree_remove_init(mod);
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-	__mod_tree_remove(&mod->data_layout.mtn, &mod_data_tree);
-#endif
+	for_each_mod_mem_type(type) {
+		if (mod->mem[type].size)
+			__mod_tree_remove(&mod->mem[type].mtn, &mod_tree);
+	}
 }
 
 struct module *mod_find(unsigned long addr, struct mod_tree_root *tree)