From patchwork Fri Mar 14 21:39:30 2025 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Deepak Gupta X-Patchwork-Id: 14017498 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id B0FC7C282EC for ; Fri, 14 Mar 2025 21:40:01 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id D02D028001B; Fri, 14 Mar 2025 17:39:57 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id CAD00280017; Fri, 14 Mar 2025 17:39:57 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id A915528001B; Fri, 14 Mar 2025 17:39:57 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0016.hostedemail.com [216.40.44.16]) by kanga.kvack.org (Postfix) with ESMTP id 840DC280017 for ; Fri, 14 Mar 2025 17:39:57 -0400 (EDT) Received: from smtpin09.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay08.hostedemail.com (Postfix) with ESMTP id B2100140FCC for ; Fri, 14 Mar 2025 21:39:58 +0000 (UTC) X-FDA: 83221474476.09.4E062A2 Received: from mail-pl1-f169.google.com (mail-pl1-f169.google.com [209.85.214.169]) by imf09.hostedemail.com (Postfix) with ESMTP id F1251140003 for ; Fri, 14 Mar 2025 21:39:56 +0000 (UTC) Authentication-Results: imf09.hostedemail.com; dkim=pass header.d=rivosinc-com.20230601.gappssmtp.com header.s=20230601 header.b=JSZyP5QT; spf=pass (imf09.hostedemail.com: domain of debug@rivosinc.com designates 209.85.214.169 as permitted sender) smtp.mailfrom=debug@rivosinc.com; dmarc=none ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1741988397; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=DxdMdi2cubLlt4TIGXXJrBSyWEjwvqF79MEiREZoHyk=; b=KvpV69GneDYX7HS+sIbfJWBrfyp15J0c4XXIkgBbPhPPSMjHv0TKgJ5OK+67JWHgZ0wRdq BV0SetdgvDXSrPxE/G/v+1f6JA3LB/tpEE+9+cP7aWFBofOgPofVsyLcS1nmIz+lJBlBVC nasO5tRoKCkbaGpc6GqD3PdIHGwP2/s= ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1741988397; a=rsa-sha256; cv=none; b=Nh5rTyqYgYj38R7xGPG7+Bdsr7yU/WL4uym7YwFuD5hP+GeHiQ2ZL+jtGXR/1NuT2TAals 0FQ1VS3EEd30xtplAw/wU7JP3rjz7twq7TDduuVUrgWiIQco9ddlp6A0041LIVBS2PC51Y OmhiFdmpY10GMWkGL0JVxXSI8qgp53U= ARC-Authentication-Results: i=1; imf09.hostedemail.com; dkim=pass header.d=rivosinc-com.20230601.gappssmtp.com header.s=20230601 header.b=JSZyP5QT; spf=pass (imf09.hostedemail.com: domain of debug@rivosinc.com designates 209.85.214.169 as permitted sender) smtp.mailfrom=debug@rivosinc.com; dmarc=none Received: by mail-pl1-f169.google.com with SMTP id d9443c01a7336-2240b4de12bso69472695ad.2 for ; Fri, 14 Mar 2025 14:39:56 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=rivosinc-com.20230601.gappssmtp.com; s=20230601; t=1741988396; x=1742593196; darn=kvack.org; h=cc:to:in-reply-to:references:message-id:content-transfer-encoding :mime-version:subject:date:from:from:to:cc:subject:date:message-id :reply-to; bh=DxdMdi2cubLlt4TIGXXJrBSyWEjwvqF79MEiREZoHyk=; b=JSZyP5QTqeyG/YmIHcPeoJR7R/65DfBU71I5cQH5iLeo9Z9crIlu4lI/MAyxOOY8t1 At1faVwqoMU5pCVByO6W0GgwEXiwdaEiwcS4KYUhOdUc6RNnGcVDKxWyhrIuLmCqTYhO VNCAAdOF8jkoak3HORZdhoLQbw4NP4DY7b5Lhuudg3FqkcFjtlQ3l+NUhfPXDYm9pPJf ty2UR1UrLIw5UCsIfXUHROAZmLAagvSQjKcchasLipmz4ybE7HoZkEwkM4zxWwWJOpr4 GY68QdCh0ikj6aL5AZg2BTlhzzr7DaiPd0EnOx86F1G04WaqCfL1PK885s6Bw6WM1Xjl vGmw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1741988396; x=1742593196; h=cc:to:in-reply-to:references:message-id:content-transfer-encoding :mime-version:subject:date:from:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=DxdMdi2cubLlt4TIGXXJrBSyWEjwvqF79MEiREZoHyk=; b=u9dqlYBT9F65rXVg6+3eGFZLwuj8J7kvkENpa6N6nvkTzQ0L0GtwAyw7dyaEmpr5Vo 2BPe10+tifs7c1H9kOugKCuUuIpXVQzS75DGFLBNiNLJZY5JmeOA8azoFBGg1uucEsxA u+cYgwtfEGDaNwX71a1ztU0fvWGGqfUIR7kCFyTjtY+o55DIsXUXHH/ebjGYFY5S2epU Nqg7mrvp6x7P/Ua0ur8R2VSjtleIbzQdt6dRTidmSTZkJgM/1r8wAgLJhWlXVFzYeVmS MMq2nwiXdAK8jvi4DHZdThziJNTAC+QHZIgvqY+0cmNC5IGdXTLeOY4JW3FiSiYl3zH8 m+Eg== X-Forwarded-Encrypted: i=1; AJvYcCV8KCCEJkwOAdy3XrZreqLIf2NGtIZJjjm/mCBysVTdLiUuyN80JXDZ6AXLytwsp2Ylepx8PnHOpw==@kvack.org X-Gm-Message-State: AOJu0YxKoNPiJYA7q3279h4DnVPSFGcsdnJS3B1pcrz61WCWzIttWSh3 8+b1FXio3RvlzxzH/IyO/Z3MjN1oANz/o3qQ70Y+0ZYCysmphuOGBMKjIR6XDdg= X-Gm-Gg: ASbGncuqlbCEdFO4c2+qhNAaqSHdiHFJo4L3coO2OkiyAEVPzEiSH2Y5l6gxqq3r1oS WIdSuQIddwVEF8CmNeqcIUfoOYwYPmgGvoV5aiQZhA2gMKdui2LNiVWAsfvqDkO8trYEUyEGa2o uYhRN+jXSuY/+2aBdsXN7LUXZaOsFkE+/+Hh6uPPH45RCu5KscWKqorgapWWHYQPKSgyslYlR3a XNWIGWRbWA9+w5Y3VAi1e/mTJwLUvQwsBtdlTbJCZeN4nED5Q+uYuajI/8SEH9H8cz7vOe7UKAr RdRV9Z5I5QSaw8tWOIQOcRPIY8gACB8yeTAHP+lQeCN/wV6y2dPspwO9KuhTWcVftA== X-Google-Smtp-Source: AGHT+IH+xCwTKv/QOpKiV4NU+zofe+/FULTNwPU3AjuhKtiuzJi1q2DiCU071AO1IM2f30UQ36o6NQ== X-Received: by 2002:a17:903:3c6f:b0:224:1ce1:a3f4 with SMTP id d9443c01a7336-225e0a62f48mr57173645ad.1.1741988395873; Fri, 14 Mar 2025 14:39:55 -0700 (PDT) Received: from debug.ba.rivosinc.com ([64.71.180.162]) by smtp.gmail.com with ESMTPSA id d9443c01a7336-225c68a6e09sm33368855ad.55.2025.03.14.14.39.53 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 14 Mar 2025 14:39:55 -0700 (PDT) From: Deepak Gupta Date: Fri, 14 Mar 2025 14:39:30 -0700 Subject: [PATCH v12 11/28] riscv/shstk: If needed allocate a new shadow stack on clone MIME-Version: 1.0 Message-Id: <20250314-v5_user_cfi_series-v12-11-e51202b53138@rivosinc.com> References: <20250314-v5_user_cfi_series-v12-0-e51202b53138@rivosinc.com> In-Reply-To: <20250314-v5_user_cfi_series-v12-0-e51202b53138@rivosinc.com> To: Thomas Gleixner , Ingo Molnar , Borislav Petkov , Dave Hansen , x86@kernel.org, "H. Peter Anvin" , Andrew Morton , "Liam R. Howlett" , Vlastimil Babka , Lorenzo Stoakes , Paul Walmsley , Palmer Dabbelt , Albert Ou , Conor Dooley , Rob Herring , Krzysztof Kozlowski , Arnd Bergmann , Christian Brauner , Peter Zijlstra , Oleg Nesterov , Eric Biederman , Kees Cook , Jonathan Corbet , Shuah Khan , Jann Horn , Conor Dooley Cc: linux-kernel@vger.kernel.org, linux-fsdevel@vger.kernel.org, linux-mm@kvack.org, linux-riscv@lists.infradead.org, devicetree@vger.kernel.org, linux-arch@vger.kernel.org, linux-doc@vger.kernel.org, linux-kselftest@vger.kernel.org, alistair.francis@wdc.com, richard.henderson@linaro.org, jim.shu@sifive.com, andybnac@gmail.com, kito.cheng@sifive.com, charlie@rivosinc.com, atishp@rivosinc.com, evan@rivosinc.com, cleger@rivosinc.com, alexghiti@rivosinc.com, samitolvanen@google.com, broonie@kernel.org, rick.p.edgecombe@intel.com, Zong Li , Deepak Gupta X-Mailer: b4 0.14.0 X-Rspam-User: X-Rspamd-Server: rspam11 X-Rspamd-Queue-Id: F1251140003 X-Stat-Signature: 7zqs9cfbaokeixbqhzt8b1t4b3bx87tm X-HE-Tag: 1741988396-669298 X-HE-Meta: U2FsdGVkX188TwB+TNVnkhAhpJ4fxqaXOi9gXtd2M1ic7DQrLFSpkihERfKmrLtAQMkjXieD/is5/4zyjzqrUoJYxuIQpffobkvq/1x5x5aQ0o+qZc8f8i+t5t2rHPqR5AE37SuNOrbf/fxXRLqrcjJljAFag0GqRaeHXsI+EwUf/63hcUdFTRBR4C2Y8fDqF1cbPRMrAuGv9ro+sE9/hEKDJy9a4o8HTpzxHIP/Jy+zbTOWNS6b9rgAUY9YaljXQoidMa5FA7xxxFO1P5+MRV/Ul1yOBlMruS4qfqoohCwrua3dHHUtbT0zFNn2Tqx5o40JXQQGGyssgqigscHy7xWPOXCSB9YT2RpDtBiEpa0qm0M3RcoLOoke0O5WCBov1SH8u3oT1vuUPOF6B7xoKqRSZjlUTKoBauwtpgjE0MNBiU+wajfP3Kjd0XoIydjPF1FHKJ2X34sj9QbOR/v9h6eFLSKnY5aahFRAqkJ7GJ1HmkcnIPopwrCy2Zr46DBguov58AW9bN+rLIG6lpPM5U9fqu8atAPyCSih7WjsBY5oxVk1PA2FgHKKMxNMLUU/YvXhAm06jjlgESCrETlRhTClczJNMMOGj+hklk/qMfWaDkXaJB6mAY+nf4oKmtEPLXOIYTqlOtolW5dKVdMZy2wZN8DBYJMRi1VpPoBngGQPzC5CtHhDyGhUmbYibJRQgU3SKFsaXLMv8PuLzAsOr0/yw1PilWyZlftbcGqkbvv2RyTgAuxizNlekB4F0baHAOA6MP7UXXZxMQPuk/u04We4hpFZToyOPg8YSlBMtjf3u+E9wRPZRu9xC+3nsqGp7+eCkKcnKa3t4mTW9nI+j9EnJH7Hb30Eg0OuKi+whOsg+mFyUaaQ57QGXesPJbRwfPYP7oOasOj6oXbDLvcVcYKv5xD94bHl1wnXFTrk1M8Iq95rGxFey5l48/JrtYEzFr6t2INioEa1xSdJmV+ zagxcTdi h2o5ne0XqtHtg9vNAGJLR9JpwdZVDGbFl1gFLmcdnOXY82RXIF7DLuY0M2ugi5TjR2eYFyD/aUa1Lih3Q374JjwWWOQewuG7ir7NC4oDyx9q7C6Q= X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: Userspace specifies CLONE_VM to share address space and spawn new thread. `clone` allow userspace to specify a new stack for new thread. However there is no way to specify new shadow stack base address without changing API. This patch allocates a new shadow stack whenever CLONE_VM is given. In case of CLONE_VFORK, parent is suspended until child finishes and thus can child use parent shadow stack. In case of !CLONE_VM, COW kicks in because entire address space is copied from parent to child. `clone3` is extensible and can provide mechanisms using which shadow stack as an input parameter can be provided. This is not settled yet and being extensively discussed on mailing list. Once that's settled, this commit will adapt to that. Reviewed-by: Zong Li Signed-off-by: Deepak Gupta --- arch/riscv/include/asm/mmu_context.h | 7 ++ arch/riscv/include/asm/usercfi.h | 25 ++++++++ arch/riscv/kernel/process.c | 9 +++ arch/riscv/kernel/usercfi.c | 120 +++++++++++++++++++++++++++++++++++ 4 files changed, 161 insertions(+) diff --git a/arch/riscv/include/asm/mmu_context.h b/arch/riscv/include/asm/mmu_context.h index 8c4bc49a3a0f..dbf27a78df6c 100644 --- a/arch/riscv/include/asm/mmu_context.h +++ b/arch/riscv/include/asm/mmu_context.h @@ -48,6 +48,13 @@ static inline unsigned long mm_untag_mask(struct mm_struct *mm) } #endif +#define deactivate_mm deactivate_mm +static inline void deactivate_mm(struct task_struct *tsk, + struct mm_struct *mm) +{ + shstk_release(tsk); +} + #include #endif /* _ASM_RISCV_MMU_CONTEXT_H */ diff --git a/arch/riscv/include/asm/usercfi.h b/arch/riscv/include/asm/usercfi.h index 5f2027c51917..82d28ac98d76 100644 --- a/arch/riscv/include/asm/usercfi.h +++ b/arch/riscv/include/asm/usercfi.h @@ -8,6 +8,9 @@ #ifndef __ASSEMBLY__ #include +struct task_struct; +struct kernel_clone_args; + #ifdef CONFIG_RISCV_USER_CFI struct cfi_status { unsigned long ubcfi_en : 1; /* Enable for backward cfi. */ @@ -17,6 +20,28 @@ struct cfi_status { unsigned long shdw_stk_size; /* size of shadow stack */ }; +unsigned long shstk_alloc_thread_stack(struct task_struct *tsk, + const struct kernel_clone_args *args); +void shstk_release(struct task_struct *tsk); +void set_shstk_base(struct task_struct *task, unsigned long shstk_addr, unsigned long size); +unsigned long get_shstk_base(struct task_struct *task, unsigned long *size); +void set_active_shstk(struct task_struct *task, unsigned long shstk_addr); +bool is_shstk_enabled(struct task_struct *task); + +#else + +#define shstk_alloc_thread_stack(tsk, args) 0 + +#define shstk_release(tsk) + +#define get_shstk_base(task, size) 0UL + +#define set_shstk_base(task, shstk_addr, size) + +#define set_active_shstk(task, shstk_addr) + +#define is_shstk_enabled(task) false + #endif /* CONFIG_RISCV_USER_CFI */ #endif /* __ASSEMBLY__ */ diff --git a/arch/riscv/kernel/process.c b/arch/riscv/kernel/process.c index 7c244de77180..99acb6342a37 100644 --- a/arch/riscv/kernel/process.c +++ b/arch/riscv/kernel/process.c @@ -29,6 +29,7 @@ #include #include #include +#include #if defined(CONFIG_STACKPROTECTOR) && !defined(CONFIG_STACKPROTECTOR_PER_TASK) #include @@ -211,6 +212,7 @@ int copy_thread(struct task_struct *p, const struct kernel_clone_args *args) unsigned long clone_flags = args->flags; unsigned long usp = args->stack; unsigned long tls = args->tls; + unsigned long ssp = 0; struct pt_regs *childregs = task_pt_regs(p); /* Ensure all threads in this mm have the same pointer masking mode. */ @@ -229,11 +231,18 @@ int copy_thread(struct task_struct *p, const struct kernel_clone_args *args) p->thread.s[0] = (unsigned long)args->fn; p->thread.s[1] = (unsigned long)args->fn_arg; } else { + /* allocate new shadow stack if needed. In case of CLONE_VM we have to */ + ssp = shstk_alloc_thread_stack(p, args); + if (IS_ERR_VALUE(ssp)) + return PTR_ERR((void *)ssp); + *childregs = *(current_pt_regs()); /* Turn off status.VS */ riscv_v_vstate_off(childregs); if (usp) /* User fork */ childregs->sp = usp; + /* if needed, set new ssp */ + ssp ? set_active_shstk(p, ssp) : 0; if (clone_flags & CLONE_SETTLS) childregs->tp = tls; childregs->a0 = 0; /* Return value of fork() */ diff --git a/arch/riscv/kernel/usercfi.c b/arch/riscv/kernel/usercfi.c index 24022809a7b5..73cf87dab186 100644 --- a/arch/riscv/kernel/usercfi.c +++ b/arch/riscv/kernel/usercfi.c @@ -19,6 +19,41 @@ #define SHSTK_ENTRY_SIZE sizeof(void *) +bool is_shstk_enabled(struct task_struct *task) +{ + return task->thread_info.user_cfi_state.ubcfi_en ? true : false; +} + +void set_shstk_base(struct task_struct *task, unsigned long shstk_addr, unsigned long size) +{ + task->thread_info.user_cfi_state.shdw_stk_base = shstk_addr; + task->thread_info.user_cfi_state.shdw_stk_size = size; +} + +unsigned long get_shstk_base(struct task_struct *task, unsigned long *size) +{ + if (size) + *size = task->thread_info.user_cfi_state.shdw_stk_size; + return task->thread_info.user_cfi_state.shdw_stk_base; +} + +void set_active_shstk(struct task_struct *task, unsigned long shstk_addr) +{ + task->thread_info.user_cfi_state.user_shdw_stk = shstk_addr; +} + +/* + * If size is 0, then to be compatible with regular stack we want it to be as big as + * regular stack. Else PAGE_ALIGN it and return back + */ +static unsigned long calc_shstk_size(unsigned long size) +{ + if (size) + return PAGE_ALIGN(size); + + return PAGE_ALIGN(min_t(unsigned long long, rlimit(RLIMIT_STACK), SZ_4G)); +} + /* * Writes on shadow stack can either be `sspush` or `ssamoswap`. `sspush` can happen * implicitly on current shadow stack pointed to by CSR_SSP. `ssamoswap` takes pointer to @@ -142,3 +177,88 @@ SYSCALL_DEFINE3(map_shadow_stack, unsigned long, addr, unsigned long, size, unsi return allocate_shadow_stack(addr, aligned_size, size, set_tok); } + +/* + * This gets called during clone/clone3/fork. And is needed to allocate a shadow stack for + * cases where CLONE_VM is specified and thus a different stack is specified by user. We + * thus need a separate shadow stack too. How does separate shadow stack is specified by + * user is still being debated. Once that's settled, remove this part of the comment. + * This function simply returns 0 if shadow stack are not supported or if separate shadow + * stack allocation is not needed (like in case of !CLONE_VM) + */ +unsigned long shstk_alloc_thread_stack(struct task_struct *tsk, + const struct kernel_clone_args *args) +{ + unsigned long addr, size; + + /* If shadow stack is not supported, return 0 */ + if (!cpu_supports_shadow_stack()) + return 0; + + /* + * If shadow stack is not enabled on the new thread, skip any + * switch to a new shadow stack. + */ + if (!is_shstk_enabled(tsk)) + return 0; + + /* + * For CLONE_VFORK the child will share the parents shadow stack. + * Set base = 0 and size = 0, this is special means to track this state + * so the freeing logic run for child knows to leave it alone. + */ + if (args->flags & CLONE_VFORK) { + set_shstk_base(tsk, 0, 0); + return 0; + } + + /* + * For !CLONE_VM the child will use a copy of the parents shadow + * stack. + */ + if (!(args->flags & CLONE_VM)) + return 0; + + /* + * reaching here means, CLONE_VM was specified and thus a separate shadow + * stack is needed for new cloned thread. Note: below allocation is happening + * using current mm. + */ + size = calc_shstk_size(args->stack_size); + addr = allocate_shadow_stack(0, size, 0, false); + if (IS_ERR_VALUE(addr)) + return addr; + + set_shstk_base(tsk, addr, size); + + return addr + size; +} + +void shstk_release(struct task_struct *tsk) +{ + unsigned long base = 0, size = 0; + /* If shadow stack is not supported or not enabled, nothing to release */ + if (!cpu_supports_shadow_stack() || !is_shstk_enabled(tsk)) + return; + + /* + * When fork() with CLONE_VM fails, the child (tsk) already has a + * shadow stack allocated, and exit_thread() calls this function to + * free it. In this case the parent (current) and the child share + * the same mm struct. Move forward only when they're same. + */ + if (!tsk->mm || tsk->mm != current->mm) + return; + + /* + * We know shadow stack is enabled but if base is NULL, then + * this task is not managing its own shadow stack (CLONE_VFORK). So + * skip freeing it. + */ + base = get_shstk_base(tsk, &size); + if (!base) + return; + + vm_munmap(base, size); + set_shstk_base(tsk, 0, 0); +}