From patchwork Fri Mar 21 17:37:27 2025 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Sourav Panda X-Patchwork-Id: 14025837 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 9B93BC36000 for ; Fri, 21 Mar 2025 17:37:44 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 93956280006; Fri, 21 Mar 2025 13:37:42 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 8BE9B280001; Fri, 21 Mar 2025 13:37:42 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 67943280006; Fri, 21 Mar 2025 13:37:42 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0016.hostedemail.com [216.40.44.16]) by kanga.kvack.org (Postfix) with ESMTP id 38F8D280001 for ; Fri, 21 Mar 2025 13:37:42 -0400 (EDT) Received: from smtpin01.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay06.hostedemail.com (Postfix) with ESMTP id 667ADB7917 for ; Fri, 21 Mar 2025 17:37:43 +0000 (UTC) X-FDA: 83246265606.01.987D2B9 Received: from mail-pj1-f73.google.com (mail-pj1-f73.google.com [209.85.216.73]) by imf19.hostedemail.com (Postfix) with ESMTP id 853421A000F for ; Fri, 21 Mar 2025 17:37:41 +0000 (UTC) Authentication-Results: imf19.hostedemail.com; dkim=pass header.d=google.com header.s=20230601 header.b=YbB2Ilud; spf=pass (imf19.hostedemail.com: domain of 35KPdZwsKCOcbXdaJeYJWMJPXXPUN.LXVURWdg-VVTeJLT.XaP@flex--souravpanda.bounces.google.com designates 209.85.216.73 as permitted sender) smtp.mailfrom=35KPdZwsKCOcbXdaJeYJWMJPXXPUN.LXVURWdg-VVTeJLT.XaP@flex--souravpanda.bounces.google.com; dmarc=pass (policy=reject) header.from=google.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1742578661; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:mime-version:mime-version: content-type:content-type:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=w+86G2SUYfeiHLweFc9OnEaVmlu/kSLmPHY1T1dM0HE=; b=T66OfsK8HwsbBMlA7aHPezaVNDeTrjj8sG3YqPu5S6Pi72wsNZuDZVwjkEfG4OXpJnH7ky gnHJRAxrDFBljDw4KoaqqBCEPO6QxzgZesjVljRUuLwqeEDZnMZdr+83GHqbAt6VC4mMlN juCJM8DcIFmLwKDc829t6JBHaBbcVsY= ARC-Authentication-Results: i=1; imf19.hostedemail.com; dkim=pass header.d=google.com header.s=20230601 header.b=YbB2Ilud; spf=pass (imf19.hostedemail.com: domain of 35KPdZwsKCOcbXdaJeYJWMJPXXPUN.LXVURWdg-VVTeJLT.XaP@flex--souravpanda.bounces.google.com designates 209.85.216.73 as permitted sender) smtp.mailfrom=35KPdZwsKCOcbXdaJeYJWMJPXXPUN.LXVURWdg-VVTeJLT.XaP@flex--souravpanda.bounces.google.com; dmarc=pass (policy=reject) header.from=google.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1742578661; a=rsa-sha256; cv=none; b=FqX8ZrF9sdXMnAQgP1fhlU9ezCUO0zMsRGfeFU0OPGIIjLX6XnfhTjJUd8LwYSxZa6pPZm VXsF7KYnm9dBxYtvzCcfuHSc5THL0R7Gv0WA7E2tD8wQ4KA8QdghJQLaHlzzrFSdbsB6LO 3lmPFLA6+bcRpRxjj8PlzcnOlavoUgM= Received: by mail-pj1-f73.google.com with SMTP id 98e67ed59e1d1-3011bee1751so3467473a91.1 for ; Fri, 21 Mar 2025 10:37:41 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20230601; t=1742578660; x=1743183460; darn=kvack.org; h=to:from:subject:message-id:references:mime-version:in-reply-to:date :from:to:cc:subject:date:message-id:reply-to; bh=w+86G2SUYfeiHLweFc9OnEaVmlu/kSLmPHY1T1dM0HE=; b=YbB2IludbCEl7WBx4lRs7/BRDVNCHBQZCMuw83y63LElBQDhmNmZf2R9tDfAt4eyNP xGX+4F6iSVszRlBRmUSMDIRTRnYLCQjK40vwZ+3oZWeqiuMK5e1H2Kf+xfkmUgEVgaeI 4tJVJ/NDtrBChMgRQQ9bYUFZntLiAPrmKKKcUKYVypFjlQWKmnGs14UNwJ2Evb7258BL zCP2+hMmH7lo/kg77QJ5W1NKR0Ar00xktWsm+20e0Ox3dJOBmvQq0zhgbiv+Bza7W42V yjEw36/A8cKqLfj96qcUsaLUzruL0pNKD1KxNQHuUN9QYejf/oFC1dsR6MD1qywGE4Nk hHNQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1742578660; x=1743183460; h=to:from:subject:message-id:references:mime-version:in-reply-to:date :x-gm-message-state:from:to:cc:subject:date:message-id:reply-to; bh=w+86G2SUYfeiHLweFc9OnEaVmlu/kSLmPHY1T1dM0HE=; b=M/Cy6TxraGC0YI5mTs6WN2ytX/+je6JRbA6ff5wYgfcuADvQehgFirH/AAS07cRr51 5yUeKSOu8zltOEMPNdYzsvz1JrDl7Ycvrm2xclMzpmxhfHRBwBH6iXwjlyfAMrvLl0+1 Umf5TxpJy5Yevj6xpUizgp7MXegZfdt6/HJGyzTETd17S6ZGS0/sIxh4QR5N6N5+He8h v2/0uzrRf72DD5L6XJi3HQpl+ucYGGmVYdZOW7WZVFjx/UtuKQQNrh9SjRQSZLCybYsy yeLd+kn1CUVDvMwzO38sflmZICIylH6EmPkipCtDb64ZTt3W2ypBnW6sLoapXwmbWbXT qGjw== X-Forwarded-Encrypted: i=1; AJvYcCVTwnafgrRvWMWeyBAY7uEkNveVPw3BhuKGjuVy9nt/3e3Hi/61b/xwNNcwj6ovzNGrwCS8POX7ZA==@kvack.org X-Gm-Message-State: AOJu0YxaYFQDKuEc2fnHiw8lrAOgRF4/A5y9P+WUezXjab6pxhfTLR8r uSY/Vd73AxAcqyQoSDXLlDqVw+otgQj1kl8fDN9C62ay+FUGJ2/efmWJKZIIsdzYqj+e66clxp4 8CaP/qCmIdU2fyIwepHOrFw== X-Google-Smtp-Source: AGHT+IGm4f7njleqOWwtfaudXh+F/G/hZodt/w5YdBSQxu8dXfprXgYhidXkVGY03bHplMdrG0x4phFmYcftkI/tzg== X-Received: from pjtd12.prod.google.com ([2002:a17:90b:4c:b0:2fa:27e2:a64d]) (user=souravpanda job=prod-delivery.src-stubby-dispatcher) by 2002:a17:90b:5450:b0:2ff:5714:6a with SMTP id 98e67ed59e1d1-3030fe98134mr5881396a91.19.1742578660362; Fri, 21 Mar 2025 10:37:40 -0700 (PDT) Date: Fri, 21 Mar 2025 17:37:27 +0000 In-Reply-To: <20250321173729.3175898-1-souravpanda@google.com> Mime-Version: 1.0 References: <20250321173729.3175898-1-souravpanda@google.com> X-Mailer: git-send-email 2.49.0.395.g12beb8f557-goog Message-ID: <20250321173729.3175898-5-souravpanda@google.com> Subject: [RFC PATCH 4/6] mm: create dedicated trees for SELECTIVE KSM partitions From: Sourav Panda To: mathieu.desnoyers@efficios.com, willy@infradead.org, david@redhat.com, pasha.tatashin@soleen.com, rientjes@google.com, akpm@linux-foundation.org, linux-mm@kvack.org, linux-kernel@vger.kernel.org, weixugc@google.com, gthelen@google.com, souravpanda@google.com, surenb@google.com X-Rspamd-Queue-Id: 853421A000F X-Rspamd-Server: rspam05 X-Rspam-User: X-Stat-Signature: 7sbn4bp84o8rqhey1uaffw9fy8dqc1aj X-HE-Tag: 1742578661-174765 X-HE-Meta: U2FsdGVkX1/pPn3pltlAtvyFXXDk3Y6TGAty3Z765LhnUuUfS6J60S6CRj2v7jwqrPHWUarree/9LsrWjO4TmjK27AqAMOZ6ZWsYrC7e0w0u9GiTSRvaVOEBbLPNDiNjyf3Gc/OMFcf379US+t3pYQwQlvmXszQMprmkiyjpnErndboXuwm2lex7HNnex0ZKm/+WTqRvCK0tjEBVHu3RNNBffSk/pY8gs9cD/kZtAxD8C0tXsf1G2ooVNh5kVZy20PTvyscVHpSMDcaTCBhxMoRICBgRlprG0YgMzxGpsTA6lEukd34keoHC6OItgYATLDvlDv1oDSiG6ovkO4MOiVdqAsATJ78fsTh7hybcV61jIhm7Ui4ddGvOOXCCBT9f1VFQLF78JGB1elzOc9UShZpGMoP3pkkXQRm1ZUA1dOLeCvFmzrqoJIuqXSHuws0aqTnqO1jJ1uBBNFxWPIazSheXecrjvXvWX3MqSIujjzoH6sVY4ISA5K+saAnr2w/EO5nqN09k3mT+2QUYg/jN0vt8qorD85t9ZR8jb1mw5TnBWW3buhvSAaK/ZMvtaCoXCsm8HLU/bWaEuJZT7oZmjC1CC0/UOy5y/BTsnebgsbjyf+PrALEfG0keoFsLCy5sI66YM8VfdAZptNmVZdC8xI1Z6xBEEx9GE+nE5fms3Vnj53XLTnEySvNICnK0yJrwCapI1i3Q+FrkXJzzl82SIV05/1q57Y18iYDz36dT5vrL2Q+PTNYypt/zUinIWz8HdYJDBD+xNSsa2vhzqDGq2yJewtAa2j5mu3jo1naFV1ruzkzWbQOTvMO/PYHmtBdV6f3/4Q9ch//Hd4dieZn/BL7SLZYuDjLjPUgCzs+Jbudtp9EVjiz1w2lqahpusfb5Q+oALWdnnEczG0CVLwViLDqq67PunH8h+6uB7PbEibGXCFMd12anpVb6P06slb5NJJVYZFZXd+pnHeJa3ux xUxNB5zr DaFEoHgnrmubj+ZSwHD0cL22TrNPGCpos2TELjqN3trbEX/UdUoaw6rMc/t4WMpbenXcveJTxavATBYI8bwmMQdoQN1aEo+Qz/15FO9PhPDIFWYFfcZn1qj8Ae5H/FRRqRwLOTLEzsh4MK2NTg+VE/owjAL82ITkSbXgPAy8W9OveHz6wV09eM7X0VP9qQLxPNV/Y5PEWu2Qm2Wnj1xJhFnBbsJ9g93ajo7DLV83MhOXc0A5P5O/3e/phNVDyrPq7bFocK9xEEhZ7DAu9LThw16YL2EVpS+/s2zjKqOGRnv/rFBl9ChfefQQfe1AQrsT76jO595zazUr48g0JlJEFEsnZUAkAHLk5SIscoIGLE6G1fD3qkajE2EoBWk6FAR0Hstox1Ale7uK7sOiZBaUyNMGX8NQaCpohlfHzNv+0MDYUwmUXZAtIR4yVlqCTSbihtS1CHx3FwHlV8Z3qHkYMpqK2R7LpPo9pVRs+H11OzrVqkMUOPFsZf62GTLG8YDLKwgYBuZ6qB5rlhpbztsnX4trg03cdTpe4ITPf0PinxKmw7cMml9PJLWAACVcT6Pj+ggcfXOjh+5K5YgA= X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: Extend ksm to create dedicated unstable and stable trees for each partition. Signed-off-by: Sourav Panda --- mm/ksm.c | 165 +++++++++++++++++++++++++++++++++++++------------------ 1 file changed, 111 insertions(+), 54 deletions(-) diff --git a/mm/ksm.c b/mm/ksm.c index 927e257c48b5..b575250aaf45 100644 --- a/mm/ksm.c +++ b/mm/ksm.c @@ -144,6 +144,28 @@ struct ksm_scan { unsigned long seqnr; }; +static struct kobject *ksm_base_kobj; + +struct partition_kobj { + struct kobject *kobj; + struct list_head list; + struct rb_root *root_stable_tree; + struct rb_root *root_unstable_tree; +}; + +static LIST_HEAD(partition_list); + +static struct partition_kobj *find_partition_by_kobj(struct kobject *kobj) +{ + struct partition_kobj *partition; + + list_for_each_entry(partition, &partition_list, list) { + if (partition->kobj == kobj) + return partition; + } + return NULL; +} + /** * struct ksm_stable_node - node of the stable rbtree * @node: rb node of this ksm page in the stable tree @@ -182,6 +204,7 @@ struct ksm_stable_node { #ifdef CONFIG_NUMA int nid; #endif + struct partition_kobj *partition; }; /** @@ -218,6 +241,7 @@ struct ksm_rmap_item { struct hlist_node hlist; }; }; + struct partition_kobj *partition; }; #define SEQNR_MASK 0x0ff /* low bits of unstable tree seqnr */ @@ -227,8 +251,6 @@ struct ksm_rmap_item { /* The stable and unstable tree heads */ static struct rb_root one_stable_tree[1] = { RB_ROOT }; static struct rb_root one_unstable_tree[1] = { RB_ROOT }; -static struct rb_root *root_stable_tree = one_stable_tree; -static struct rb_root *root_unstable_tree = one_unstable_tree; /* Recently migrated nodes of stable tree, pending proper placement */ static LIST_HEAD(migrate_nodes); @@ -555,7 +577,7 @@ static inline void stable_node_dup_del(struct ksm_stable_node *dup) if (is_stable_node_dup(dup)) __stable_node_dup_del(dup); else - rb_erase(&dup->node, root_stable_tree + NUMA(dup->nid)); + rb_erase(&dup->node, dup->partition->root_stable_tree + NUMA(dup->nid)); #ifdef CONFIG_DEBUG_VM dup->head = NULL; #endif @@ -580,14 +602,20 @@ static inline void free_rmap_item(struct ksm_rmap_item *rmap_item) kmem_cache_free(rmap_item_cache, rmap_item); } -static inline struct ksm_stable_node *alloc_stable_node(void) +static inline struct ksm_stable_node *alloc_stable_node(struct partition_kobj *partition) { /* * The allocation can take too long with GFP_KERNEL when memory is under * pressure, which may lead to hung task warnings. Adding __GFP_HIGH * grants access to memory reserves, helping to avoid this problem. */ - return kmem_cache_alloc(stable_node_cache, GFP_KERNEL | __GFP_HIGH); + struct ksm_stable_node *node = kmem_cache_alloc(stable_node_cache, + GFP_KERNEL | __GFP_HIGH); + + if (node) + node->partition = partition; + + return node; } static inline void free_stable_node(struct ksm_stable_node *stable_node) @@ -777,9 +805,10 @@ static inline int get_kpfn_nid(unsigned long kpfn) } static struct ksm_stable_node *alloc_stable_node_chain(struct ksm_stable_node *dup, - struct rb_root *root) + struct rb_root *root, + struct partition_kobj *partition) { - struct ksm_stable_node *chain = alloc_stable_node(); + struct ksm_stable_node *chain = alloc_stable_node(partition); VM_BUG_ON(is_stable_node_chain(dup)); if (likely(chain)) { INIT_HLIST_HEAD(&chain->hlist); @@ -1016,7 +1045,8 @@ static void remove_rmap_item_from_tree(struct ksm_rmap_item *rmap_item) unsigned char age = get_rmap_item_age(rmap_item); if (!age) rb_erase(&rmap_item->node, - root_unstable_tree + NUMA(rmap_item->nid)); + rmap_item->partition->root_unstable_tree + + NUMA(rmap_item->nid)); ksm_pages_unshared--; rmap_item->address &= PAGE_MASK; } @@ -1154,17 +1184,23 @@ static int remove_all_stable_nodes(void) struct ksm_stable_node *stable_node, *next; int nid; int err = 0; - - for (nid = 0; nid < ksm_nr_node_ids; nid++) { - while (root_stable_tree[nid].rb_node) { - stable_node = rb_entry(root_stable_tree[nid].rb_node, - struct ksm_stable_node, node); - if (remove_stable_node_chain(stable_node, - root_stable_tree + nid)) { - err = -EBUSY; - break; /* proceed to next nid */ + struct partition_kobj *partition; + struct rb_root *root_stable_tree; + + list_for_each_entry(partition, &partition_list, list) { + root_stable_tree = partition->root_stable_tree; + + for (nid = 0; nid < ksm_nr_node_ids; nid++) { + while (root_stable_tree[nid].rb_node) { + stable_node = rb_entry(root_stable_tree[nid].rb_node, + struct ksm_stable_node, node); + if (remove_stable_node_chain(stable_node, + root_stable_tree + nid)) { + err = -EBUSY; + break; /* proceed to next nid */ + } + cond_resched(); } - cond_resched(); } } list_for_each_entry_safe(stable_node, next, &migrate_nodes, list) { @@ -1802,7 +1838,8 @@ static __always_inline struct folio *chain(struct ksm_stable_node **s_n_d, * This function returns the stable tree node of identical content if found, * -EBUSY if the stable node's page is being migrated, NULL otherwise. */ -static struct folio *stable_tree_search(struct page *page) +static struct folio *stable_tree_search(struct page *page, + struct partition_kobj *partition) { int nid; struct rb_root *root; @@ -1821,7 +1858,7 @@ static struct folio *stable_tree_search(struct page *page) } nid = get_kpfn_nid(folio_pfn(folio)); - root = root_stable_tree + nid; + root = partition->root_stable_tree + nid; again: new = &root->rb_node; parent = NULL; @@ -1991,7 +2028,7 @@ static struct folio *stable_tree_search(struct page *page) VM_BUG_ON(is_stable_node_dup(stable_node_dup)); /* chain is missing so create it */ stable_node = alloc_stable_node_chain(stable_node_dup, - root); + root, partition); if (!stable_node) return NULL; } @@ -2016,7 +2053,8 @@ static struct folio *stable_tree_search(struct page *page) * This function returns the stable tree node just allocated on success, * NULL otherwise. */ -static struct ksm_stable_node *stable_tree_insert(struct folio *kfolio) +static struct ksm_stable_node *stable_tree_insert(struct folio *kfolio, + struct partition_kobj *partition) { int nid; unsigned long kpfn; @@ -2028,7 +2066,7 @@ static struct ksm_stable_node *stable_tree_insert(struct folio *kfolio) kpfn = folio_pfn(kfolio); nid = get_kpfn_nid(kpfn); - root = root_stable_tree + nid; + root = partition->root_stable_tree + nid; again: parent = NULL; new = &root->rb_node; @@ -2067,7 +2105,7 @@ static struct ksm_stable_node *stable_tree_insert(struct folio *kfolio) } } - stable_node_dup = alloc_stable_node(); + stable_node_dup = alloc_stable_node(partition); if (!stable_node_dup) return NULL; @@ -2082,7 +2120,8 @@ static struct ksm_stable_node *stable_tree_insert(struct folio *kfolio) if (!is_stable_node_chain(stable_node)) { struct ksm_stable_node *orig = stable_node; /* chain is missing so create it */ - stable_node = alloc_stable_node_chain(orig, root); + stable_node = alloc_stable_node_chain(orig, root, + partition); if (!stable_node) { free_stable_node(stable_node_dup); return NULL; @@ -2121,7 +2160,7 @@ struct ksm_rmap_item *unstable_tree_search_insert(struct ksm_rmap_item *rmap_ite int nid; nid = get_kpfn_nid(page_to_pfn(page)); - root = root_unstable_tree + nid; + root = rmap_item->partition->root_unstable_tree + nid; new = &root->rb_node; while (*new) { @@ -2291,7 +2330,7 @@ static void cmp_and_merge_page(struct page *page, struct ksm_rmap_item *rmap_ite } /* Start by searching for the folio in the stable tree */ - kfolio = stable_tree_search(page); + kfolio = stable_tree_search(page, rmap_item->partition); if (&kfolio->page == page && rmap_item->head == stable_node) { folio_put(kfolio); return; @@ -2344,7 +2383,8 @@ static void cmp_and_merge_page(struct page *page, struct ksm_rmap_item *rmap_ite * node in the stable tree and add both rmap_items. */ folio_lock(kfolio); - stable_node = stable_tree_insert(kfolio); + stable_node = stable_tree_insert(kfolio, + rmap_item->partition); if (stable_node) { stable_tree_append(tree_rmap_item, stable_node, false); @@ -2502,7 +2542,8 @@ static struct ksm_rmap_item *retrieve_rmap_item(struct page **page, } static void ksm_sync_merge(struct mm_struct *mm, - unsigned long start, unsigned long end) + unsigned long start, unsigned long end, + struct partition_kobj *partition) { struct ksm_rmap_item *rmap_item; struct page *page; @@ -2510,6 +2551,7 @@ static void ksm_sync_merge(struct mm_struct *mm, rmap_item = retrieve_rmap_item(&page, mm, start, end); if (!rmap_item) return; + rmap_item->partition = partition; cmp_and_merge_page(page, rmap_item); put_page(page); } @@ -3328,19 +3370,23 @@ static void ksm_check_stable_tree(unsigned long start_pfn, struct ksm_stable_node *stable_node, *next; struct rb_node *node; int nid; - - for (nid = 0; nid < ksm_nr_node_ids; nid++) { - node = rb_first(root_stable_tree + nid); - while (node) { - stable_node = rb_entry(node, struct ksm_stable_node, node); - if (stable_node_chain_remove_range(stable_node, - start_pfn, end_pfn, - root_stable_tree + - nid)) - node = rb_first(root_stable_tree + nid); - else - node = rb_next(node); - cond_resched(); + struct rb_root *root_stable_tree + + list_for_each_entry(partition, &partition_list, list) { + root_stable_tree = partition->root_stable_tree; + + for (nid = 0; nid < ksm_nr_node_ids; nid++) { + node = rb_first(root_stable_tree + nid); + while (node) { + stable_node = rb_entry(node, struct ksm_stable_node, node); + if (stable_node_chain_remove_range(stable_node, + start_pfn, end_pfn, + root_stable_tree + nid)) + node = rb_first(root_stable_tree + nid); + else + node = rb_next(node); + cond_resched(); + } } } list_for_each_entry_safe(stable_node, next, &migrate_nodes, list) { @@ -3551,6 +3597,7 @@ static ssize_t trigger_merge_store(struct kobject *kobj, int ret; struct task_struct *task; struct mm_struct *mm; + struct partition_kobj *partition; input = kstrdup(buf, GFP_KERNEL); if (!input) @@ -3583,9 +3630,13 @@ static ssize_t trigger_merge_store(struct kobject *kobj, if (!mm) return -EINVAL; + partition = find_partition_by_kobj(kobj); + if (!partition) + return -EINVAL; + mutex_lock(&ksm_thread_mutex); wait_while_offlining(); - ksm_sync_merge(mm, start, end); + ksm_sync_merge(mm, start, end, partition); mutex_unlock(&ksm_thread_mutex); mmput(mm); @@ -3606,6 +3657,8 @@ static ssize_t merge_across_nodes_store(struct kobject *kobj, { int err; unsigned long knob; + struct rb_root *root_stable_tree; + struct partition_kobj *partition; err = kstrtoul(buf, 10, &knob); if (err) @@ -3615,6 +3668,10 @@ static ssize_t merge_across_nodes_store(struct kobject *kobj, mutex_lock(&ksm_thread_mutex); wait_while_offlining(); + + partition = find_partition_by_kobj(kobj); + root_stable_tree = partition->root_stable_tree; + if (ksm_merge_across_nodes != knob) { if (ksm_pages_shared || remove_all_stable_nodes()) err = -EBUSY; @@ -3633,10 +3690,10 @@ static ssize_t merge_across_nodes_store(struct kobject *kobj, if (!buf) err = -ENOMEM; else { - root_stable_tree = buf; - root_unstable_tree = buf + nr_node_ids; + partition->root_stable_tree = buf; + partition->root_unstable_tree = buf + nr_node_ids; /* Stable tree is empty but not the unstable */ - root_unstable_tree[0] = one_unstable_tree[0]; + partition->root_unstable_tree[0] = one_unstable_tree[0]; } } if (!err) { @@ -3834,14 +3891,6 @@ KSM_ATTR_RO(full_scans); #ifdef CONFIG_SELECTIVE_KSM static struct kobject *ksm_base_kobj; - -struct partition_kobj { - struct kobject *kobj; - struct list_head list; -}; - -static LIST_HEAD(partition_list); - #else /* CONFIG_SELECTIVE_KSM */ static ssize_t smart_scan_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf) @@ -4055,6 +4104,7 @@ static ssize_t add_partition_store(struct kobject *kobj, struct partition_kobj *new_partition_kobj; char partition_name[50]; int err; + struct rb_root *tree_root; mutex_lock(&ksm_thread_mutex); @@ -4081,6 +4131,13 @@ static ssize_t add_partition_store(struct kobject *kobj, goto unlock; } + tree_root = kcalloc(nr_node_ids + nr_node_ids, sizeof(*tree_root), GFP_KERNEL); + if (!tree_root) { + err = -ENOMEM; + goto unlock; + } + new_partition_kobj->root_stable_tree = tree_root; + new_partition_kobj->root_unstable_tree = tree_root + nr_node_ids; err = sysfs_create_group(new_partition_kobj->kobj, &ksm_attr_group); if (err) { pr_err("ksm: register sysfs failed\n");