From patchwork Tue Nov 6 09:30:59 2018 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Kuo-Hsin Yang X-Patchwork-Id: 10669967 Return-Path: Received: from mail.wl.linuxfoundation.org (pdx-wl-mail.web.codeaurora.org [172.30.200.125]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id 5B3F313BF for ; Tue, 6 Nov 2018 09:31:14 +0000 (UTC) Received: from mail.wl.linuxfoundation.org (localhost [127.0.0.1]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 4990C2A281 for ; Tue, 6 Nov 2018 09:31:14 +0000 (UTC) Received: by mail.wl.linuxfoundation.org (Postfix, from userid 486) id 3D2522A294; Tue, 6 Nov 2018 09:31:14 +0000 (UTC) X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on pdx-wl-mail.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-3.0 required=2.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,MAILING_LIST_MULTI,RCVD_IN_DNSWL_NONE autolearn=ham version=3.3.1 Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 670B62A281 for ; Tue, 6 Nov 2018 09:31:13 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 5879C6B02F0; Tue, 6 Nov 2018 04:31:07 -0500 (EST) Delivered-To: linux-mm-outgoing@kvack.org Received: by kanga.kvack.org (Postfix, from userid 40) id 539D06B02F2; Tue, 6 Nov 2018 04:31:07 -0500 (EST) X-Original-To: int-list-linux-mm@kvack.org X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 427916B02F3; Tue, 6 Nov 2018 04:31:07 -0500 (EST) X-Original-To: linux-mm@kvack.org X-Delivered-To: linux-mm@kvack.org Received: from mail-pl1-f197.google.com (mail-pl1-f197.google.com [209.85.214.197]) by kanga.kvack.org (Postfix) with ESMTP id F398A6B02F0 for ; Tue, 6 Nov 2018 04:31:06 -0500 (EST) Received: by mail-pl1-f197.google.com with SMTP id e97-v6so12826221plb.10 for ; Tue, 06 Nov 2018 01:31:06 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:dkim-signature:from:to:cc:subject:date :message-id:mime-version:content-transfer-encoding; bh=03j09s75VsTqkZOJHb0MeFK1toWMKtdV6C+k4g28cQY=; b=Cxnku8k/L8ltwqYZH9UM3glF2gPpZqV4Q0HTvn+a6QO0pyKtSQKvbn09CMH8dunrPl O84ia+NoIcyeI/uj+2WDxbKIxucRe0zu54SXjpQhpgwwztDSw9fGNMIETLiegZZHI6H4 jwvCKUoTDKYxHTNWfUMRPi/auRqZEEd3Dx4RlQZfh7yjlYAX9SvN3TTYp0nwKXjY7bat VMuYwx8i58pLYwz8s1VYiDPdnxx5TkRFC9t4T0vzobJZO0Hok4gnVirIPslhF8ztZCij cIVPzswBERxYNbptkS4G+0mwUOBRA6Kh7DFxm2FY77NLd68gK3LFB90fe02PvsSqTiNh 0HJw== X-Gm-Message-State: AGRZ1gKkrMVnv/bjwV3w9HPsBX7VQZfELGxpm19yLsQpYF35xJoQorrT wTWdTIdVeChewmpvBxHifIjDl13/aVz9s6KXeByJIuZYIEDzJrIOhkBHU1tq/X8CaFpkdK9fqRo 2eQ1E4M6dKeOz+eNcZEQpX4Q3k4KYUfuX/AyfvEnyZgE3pzAjDrx/k5eAKGVgOgk2vKxbXNwsEG NHAADYVSKOz/UiIUm8zpIpvBZT42QrZ/zgb6OVZbwGuho7ClVu7sMGypoWztmm/KWWTHZQkfk0K /kCifI9dg6xh97/oTLp5a32biNEhP5H2zdyw+KyLvzUCmUiN2usY0F7p4ZzT6S40VayjxY0jfjj t5EecVoS6hMY7WFLJ3Pw9WkS/OgA+mzuNrQZewZCvhir2flljnYr+PYeMLM7TTgOhbIF0nOLqtD M X-Received: by 2002:a63:d52:: with SMTP id 18-v6mr23296174pgn.107.1541496666600; Tue, 06 Nov 2018 01:31:06 -0800 (PST) X-Received: by 2002:a63:d52:: with SMTP id 18-v6mr23296112pgn.107.1541496665587; Tue, 06 Nov 2018 01:31:05 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1541496665; cv=none; d=google.com; s=arc-20160816; b=weq6gzTRlG+l3TLE0wvZGAm1xz9u65C/s+W1RoQUaEzvXGKEu4RlbCJ9TY8e+EnDh0 I60ftnJkpLgF/NARwUIqk/+xt83fpc7RPXDdxt25M6jQfU0dKElTTNswJmREuMx4mJ2q rihBCBQEjZHWI+rz0qjIscPywD4rpWiRopunXfEG3jXtXGJmb+WFOTXJgmDnLV3KGEP5 q1TV4k2P7DyqbGZlGlUcjCm6OUk4Ydr5OHPC+X/qfgcY4JfLe5A8W5fYN8z58AHPSFM/ SVDUsN+YM+rgQZyanneiz5OpOzXw6XiLigYN68hACFqC2PpI1qrnDDCGPjfJ4EKqNT0U l6kg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=content-transfer-encoding:mime-version:message-id:date:subject:cc :to:from:dkim-signature; bh=03j09s75VsTqkZOJHb0MeFK1toWMKtdV6C+k4g28cQY=; b=QnQzk9aLlMyDli6qGL3Z0H75y7wURj9g3uK1BwbX4lGeY3ME5u+FyzaS+t+NZE9mYT fraC46megvxLHGLq4Lbv1q39IDM5007VZzolXO1xSOaGrJb6vV6wiRo5dqfkjdZEPOY7 Dauu3JueXMbfBsqyAFIiY5Ndm1HNRP59GG2g3iunizmET+eA7R9I80ppIvoUR/p5MmWJ hETuOazxuU+Q0vitfPEvRHdUgB7ZFsezfa2U9iEFHBNRMruawXOrkbWeA2Z/5OYYW0ma ORNhjwdD2nl03CWbdAdLdfzpXVZEgBLFog8hsFBAHEmYbbRfugjE+w7c/Rgq5BWuK3S1 lyVg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@chromium.org header.s=google header.b="V6TVW/K4"; spf=pass (google.com: domain of vovoy@chromium.org designates 209.85.220.65 as permitted sender) smtp.mailfrom=vovoy@chromium.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=chromium.org Received: from mail-sor-f65.google.com (mail-sor-f65.google.com. [209.85.220.65]) by mx.google.com with SMTPS id e16-v6sor40523732pgh.17.2018.11.06.01.31.05 for (Google Transport Security); Tue, 06 Nov 2018 01:31:05 -0800 (PST) Received-SPF: pass (google.com: domain of vovoy@chromium.org designates 209.85.220.65 as permitted sender) client-ip=209.85.220.65; Authentication-Results: mx.google.com; dkim=pass header.i=@chromium.org header.s=google header.b="V6TVW/K4"; spf=pass (google.com: domain of vovoy@chromium.org designates 209.85.220.65 as permitted sender) smtp.mailfrom=vovoy@chromium.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=chromium.org DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=chromium.org; s=google; h=from:to:cc:subject:date:message-id:mime-version :content-transfer-encoding; bh=03j09s75VsTqkZOJHb0MeFK1toWMKtdV6C+k4g28cQY=; b=V6TVW/K4V6TiC04O5bkrxC8N4QPwOTya1DZM9vP+oSRu3TYn4aJmqA3PXnBVB++c6N JF/SCNTbV1vg7C6XvvbqVVygfuYDzqjd3DXR1Tl8xZ3Th6/rGbr3T/I0FsLP91mKwXhv CGs/i9x8fWnu2bKS0D9L6wjwllLBfTrN2M4iI= X-Google-Smtp-Source: AJdET5erv76fRKDwBMAgfSbNi4NU0/g45MgseahPRsTShELMByBhhk5hvfvuVs3GAlLXBzNTDw+MHg== X-Received: by 2002:a63:2f86:: with SMTP id v128mr22401055pgv.407.1541496665000; Tue, 06 Nov 2018 01:31:05 -0800 (PST) Received: from vovoy-z840.tpe.corp.google.com ([2401:fa00:1:b:c11e:9571:ae83:d95b]) by smtp.gmail.com with ESMTPSA id 64-v6sm55227192pgb.74.2018.11.06.01.31.02 (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Tue, 06 Nov 2018 01:31:04 -0800 (PST) From: Kuo-Hsin Yang To: linux-kernel@vger.kernel.org, intel-gfx@lists.freedesktop.org, linux-mm@kvack.org Cc: Kuo-Hsin Yang , Chris Wilson , Joonas Lahtinen , Peter Zijlstra , Andrew Morton , Dave Hansen , Michal Hocko Subject: [PATCH v6] mm, drm/i915: mark pinned shmemfs pages as unevictable Date: Tue, 6 Nov 2018 17:30:59 +0800 Message-Id: <20181106093100.71829-1-vovoy@chromium.org> X-Mailer: git-send-email 2.19.1.930.g4563a0d9d0-goog MIME-Version: 1.0 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: X-Virus-Scanned: ClamAV using ClamSMTP The i915 driver uses shmemfs to allocate backing storage for gem objects. These shmemfs pages can be pinned (increased ref count) by shmem_read_mapping_page_gfp(). When a lot of pages are pinned, vmscan wastes a lot of time scanning these pinned pages. In some extreme case, all pages in the inactive anon lru are pinned, and only the inactive anon lru is scanned due to inactive_ratio, the system cannot swap and invokes the oom-killer. Mark these pinned pages as unevictable to speed up vmscan. Export pagevec API check_move_unevictable_pages(). This patch was inspired by Chris Wilson's change [1]. [1]: https://patchwork.kernel.org/patch/9768741/ Cc: Chris Wilson Cc: Joonas Lahtinen Cc: Peter Zijlstra Cc: Andrew Morton Cc: Dave Hansen Signed-off-by: Kuo-Hsin Yang Acked-by: Michal Hocko # mm part --- Changes for v6: Tweak the acked-by. Changes for v5: Modify doc and comments. Remove the ifdef surrounding check_move_unevictable_pages. Changes for v4: Export pagevec API check_move_unevictable_pages(). Changes for v3: Use check_move_lru_page instead of shmem_unlock_mapping to move pages to appropriate lru lists. Changes for v2: Squashed the two patches. Documentation/vm/unevictable-lru.rst | 6 +++++- drivers/gpu/drm/i915/i915_gem.c | 28 ++++++++++++++++++++++++++-- include/linux/swap.h | 4 +++- mm/shmem.c | 2 +- mm/vmscan.c | 22 +++++++++++----------- 5 files changed, 46 insertions(+), 16 deletions(-) diff --git a/Documentation/vm/unevictable-lru.rst b/Documentation/vm/unevictable-lru.rst index fdd84cb8d511..b8e29f977f2d 100644 --- a/Documentation/vm/unevictable-lru.rst +++ b/Documentation/vm/unevictable-lru.rst @@ -143,7 +143,7 @@ using a number of wrapper functions: Query the address space, and return true if it is completely unevictable. -These are currently used in two places in the kernel: +These are currently used in three places in the kernel: (1) By ramfs to mark the address spaces of its inodes when they are created, and this mark remains for the life of the inode. @@ -154,6 +154,10 @@ These are currently used in two places in the kernel: swapped out; the application must touch the pages manually if it wants to ensure they're in memory. + (3) By the i915 driver to mark pinned address space until it's unpinned. The + amount of unevictable memory marked by i915 driver is roughly the bounded + object size in debugfs/dri/0/i915_gem_objects. + Detecting Unevictable Pages --------------------------- diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index 0c8aa57ce83b..c620891e0d02 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c @@ -2381,12 +2381,25 @@ void __i915_gem_object_invalidate(struct drm_i915_gem_object *obj) invalidate_mapping_pages(mapping, 0, (loff_t)-1); } +/** + * Move pages to appropriate lru and release the pagevec. Decrement the ref + * count of these pages. + */ +static inline void check_release_pagevec(struct pagevec *pvec) +{ + if (pagevec_count(pvec)) { + check_move_unevictable_pages(pvec); + __pagevec_release(pvec); + } +} + static void i915_gem_object_put_pages_gtt(struct drm_i915_gem_object *obj, struct sg_table *pages) { struct sgt_iter sgt_iter; struct page *page; + struct pagevec pvec; __i915_gem_object_release_shmem(obj, pages, true); @@ -2395,6 +2408,9 @@ i915_gem_object_put_pages_gtt(struct drm_i915_gem_object *obj, if (i915_gem_object_needs_bit17_swizzle(obj)) i915_gem_object_save_bit_17_swizzle(obj, pages); + mapping_clear_unevictable(file_inode(obj->base.filp)->i_mapping); + + pagevec_init(&pvec); for_each_sgt_page(page, sgt_iter, pages) { if (obj->mm.dirty) set_page_dirty(page); @@ -2402,8 +2418,10 @@ i915_gem_object_put_pages_gtt(struct drm_i915_gem_object *obj, if (obj->mm.madv == I915_MADV_WILLNEED) mark_page_accessed(page); - put_page(page); + if (!pagevec_add(&pvec, page)) + check_release_pagevec(&pvec); } + check_release_pagevec(&pvec); obj->mm.dirty = false; sg_free_table(pages); @@ -2526,6 +2544,7 @@ static int i915_gem_object_get_pages_gtt(struct drm_i915_gem_object *obj) unsigned int sg_page_sizes; gfp_t noreclaim; int ret; + struct pagevec pvec; /* * Assert that the object is not currently in any GPU domain. As it @@ -2559,6 +2578,7 @@ static int i915_gem_object_get_pages_gtt(struct drm_i915_gem_object *obj) * Fail silently without starting the shrinker */ mapping = obj->base.filp->f_mapping; + mapping_set_unevictable(mapping); noreclaim = mapping_gfp_constraint(mapping, ~__GFP_RECLAIM); noreclaim |= __GFP_NORETRY | __GFP_NOWARN; @@ -2673,8 +2693,12 @@ static int i915_gem_object_get_pages_gtt(struct drm_i915_gem_object *obj) err_sg: sg_mark_end(sg); err_pages: + mapping_clear_unevictable(mapping); + pagevec_init(&pvec); for_each_sgt_page(page, sgt_iter, st) - put_page(page); + if (!pagevec_add(&pvec, page)) + check_release_pagevec(&pvec); + check_release_pagevec(&pvec); sg_free_table(st); kfree(st); diff --git a/include/linux/swap.h b/include/linux/swap.h index d8a07a4f171d..a8f6d5d89524 100644 --- a/include/linux/swap.h +++ b/include/linux/swap.h @@ -18,6 +18,8 @@ struct notifier_block; struct bio; +struct pagevec; + #define SWAP_FLAG_PREFER 0x8000 /* set if swap priority specified */ #define SWAP_FLAG_PRIO_MASK 0x7fff #define SWAP_FLAG_PRIO_SHIFT 0 @@ -369,7 +371,7 @@ static inline int node_reclaim(struct pglist_data *pgdat, gfp_t mask, #endif extern int page_evictable(struct page *page); -extern void check_move_unevictable_pages(struct page **, int nr_pages); +extern void check_move_unevictable_pages(struct pagevec *pvec); extern int kswapd_run(int nid); extern void kswapd_stop(int nid); diff --git a/mm/shmem.c b/mm/shmem.c index ea26d7a0342d..de4893c904a3 100644 --- a/mm/shmem.c +++ b/mm/shmem.c @@ -756,7 +756,7 @@ void shmem_unlock_mapping(struct address_space *mapping) break; index = indices[pvec.nr - 1] + 1; pagevec_remove_exceptionals(&pvec); - check_move_unevictable_pages(pvec.pages, pvec.nr); + check_move_unevictable_pages(&pvec); pagevec_release(&pvec); cond_resched(); } diff --git a/mm/vmscan.c b/mm/vmscan.c index 62ac0c488624..d070f431ff19 100644 --- a/mm/vmscan.c +++ b/mm/vmscan.c @@ -50,6 +50,7 @@ #include #include #include +#include #include #include @@ -4182,17 +4183,16 @@ int page_evictable(struct page *page) return ret; } -#ifdef CONFIG_SHMEM /** - * check_move_unevictable_pages - check pages for evictability and move to appropriate zone lru list - * @pages: array of pages to check - * @nr_pages: number of pages to check + * check_move_unevictable_pages - check pages for evictability and move to + * appropriate zone lru list + * @pvec: pagevec with lru pages to check * - * Checks pages for evictability and moves them to the appropriate lru list. - * - * This function is only used for SysV IPC SHM_UNLOCK. + * Checks pages for evictability, if an evictable page is in the unevictable + * lru list, moves it to the appropriate evictable lru list. This function + * should be only used for lru pages. */ -void check_move_unevictable_pages(struct page **pages, int nr_pages) +void check_move_unevictable_pages(struct pagevec *pvec) { struct lruvec *lruvec; struct pglist_data *pgdat = NULL; @@ -4200,8 +4200,8 @@ void check_move_unevictable_pages(struct page **pages, int nr_pages) int pgrescued = 0; int i; - for (i = 0; i < nr_pages; i++) { - struct page *page = pages[i]; + for (i = 0; i < pvec->nr; i++) { + struct page *page = pvec->pages[i]; struct pglist_data *pagepgdat = page_pgdat(page); pgscanned++; @@ -4233,4 +4233,4 @@ void check_move_unevictable_pages(struct page **pages, int nr_pages) spin_unlock_irq(&pgdat->lru_lock); } } -#endif /* CONFIG_SHMEM */ +EXPORT_SYMBOL_GPL(check_move_unevictable_pages);