From patchwork Fri May 17 06:06:49 2019 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-Patchwork-Submitter: "chenjianhong (A)" X-Patchwork-Id: 10947379 Return-Path: Received: from mail.wl.linuxfoundation.org (pdx-wl-mail.web.codeaurora.org [172.30.200.125]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id 5ABB8912 for ; Fri, 17 May 2019 06:02:41 +0000 (UTC) Received: from mail.wl.linuxfoundation.org (localhost [127.0.0.1]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 3CAC4201BD for ; Fri, 17 May 2019 06:02:41 +0000 (UTC) Received: by mail.wl.linuxfoundation.org (Postfix, from userid 486) id 20458201F5; Fri, 17 May 2019 06:02:41 +0000 (UTC) X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on pdx-wl-mail.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-2.9 required=2.0 tests=BAYES_00,MAILING_LIST_MULTI, RCVD_IN_DNSWL_NONE autolearn=ham version=3.3.1 Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 5840B201BD for ; Fri, 17 May 2019 06:02:40 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 91B446B0005; Fri, 17 May 2019 02:02:38 -0400 (EDT) Delivered-To: linux-mm-outgoing@kvack.org Received: by kanga.kvack.org (Postfix, from userid 40) id 8A50B6B0006; Fri, 17 May 2019 02:02:38 -0400 (EDT) X-Original-To: int-list-linux-mm@kvack.org X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 744DC6B0007; Fri, 17 May 2019 02:02:38 -0400 (EDT) X-Original-To: linux-mm@kvack.org X-Delivered-To: linux-mm@kvack.org Received: from mail-ot1-f69.google.com (mail-ot1-f69.google.com [209.85.210.69]) by kanga.kvack.org (Postfix) with ESMTP id 4297D6B0005 for ; Fri, 17 May 2019 02:02:38 -0400 (EDT) Received: by mail-ot1-f69.google.com with SMTP id o98so1634484ota.11 for ; Thu, 16 May 2019 23:02:38 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-original-authentication-results:x-gm-message-state:from:to:cc :subject:date:message-id:mime-version:content-transfer-encoding; bh=DO6WLkoUqXxxlGmLPR+WFpvuFJaMgzTrrjSYWEqE6Ok=; b=PoqOQgo+23VwURAfYf+FH7FsfH1C7QJZy0JogwwZpw+JuXGGBrNqwLmQL4unSFYMUj PkL7wFELQqZ2f9okS61UxtWEDIOKnDDPHgrw49UaFc8/Kg51Y4DXg0Qile1IJ287v++C 5bOsX3sO8uE5VZphNDygSiD96xlejIHeIhZrB3cR52jyrfLBuXRW0Nqt+JCS8B2/Tfne f8oxkmht1bqE2ad7kYzNK44sCZXipiiLbQrhQeoY4qHjRpVe0lzRNtYCx69j717f8nyx m5h2Bj1+ULPlSkQtD4XJRCfuPSFiVIL/bifIXMzDc4wQRCDBWT5g5yNVtn3asNBkQ0N/ 6I5w== X-Original-Authentication-Results: mx.google.com; spf=pass (google.com: domain of chenjianhong2@huawei.com designates 45.249.212.191 as permitted sender) smtp.mailfrom=chenjianhong2@huawei.com X-Gm-Message-State: APjAAAW+PmlrRcrhZobFq2cqnT2OKes/4VcTgho3svD+997Z64+VMgpF eTm00R7kBAtVZruKQN8G6k1nbP5leJ60Ad0ZDHM8dp7Oxd6TuLavKx7jznuP4fqV1xj3D51SATZ U1g6vo8vJsihCathoMmaRZ46Q6DJc5qwrvLuRLNVklbnvh94dpKRbR4Tv7uIQRt9hEw== X-Received: by 2002:aca:dfc5:: with SMTP id w188mr455944oig.0.1558072957729; Thu, 16 May 2019 23:02:37 -0700 (PDT) X-Google-Smtp-Source: APXvYqwic9rhGzWWDnTaOR33aVttpvNKi5irchSip1HQw2lj+QvZKnDdoPSDR/XMdMlrP0fEbdGo X-Received: by 2002:aca:dfc5:: with SMTP id w188mr455928oig.0.1558072956059; Thu, 16 May 2019 23:02:36 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1558072956; cv=none; d=google.com; s=arc-20160816; b=jWP4M2Xwk+zBmpGpsrBXba3Oi6blnkRBj9+p8nfnd2yTW2XZhdWX0VY6sl0+bM6zot 9/60hqtc2GuvLV5ZftApbStsS0D4agb0qinC3o/jmqTLLvDla0SfWYCwfF1WcxICBvR5 0lKsUvYMemyZe3o7HHGZ95htiowM9DxLb3L2k8OtePe0zIuLNJonBwa/cyDuCZSNCHAS QHG6V8BupEGX5FLO6/QS0rJN7mB/Goq8AGLVi1/xK0TzZEN8d8sXvfMIXxRPYvPjwvwF y+87ovPjCVQITBvLkEbFu9KDSYUKZZPa3vqKAmmSi3n6DoZ/4Rv1XgwmMU0G6WEhON0g /Gpw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=content-transfer-encoding:mime-version:message-id:date:subject:cc :to:from; bh=DO6WLkoUqXxxlGmLPR+WFpvuFJaMgzTrrjSYWEqE6Ok=; b=HpKECbvfr8z22RftXIqvnXAi91e4w0oN+DWK3P5OAX9hUD3m+sLi4yGGhFWnTXAYWe SZ3KzmjTZlJztrmcUoPYUOyJFCVejpqCBPAJfsv+R9yLai0ZT9KazuUavPHu0VAwnNyP 66IxjPLJQILJtvtu++bMv3s79n0m+3jbHzPB40MzzLFBJyShA238Sd+LDWWCxT4khN2o 4jysUSspqPlApDz4NKO5VG9pfvpFbdNgtctbSbKsp/hZ6YsG+76VCWYqwrhxi40LwlpK XThsfYqquUOyTnO+XmVMxa/m9hAmQNwFNWRuXyuCqddjv1+w1gRbBqzkPnXPs5oxUSky AamQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of chenjianhong2@huawei.com designates 45.249.212.191 as permitted sender) smtp.mailfrom=chenjianhong2@huawei.com Received: from huawei.com (szxga05-in.huawei.com. [45.249.212.191]) by mx.google.com with ESMTPS id l15si3968853oib.157.2019.05.16.23.02.35 for (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Thu, 16 May 2019 23:02:36 -0700 (PDT) Received-SPF: pass (google.com: domain of chenjianhong2@huawei.com designates 45.249.212.191 as permitted sender) client-ip=45.249.212.191; Authentication-Results: mx.google.com; spf=pass (google.com: domain of chenjianhong2@huawei.com designates 45.249.212.191 as permitted sender) smtp.mailfrom=chenjianhong2@huawei.com Received: from DGGEMS407-HUB.china.huawei.com (unknown [172.30.72.60]) by Forcepoint Email with ESMTP id 5203840D642EC3827BCA; Fri, 17 May 2019 14:02:31 +0800 (CST) Received: from use12-sp2.huawei.com (10.67.188.162) by DGGEMS407-HUB.china.huawei.com (10.3.19.207) with Microsoft SMTP Server id 14.3.439.0; Fri, 17 May 2019 14:02:21 +0800 From: jianhong chen To: , , , , , , , , , CC: , , , Subject: [PATCH] mm/mmap: fix the adjusted length error Date: Fri, 17 May 2019 14:06:49 +0800 Message-ID: <1558073209-79549-1-git-send-email-chenjianhong2@huawei.com> X-Mailer: git-send-email 1.8.5.6 MIME-Version: 1.0 X-Originating-IP: [10.67.188.162] X-CFilter-Loop: Reflected X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: X-Virus-Scanned: ClamAV using ClamSMTP In linux version 4.4, a 32-bit process may fail to allocate 64M hugepage memory by function shmat even though there is a 64M memory gap in the process. It is the adjusted length that causes the problem, introduced from commit db4fbfb9523c935 ("mm: vm_unmapped_area() lookup function"). Accounting for the worst case alignment overhead, function unmapped_area and unmapped_area_topdown adjust the search length before searching for available vma gap. This is an estimated length, sum of the desired length and the longest alignment offset, which can cause misjudgement if the system has very few virtual memory left. For example, if the longest memory gap available is 64M, we can’t get it from the system by allocating 64M hugepage memory via shmat function. The reason is that it requires a longger length, the sum of the desired length(64M) and the longest alignment offset. To fix this error ,we can calculate the alignment offset of gap_start or gap_end to get a desired gap_start or gap_end value, before searching for the available gap. In this way, we don't need to adjust the search length. Problem reproduces procedure: 1. allocate a lot of virtual memory segments via shmat and malloc 2. release one of the biggest memory segment via shmdt 3. attach the biggest memory segment via shmat e.g. process maps: 00008000-00009000 r-xp 00000000 00:12 3385 /tmp/memory_mmap 00011000-00012000 rw-p 00001000 00:12 3385 /tmp/memory_mmap 27536000-f756a000 rw-p 00000000 00:00 0 f756a000-f7691000 r-xp 00000000 01:00 560 /lib/libc-2.11.1.so f7691000-f7699000 ---p 00127000 01:00 560 /lib/libc-2.11.1.so f7699000-f769b000 r--p 00127000 01:00 560 /lib/libc-2.11.1.so f769b000-f769c000 rw-p 00129000 01:00 560 /lib/libc-2.11.1.so f769c000-f769f000 rw-p 00000000 00:00 0 f769f000-f76c0000 r-xp 00000000 01:00 583 /lib/libgcc_s.so.1 f76c0000-f76c7000 ---p 00021000 01:00 583 /lib/libgcc_s.so.1 f76c7000-f76c8000 rw-p 00020000 01:00 583 /lib/libgcc_s.so.1 f76c8000-f76e5000 r-xp 00000000 01:00 543 /lib/ld-2.11.1.so f76e9000-f76ea000 rw-p 00000000 00:00 0 f76ea000-f76ec000 rw-p 00000000 00:00 0 f76ec000-f76ed000 r--p 0001c000 01:00 543 /lib/ld-2.11.1.so f76ed000-f76ee000 rw-p 0001d000 01:00 543 /lib/ld-2.11.1.so f7800000-f7a00000 rw-s 00000000 00:0e 0 /SYSV000000ea (deleted) fba00000-fca00000 rw-s 00000000 00:0e 65538 /SYSV000000ec (deleted) fca00000-fce00000 rw-s 00000000 00:0e 98307 /SYSV000000ed (deleted) fce00000-fd800000 rw-s 00000000 00:0e 131076 /SYSV000000ee (deleted) ff913000-ff934000 rw-p 00000000 00:00 0 [stack] ffff0000-ffff1000 r-xp 00000000 00:00 0 [vectors] from 0xf7a00000 to fba00000, it has 64M memory gap, but we can't get it from kernel. Signed-off-by: jianhong chen Cc: stable@vger.kernel.org --- mm/mmap.c | 43 +++++++++++++++++++++++++++++-------------- 1 file changed, 29 insertions(+), 14 deletions(-) diff --git a/mm/mmap.c b/mm/mmap.c index bd7b9f2..c5a5782 100644 --- a/mm/mmap.c +++ b/mm/mmap.c @@ -1865,6 +1865,22 @@ unsigned long mmap_region(struct file *file, unsigned long addr, return error; } +static inline unsigned long gap_start_offset(struct vm_unmapped_area_info *info, + unsigned long addr) +{ + /* get gap_start offset to adjust gap address to the + * desired alignment + */ + return (info->align_offset - addr) & info->align_mask; +} + +static inline unsigned long gap_end_offset(struct vm_unmapped_area_info *info, + unsigned long addr) +{ + /* get gap_end offset to adjust gap address to the desired alignment */ + return (addr - info->align_offset) & info->align_mask; +} + unsigned long unmapped_area(struct vm_unmapped_area_info *info) { /* @@ -1879,10 +1895,7 @@ unsigned long unmapped_area(struct vm_unmapped_area_info *info) struct vm_area_struct *vma; unsigned long length, low_limit, high_limit, gap_start, gap_end; - /* Adjust search length to account for worst case alignment overhead */ - length = info->length + info->align_mask; - if (length < info->length) - return -ENOMEM; + length = info->length; /* Adjust search limits by the desired length */ if (info->high_limit < length) @@ -1914,6 +1927,7 @@ unsigned long unmapped_area(struct vm_unmapped_area_info *info) } gap_start = vma->vm_prev ? vm_end_gap(vma->vm_prev) : 0; + gap_start += gap_start_offset(info, gap_start); check_current: /* Check if current node has a suitable gap */ if (gap_start > high_limit) @@ -1942,6 +1956,7 @@ unsigned long unmapped_area(struct vm_unmapped_area_info *info) struct vm_area_struct, vm_rb); if (prev == vma->vm_rb.rb_left) { gap_start = vm_end_gap(vma->vm_prev); + gap_start += gap_start_offset(info, gap_start); gap_end = vm_start_gap(vma); goto check_current; } @@ -1951,17 +1966,17 @@ unsigned long unmapped_area(struct vm_unmapped_area_info *info) check_highest: /* Check highest gap, which does not precede any rbtree node */ gap_start = mm->highest_vm_end; + gap_start += gap_start_offset(info, gap_start); gap_end = ULONG_MAX; /* Only for VM_BUG_ON below */ if (gap_start > high_limit) return -ENOMEM; found: /* We found a suitable gap. Clip it with the original low_limit. */ - if (gap_start < info->low_limit) + if (gap_start < info->low_limit) { gap_start = info->low_limit; - - /* Adjust gap address to the desired alignment */ - gap_start += (info->align_offset - gap_start) & info->align_mask; + gap_start += gap_start_offset(info, gap_start); + } VM_BUG_ON(gap_start + info->length > info->high_limit); VM_BUG_ON(gap_start + info->length > gap_end); @@ -1974,16 +1989,14 @@ unsigned long unmapped_area_topdown(struct vm_unmapped_area_info *info) struct vm_area_struct *vma; unsigned long length, low_limit, high_limit, gap_start, gap_end; - /* Adjust search length to account for worst case alignment overhead */ - length = info->length + info->align_mask; - if (length < info->length) - return -ENOMEM; + length = info->length; /* * Adjust search limits by the desired length. * See implementation comment at top of unmapped_area(). */ gap_end = info->high_limit; + gap_end -= gap_end_offset(info, gap_end); if (gap_end < length) return -ENOMEM; high_limit = gap_end - length; @@ -2020,6 +2033,7 @@ unsigned long unmapped_area_topdown(struct vm_unmapped_area_info *info) check_current: /* Check if current node has a suitable gap */ gap_end = vm_start_gap(vma); + gap_end -= gap_end_offset(info, gap_end); if (gap_end < low_limit) return -ENOMEM; if (gap_start <= high_limit && @@ -2054,13 +2068,14 @@ unsigned long unmapped_area_topdown(struct vm_unmapped_area_info *info) found: /* We found a suitable gap. Clip it with the original high_limit. */ - if (gap_end > info->high_limit) + if (gap_end > info->high_limit) { gap_end = info->high_limit; + gap_end -= gap_end_offset(info, gap_end); + } found_highest: /* Compute highest gap address at the desired alignment */ gap_end -= info->length; - gap_end -= (gap_end - info->align_offset) & info->align_mask; VM_BUG_ON(gap_end < info->low_limit); VM_BUG_ON(gap_end < gap_start);