From patchwork Mon Aug 20 13:52:33 2012 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: "Kirill A. Shutemov" X-Patchwork-Id: 1348711 Return-Path: X-Original-To: patchwork-linux-sh@patchwork.kernel.org Delivered-To: patchwork-process-083081@patchwork1.kernel.org Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by patchwork1.kernel.org (Postfix) with ESMTP id 0BFC83FD40 for ; Mon, 20 Aug 2012 13:54:59 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752341Ab2HTNwn (ORCPT ); Mon, 20 Aug 2012 09:52:43 -0400 Received: from mga03.intel.com ([143.182.124.21]:55024 "EHLO mga03.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1753122Ab2HTNwj (ORCPT ); Mon, 20 Aug 2012 09:52:39 -0400 Received: from azsmga002.ch.intel.com ([10.2.17.35]) by azsmga101.ch.intel.com with ESMTP; 20 Aug 2012 06:52:38 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="4.77,797,1336374000"; d="scan'208";a="136222982" Received: from blue.fi.intel.com ([10.237.72.50]) by AZSMGA002.ch.intel.com with ESMTP; 20 Aug 2012 06:52:32 -0700 Received: by blue.fi.intel.com (Postfix, from userid 1000) id 8962AE0085; Mon, 20 Aug 2012 16:52:43 +0300 (EEST) From: "Kirill A. Shutemov" To: linux-mm@kvack.org Cc: Thomas Gleixner , Ingo Molnar , "H. Peter Anvin" , x86@kernel.org, Andi Kleen , "Kirill A. Shutemov" , Tim Chen , Alex Shi , Jan Beulich , Robert Richter , Andy Lutomirski , Andrew Morton , Andrea Arcangeli , Johannes Weiner , Hugh Dickins , KAMEZAWA Hiroyuki , Mel Gorman , linux-kernel@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, linux-mips@linux-mips.org, linux-sh@vger.kernel.org, sparclinux@vger.kernel.org Subject: [PATCH v4 4/8] mm: pass fault address to clear_huge_page() Date: Mon, 20 Aug 2012 16:52:33 +0300 Message-Id: <1345470757-12005-5-git-send-email-kirill.shutemov@linux.intel.com> X-Mailer: git-send-email 1.7.10.4 In-Reply-To: <1345470757-12005-1-git-send-email-kirill.shutemov@linux.intel.com> References: <1345470757-12005-1-git-send-email-kirill.shutemov@linux.intel.com> Sender: linux-sh-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-sh@vger.kernel.org From: "Kirill A. Shutemov" Signed-off-by: Kirill A. Shutemov --- include/linux/mm.h | 2 +- mm/huge_memory.c | 2 +- mm/hugetlb.c | 3 ++- mm/memory.c | 7 ++++--- 4 files changed, 8 insertions(+), 6 deletions(-) diff --git a/include/linux/mm.h b/include/linux/mm.h index 311be90..2858723 100644 --- a/include/linux/mm.h +++ b/include/linux/mm.h @@ -1638,7 +1638,7 @@ extern void dump_page(struct page *page); #if defined(CONFIG_TRANSPARENT_HUGEPAGE) || defined(CONFIG_HUGETLBFS) extern void clear_huge_page(struct page *page, - unsigned long addr, + unsigned long haddr, unsigned long fault_address, unsigned int pages_per_huge_page); extern void copy_user_huge_page(struct page *dst, struct page *src, unsigned long addr, struct vm_area_struct *vma, diff --git a/mm/huge_memory.c b/mm/huge_memory.c index 6f0825b611..070bf89 100644 --- a/mm/huge_memory.c +++ b/mm/huge_memory.c @@ -644,7 +644,7 @@ static int __do_huge_pmd_anonymous_page(struct mm_struct *mm, if (unlikely(!pgtable)) return VM_FAULT_OOM; - clear_huge_page(page, haddr, HPAGE_PMD_NR); + clear_huge_page(page, haddr, address, HPAGE_PMD_NR); __SetPageUptodate(page); spin_lock(&mm->page_table_lock); diff --git a/mm/hugetlb.c b/mm/hugetlb.c index 3c86d3d..5182192 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c @@ -2718,7 +2718,8 @@ retry: ret = VM_FAULT_SIGBUS; goto out; } - clear_huge_page(page, haddr, pages_per_huge_page(h)); + clear_huge_page(page, haddr, fault_address, + pages_per_huge_page(h)); __SetPageUptodate(page); if (vma->vm_flags & VM_MAYSHARE) { diff --git a/mm/memory.c b/mm/memory.c index 5736170..dfc179b 100644 --- a/mm/memory.c +++ b/mm/memory.c @@ -3984,19 +3984,20 @@ static void clear_gigantic_page(struct page *page, } } void clear_huge_page(struct page *page, - unsigned long addr, unsigned int pages_per_huge_page) + unsigned long haddr, unsigned long fault_address, + unsigned int pages_per_huge_page) { int i; if (unlikely(pages_per_huge_page > MAX_ORDER_NR_PAGES)) { - clear_gigantic_page(page, addr, pages_per_huge_page); + clear_gigantic_page(page, haddr, pages_per_huge_page); return; } might_sleep(); for (i = 0; i < pages_per_huge_page; i++) { cond_resched(); - clear_user_highpage(page + i, addr + i * PAGE_SIZE); + clear_user_highpage(page + i, haddr + i * PAGE_SIZE); } }