From patchwork Fri Jan 29 07:39:46 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Anshuman Khandual X-Patchwork-Id: 12055419 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-17.0 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER, INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED, USER_AGENT_GIT autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 7A2A4C433E0 for ; Fri, 29 Jan 2021 07:41:50 +0000 (UTC) Received: from merlin.infradead.org (merlin.infradead.org [205.233.59.134]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 2784E64E00 for ; Fri, 29 Jan 2021 07:41:50 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 2784E64E00 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=arm.com Authentication-Results: mail.kernel.org; spf=none smtp.mailfrom=linux-arm-kernel-bounces+linux-arm-kernel=archiver.kernel.org@lists.infradead.org DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=lists.infradead.org; s=merlin.20170209; h=Sender:Content-Transfer-Encoding: Content-Type:MIME-Version:Cc:List-Subscribe:List-Help:List-Post:List-Archive: List-Unsubscribe:List-Id:References:In-Reply-To:Message-Id:Date:Subject:To: From:Reply-To:Content-ID:Content-Description:Resent-Date:Resent-From: Resent-Sender:Resent-To:Resent-Cc:Resent-Message-ID:List-Owner; bh=ElXCnzsih/3uQmUlN/2PR/P1F+RZRLOStAIKvyRCt6s=; b=cLOANip4W2BnEgi8riMVuaEzMy 4vNH3A2ZDQZRFUOYaYYKQoP1k99WQXlF58c5vt9U/o2z7/ovcMkGFzW96QlxBPlmeyc0X/ucueZpb P5oXSV6JE1dr6jCeEqPhMFfR0KvuYLEinyU+4l7UBbXghagBc9AEoCzpQzhKQhmFkBVw9W3bPxcLr E5FFUSVwlc6Udk6ffCZ0xvn1E4DeHRHppost7mBafoPFsQQ4uloHyn8glaSJPqXBTz0Da5L0tJp0O V/kAv/AvuCiAgjfUE30rgYlkLRvMfogEu9oG7esgqL0X/+mYhf4LTws/vzSff9R41CsJRD/1BZNWn 67AjJ4Eg==; Received: from localhost ([::1] helo=merlin.infradead.org) by merlin.infradead.org with esmtp (Exim 4.92.3 #3 (Red Hat Linux)) id 1l5OOU-0004uA-Ge; Fri, 29 Jan 2021 07:40:38 +0000 Received: from foss.arm.com ([217.140.110.172]) by merlin.infradead.org with esmtp (Exim 4.92.3 #3 (Red Hat Linux)) id 1l5ONs-0004hV-68 for linux-arm-kernel@lists.infradead.org; Fri, 29 Jan 2021 07:40:04 +0000 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 7775C1509; Thu, 28 Jan 2021 23:39:59 -0800 (PST) Received: from p8cg001049571a15.arm.com (unknown [10.163.92.245]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPA id 561453F719; Thu, 28 Jan 2021 23:39:54 -0800 (PST) From: Anshuman Khandual To: linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org Subject: [PATCH 2/2] arm64/mm: Reorganize pfn_valid() Date: Fri, 29 Jan 2021 13:09:46 +0530 Message-Id: <1611905986-20155-3-git-send-email-anshuman.khandual@arm.com> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1611905986-20155-1-git-send-email-anshuman.khandual@arm.com> References: <1611905986-20155-1-git-send-email-anshuman.khandual@arm.com> X-CRM114-Version: 20100106-BlameMichelson ( TRE 0.8.0 (BSD) ) MR-646709E3 X-CRM114-CacheID: sfid-20210129_024000_379093_3EB324F1 X-CRM114-Status: GOOD ( 17.48 ) X-BeenThere: linux-arm-kernel@lists.infradead.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Mark Rutland , Anshuman Khandual , Catalin Marinas , David Hildenbrand , Robin Murphy , Mike Rapoport , =?utf-8?b?SsOpcsO0bWUgR2xpc3Nl?= , James Morse , Dan Williams , Will Deacon , Ard Biesheuvel MIME-Version: 1.0 Sender: "linux-arm-kernel" Errors-To: linux-arm-kernel-bounces+linux-arm-kernel=archiver.kernel.org@lists.infradead.org There are multiple instances of pfn_to_section_nr() and __pfn_to_section() when CONFIG_SPARSEMEM is enabled. This can be just optimized if the memory section is fetched earlier. Hence bifurcate pfn_valid() into two different definitions depending on whether CONFIG_SPARSEMEM is enabled. Also replace the open coded pfn <--> addr conversion with __[pfn|phys]_to_[phys|pfn](). This does not cause any functional change. Cc: Catalin Marinas Cc: Will Deacon Cc: Ard Biesheuvel Cc: linux-arm-kernel@lists.infradead.org Cc: linux-kernel@vger.kernel.org Signed-off-by: Anshuman Khandual --- arch/arm64/mm/init.c | 38 +++++++++++++++++++++++++++++++------- 1 file changed, 31 insertions(+), 7 deletions(-) diff --git a/arch/arm64/mm/init.c b/arch/arm64/mm/init.c index 1141075e4d53..09adca90c57a 100644 --- a/arch/arm64/mm/init.c +++ b/arch/arm64/mm/init.c @@ -217,18 +217,25 @@ static void __init zone_sizes_init(unsigned long min, unsigned long max) free_area_init(max_zone_pfns); } +#ifdef CONFIG_SPARSEMEM int pfn_valid(unsigned long pfn) { - phys_addr_t addr = pfn << PAGE_SHIFT; + struct mem_section *ms = __pfn_to_section(pfn); + phys_addr_t addr = __pfn_to_phys(pfn); - if ((addr >> PAGE_SHIFT) != pfn) + /* + * Ensure the upper PAGE_SHIFT bits are clear in the + * pfn. Else it might lead to false positives when + * some of the upper bits are set, but the lower bits + * match a valid pfn. + */ + if (__phys_to_pfn(addr) != pfn) return 0; -#ifdef CONFIG_SPARSEMEM if (pfn_to_section_nr(pfn) >= NR_MEM_SECTIONS) return 0; - if (!valid_section(__pfn_to_section(pfn))) + if (!valid_section(ms)) return 0; /* @@ -240,11 +247,28 @@ int pfn_valid(unsigned long pfn) * memory sections covering all of hotplug memory including * both normal and ZONE_DEVICE based. */ - if (!early_section(__pfn_to_section(pfn))) - return pfn_section_valid(__pfn_to_section(pfn), pfn); -#endif + if (!early_section(ms)) + return pfn_section_valid(ms, pfn); + return memblock_is_map_memory(addr); } +#else +int pfn_valid(unsigned long pfn) +{ + phys_addr_t addr = __pfn_to_phys(pfn); + + /* + * Ensure the upper PAGE_SHIFT bits are clear in the + * pfn. Else it might lead to false positives when + * some of the upper bits are set, but the lower bits + * match a valid pfn. + */ + if (__phys_to_pfn(addr) != pfn) + return 0; + + return memblock_is_map_memory(addr); +} +#endif EXPORT_SYMBOL(pfn_valid); static phys_addr_t memory_limit = PHYS_ADDR_MAX;