Message ID | 20230824080408.2933205-1-stevensd@google.com (mailing list archive) |
---|---|
Headers | show
Return-Path: <kvm-owner@vger.kernel.org> X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 54002C7113B for <kvm@archiver.kernel.org>; Thu, 24 Aug 2023 08:06:03 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S236226AbjHXIFd (ORCPT <rfc822;kvm@archiver.kernel.org>); Thu, 24 Aug 2023 04:05:33 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:45990 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S235718AbjHXIFX (ORCPT <rfc822;kvm@vger.kernel.org>); Thu, 24 Aug 2023 04:05:23 -0400 Received: from mail-oi1-x22d.google.com (mail-oi1-x22d.google.com [IPv6:2607:f8b0:4864:20::22d]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 1A2A8101 for <kvm@vger.kernel.org>; Thu, 24 Aug 2023 01:04:50 -0700 (PDT) Received: by mail-oi1-x22d.google.com with SMTP id 5614622812f47-3a412653335so4757763b6e.1 for <kvm@vger.kernel.org>; Thu, 24 Aug 2023 01:04:50 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=chromium.org; s=google; t=1692864289; x=1693469089; h=content-transfer-encoding:mime-version:message-id:date:subject:cc :to:from:from:to:cc:subject:date:message-id:reply-to; bh=5EE+UzoRngnjehf25g1txAUD7jFYvBFO8xVG5NUnepk=; b=PiabMMQhe9y2vUBAt7h+ntRdrItXl5XEmXNUAkQzHdz/5Usah2vUewJPe1DxJjyMoj MPwJ8vBC0PRuHBzkM99uP/8pDt3bKa4MMVAyZvrVMT5+kHkXevDD6Xbkw9rXNNage966 g+94vbQfs3rPzxqPbviDDzoC4v8GyCo9zgfVo= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20221208; t=1692864289; x=1693469089; h=content-transfer-encoding:mime-version:message-id:date:subject:cc :to:from:x-gm-message-state:from:to:cc:subject:date:message-id :reply-to; bh=5EE+UzoRngnjehf25g1txAUD7jFYvBFO8xVG5NUnepk=; b=NltMtdyXnKDBQ5dlsFx+voFqNkHwyurGiG7efY2eEDOtB6CThvSSFGJxXzWaY59cex mGsjFVDsEoFgmq3ejKb2Vna6BVJvm1ibyjwN7yr2kbz6OEgl0wTPjs2oFDZ8bqgbaYeI S4O10g4fe2JR51vEs0ilXh9EhBKNxDM+YX1q9F7HG/KyBFPJoRmvtXR+Nm5gLKE4kX9I N2qqOPwvepLs3FtF89m0ph2Rp3Xhm/zlxh+MzQ6kZsznK1MnELeg02t38sknAz4yd87+ VWQslGAC+M9P5mAZKWoyw3WX4H1y0nX4aPTXtUxagg7LKCCQty1j97cplCkUhXyttyML dy2g== X-Gm-Message-State: AOJu0YwPVjzPjZBjooEY3bVvX7o1GWiwxiyEkBOSZNvCGf5DJM5hjNca 32yuCb/lorwcQG0nxyg9LQj04A== X-Google-Smtp-Source: AGHT+IHkBcWKFj3+pW5fuuiWfLnbD3ta+ABypKF47WzCDkfdw67HWPAGK/Ns2WRoKXuxgQGkt6DXAw== X-Received: by 2002:a05:6358:720c:b0:13c:c84b:88bb with SMTP id h12-20020a056358720c00b0013cc84b88bbmr2497118rwa.27.1692864289233; Thu, 24 Aug 2023 01:04:49 -0700 (PDT) Received: from localhost ([2401:fa00:8f:203:515:8b2a:90c3:b79e]) by smtp.gmail.com with UTF8SMTPSA id a18-20020aa78652000000b00666e649ca46sm7014327pfo.101.2023.08.24.01.04.46 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Thu, 24 Aug 2023 01:04:48 -0700 (PDT) From: David Stevens <stevensd@chromium.org> X-Google-Original-From: David Stevens <stevensd@google.com> To: Sean Christopherson <seanjc@google.com> Cc: Yu Zhang <yu.c.zhang@linux.intel.com>, Isaku Yamahata <isaku.yamahata@gmail.com>, Marc Zyngier <maz@kernel.org>, Michael Ellerman <mpe@ellerman.id.au>, Peter Xu <peterx@redhat.com>, linux-arm-kernel@lists.infradead.org, kvmarm@lists.linux.dev, linux-kernel@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, kvm@vger.kernel.org, David Stevens <stevensd@chromium.org> Subject: [PATCH v8 0/8] KVM: allow mapping non-refcounted pages Date: Thu, 24 Aug 2023 17:04:00 +0900 Message-ID: <20230824080408.2933205-1-stevensd@google.com> X-Mailer: git-send-email 2.42.0.rc1.204.g551eb34607-goog MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Precedence: bulk List-ID: <kvm.vger.kernel.org> X-Mailing-List: kvm@vger.kernel.org |
Series |
KVM: allow mapping non-refcounted pages
|
expand
|
From: David Stevens <stevensd@chromium.org> This patch series adds support for mapping VM_IO and VM_PFNMAP memory that is backed by struct pages that aren't currently being refcounted (e.g. tail pages of non-compound higher order allocations) into the guest. Our use case is virtio-gpu blob resources [1], which directly map host graphics buffers into the guest as "vram" for the virtio-gpu device. This feature currently does not work on systems using the amdgpu driver, as that driver allocates non-compound higher order pages via ttm_pool_alloc_page. First, this series replaces the __gfn_to_pfn_memslot API with a more extensible __kvm_faultin_pfn API. The updated API rearranges __gfn_to_pfn_memslot's args into a struct and where possible packs the bool arguments into a FOLL_ flags argument. The refactoring changes do not change any behavior, except as noted in the PPC change. When introduced in the refactoring, __kvm_faultin_pfn implies FOLL_GET to preserve existing behavior. From there, the API is made to support mapping non-refcounted pages by respecting the FOLL_GET flag. This series only adds support for non-refcounted pages to the x86 MMU. Other MMUs can likely be updated without too much difficulty, but it is not needed at this point. Updating other parts of KVM (e.g. pfncache) is not straightforward [2]. [1] https://patchwork.kernel.org/project/dri-devel/cover/20200814024000.2485-1-gurchetansingh@chromium.org/ [2] https://lore.kernel.org/all/ZBEEQtmtNPaEqU1i@google.com/ v7 -> v8: - Set access bits before releasing mmu_lock. - Pass FOLL_GET on 32-bit x86 or !tdp_enabled. - Refactor FOLL_GET handling, add kvm_follow_refcounted_pfn helper. - Set refcounted bit on >4k pages. - Add comments and apply formatting suggestions. - rebase on kvm next branch. v6 -> v7: - Replace __gfn_to_pfn_memslot with a more flexible __kvm_faultin_pfn, and extend that API to support non-refcounted pages. v5 -> v6: - rebase on kvm next branch - rename gfn_to_pfn_page to gfn_to_pfn_noref - fix uninitialized outparam in error case of __kvm_faultin_pfn - add kvm_release_pfn_noref_clean for releasing pfn/page pair v4 -> v5: - rebase on kvm next branch again v3 -> v4: - rebase on kvm next branch again - Add some more context to a comment in ensure_pfn_ref v2 -> v3: - rebase on kvm next branch v1 -> v2: - Introduce new gfn_to_pfn_page functions instead of modifying the behavior of existing gfn_to_pfn functions, to make the change less invasive. - Drop changes to mmu_audit.c - Include Nicholas Piggin's patch to avoid corrupting refcount in the follow_pte case, and use it in depreciated gfn_to_pfn functions. - Rebase on kvm/next David Stevens (7): KVM: mmu: Introduce __kvm_follow_pfn function KVM: mmu: Make __kvm_follow_pfn not imply FOLL_GET KVM: x86/mmu: Migrate to __kvm_follow_pfn KVM: x86/mmu: Don't pass FOLL_GET to __kvm_follow_pfn KVM: arm64: Migrate to __kvm_follow_pfn KVM: PPC: Migrate to __kvm_follow_pfn KVM: mmu: remove __gfn_to_pfn_memslot Sean Christopherson (1): KVM: Assert that a page's refcount is elevated when marking accessed/dirty arch/arm64/kvm/mmu.c | 25 +-- arch/powerpc/include/asm/kvm_book3s.h | 2 +- arch/powerpc/kvm/book3s_64_mmu_hv.c | 38 ++-- arch/powerpc/kvm/book3s_64_mmu_radix.c | 50 +++--- arch/powerpc/kvm/book3s_hv_nested.c | 4 +- arch/x86/kvm/mmu/mmu.c | 94 +++++++--- arch/x86/kvm/mmu/mmu_internal.h | 1 + arch/x86/kvm/mmu/paging_tmpl.h | 8 +- arch/x86/kvm/mmu/spte.c | 4 +- arch/x86/kvm/mmu/spte.h | 12 +- arch/x86/kvm/mmu/tdp_mmu.c | 22 ++- include/linux/kvm_host.h | 26 +++ virt/kvm/kvm_main.c | 231 ++++++++++++++----------- virt/kvm/kvm_mm.h | 3 +- virt/kvm/pfncache.c | 10 +- 15 files changed, 316 insertions(+), 214 deletions(-)