Message ID | 20230911021637.1941096-7-stevensd@google.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | KVM: allow mapping non-refcounted pages | expand |
On 9/11/23 05:16, David Stevens wrote: > --- a/arch/x86/kvm/mmu/paging_tmpl.h > +++ b/arch/x86/kvm/mmu/paging_tmpl.h > @@ -848,7 +848,8 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault > > out_unlock: > write_unlock(&vcpu->kvm->mmu_lock); > - kvm_release_pfn_clean(fault->pfn); > + if (fault->is_refcounted_page) > + kvm_set_page_accessed(pfn_to_page(fault->pfn)); The other similar occurrences in the code that replaced kvm_release_pfn_clean() with kvm_set_page_accessed() did it under the held mmu_lock. Does kvm_set_page_accessed() needs to be invoked under the lock?
On 9/11/23 05:16, David Stevens wrote: > From: David Stevens <stevensd@chromium.org> > > Handle non-refcounted pages in __kvm_faultin_pfn. This allows the host > to map memory into the guest that is backed by non-refcounted struct > pages - for example, the tail pages of higher order non-compound pages > allocated by the amdgpu driver via ttm_pool_alloc_page. > > The bulk of this change is tracking the is_refcounted_page flag so that > non-refcounted pages don't trigger page_count() == 0 warnings. This is > done by storing the flag in an unused bit in the sptes. There are no > bits available in PAE SPTEs, so non-refcounted pages can only be handled > on TDP and x86-64. > > Signed-off-by: David Stevens <stevensd@chromium.org> > --- > arch/x86/kvm/mmu/mmu.c | 52 +++++++++++++++++++++++---------- > arch/x86/kvm/mmu/mmu_internal.h | 1 + > arch/x86/kvm/mmu/paging_tmpl.h | 8 +++-- > arch/x86/kvm/mmu/spte.c | 4 ++- > arch/x86/kvm/mmu/spte.h | 12 +++++++- > arch/x86/kvm/mmu/tdp_mmu.c | 22 ++++++++------ > include/linux/kvm_host.h | 3 ++ > virt/kvm/kvm_main.c | 6 ++-- > 8 files changed, 76 insertions(+), 32 deletions(-) Could you please tell which kernel tree you used for the base of this series? This patch #6 doesn't apply cleanly to stable/mainline/next/kvm error: sha1 information is lacking or useless (arch/x86/kvm/mmu/mmu.c). error: could not build fake ancestor
On 9/18/23 12:58, Dmitry Osipenko wrote: > On 9/11/23 05:16, David Stevens wrote: >> From: David Stevens <stevensd@chromium.org> >> >> Handle non-refcounted pages in __kvm_faultin_pfn. This allows the host >> to map memory into the guest that is backed by non-refcounted struct >> pages - for example, the tail pages of higher order non-compound pages >> allocated by the amdgpu driver via ttm_pool_alloc_page. >> >> The bulk of this change is tracking the is_refcounted_page flag so that >> non-refcounted pages don't trigger page_count() == 0 warnings. This is >> done by storing the flag in an unused bit in the sptes. There are no >> bits available in PAE SPTEs, so non-refcounted pages can only be handled >> on TDP and x86-64. >> >> Signed-off-by: David Stevens <stevensd@chromium.org> >> --- >> arch/x86/kvm/mmu/mmu.c | 52 +++++++++++++++++++++++---------- >> arch/x86/kvm/mmu/mmu_internal.h | 1 + >> arch/x86/kvm/mmu/paging_tmpl.h | 8 +++-- >> arch/x86/kvm/mmu/spte.c | 4 ++- >> arch/x86/kvm/mmu/spte.h | 12 +++++++- >> arch/x86/kvm/mmu/tdp_mmu.c | 22 ++++++++------ >> include/linux/kvm_host.h | 3 ++ >> virt/kvm/kvm_main.c | 6 ++-- >> 8 files changed, 76 insertions(+), 32 deletions(-) > > Could you please tell which kernel tree you used for the base of this > series? This patch #6 doesn't apply cleanly to stable/mainline/next/kvm > > error: sha1 information is lacking or useless (arch/x86/kvm/mmu/mmu.c). > error: could not build fake ancestor I applied the patch manually to v6.5.2 and tested Venus using Intel TGL iGPU, the intel driver is crashing: BUG: kernel NULL pointer dereference, address: 0000000000000058 #PF: supervisor read access in kernel mode #PF: error_code(0x0000) - not-present page PGD 0 P4D 0 Oops: 0000 [#1] PREEMPT SMP CPU: 1 PID: 5926 Comm: qemu-system-x86 Not tainted 6.5.2+ #114 Hardware name: LENOVO 20VE/LNVNB161216, BIOS F8CN43WW(V2.06) 08/12/2021 RIP: 0010:gen8_ppgtt_insert+0x50b/0x8f0 Code: 00 00 f7 c2 00 00 20 00 74 15 f7 c3 ff ff 1f 00 75 0d 41 81 fc ff ff 1f 00 0f 87 0e 02 00 00 48 8b 74 24 08 44 89 c0 45 85 ed <48> 8b 4e 58 48 8b 04 c1 0f 85 0b 02 00 00 81 e2 00 00 01 00 0f 84 RSP: 0018:ffffafc085afb820 EFLAGS: 00010246 RAX: 0000000000000000 RBX: 00000000e9604000 RCX: 000000000000001b RDX: 0000000000211000 RSI: 0000000000000000 RDI: ffff9513d44c1000 RBP: ffff951106f8dfc0 R08: 0000000000000000 R09: 0000000000000003 R10: 0000000000000fff R11: 00000000e9800000 R12: 00000000001fc000 R13: 0000000000000000 R14: 0000000000001000 R15: 0000ffff00000000 FS: 00007f2a5bcced80(0000) GS:ffff951a87a40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000058 CR3: 0000000116f16006 CR4: 0000000000772ee0 PKRU: 55555554 Call Trace: <TASK> ? __die+0x1f/0x60 ? page_fault_oops+0x14d/0x420 ? exc_page_fault+0x3d7/0x880 ? lock_acquire+0xc9/0x290 ? asm_exc_page_fault+0x22/0x30 ? gen8_ppgtt_insert+0x50b/0x8f0 ppgtt_bind_vma+0x4f/0x60 fence_work+0x1b/0x70 fence_notify+0x8f/0x130 __i915_sw_fence_complete+0x58/0x230 i915_vma_pin_ww+0x513/0xa80 eb_validate_vmas+0x17e/0x9e0 ? eb_pin_engine+0x2bb/0x340 i915_gem_do_execbuffer+0xc85/0x2bf0 ? __lock_acquire+0x3b6/0x21c0 i915_gem_execbuffer2_ioctl+0xee/0x240 ? i915_gem_do_execbuffer+0x2bf0/0x2bf0 drm_ioctl_kernel+0x9d/0x140 drm_ioctl+0x1dd/0x410 ? i915_gem_do_execbuffer+0x2bf0/0x2bf0 ? __fget_files+0xc5/0x170 __x64_sys_ioctl+0x8c/0xc0 do_syscall_64+0x34/0x80 entry_SYSCALL_64_after_hwframe+0x46/0xb0 RIP: 0033:0x7f2a60b0c9df $ ./scripts/faddr2line ./vmlinux gen8_ppgtt_insert+0x50b/0x8f0 gen8_ppgtt_insert+0x50b/0x8f0: i915_pt_entry at drivers/gpu/drm/i915/gt/intel_gtt.h:557 (inlined by) gen8_ppgtt_insert_huge at drivers/gpu/drm/i915/gt/gen8_ppgtt.c:641 (inlined by) gen8_ppgtt_insert at drivers/gpu/drm/i915/gt/gen8_ppgtt.c:743 It's likely should be the i915 driver issue that is crashes with the NULL deref, but the origin of the bug should be the kvm page fault handling. David, could you please tell what tests you've run and post a link to yours kernel tree? Maybe I made obscure mistake while applied the patch manually.
On Mon, Sep 18, 2023 at 6:58 PM Dmitry Osipenko <dmitry.osipenko@collabora.com> wrote: > > On 9/11/23 05:16, David Stevens wrote: > > From: David Stevens <stevensd@chromium.org> > > > > Handle non-refcounted pages in __kvm_faultin_pfn. This allows the host > > to map memory into the guest that is backed by non-refcounted struct > > pages - for example, the tail pages of higher order non-compound pages > > allocated by the amdgpu driver via ttm_pool_alloc_page. > > > > The bulk of this change is tracking the is_refcounted_page flag so that > > non-refcounted pages don't trigger page_count() == 0 warnings. This is > > done by storing the flag in an unused bit in the sptes. There are no > > bits available in PAE SPTEs, so non-refcounted pages can only be handled > > on TDP and x86-64. > > > > Signed-off-by: David Stevens <stevensd@chromium.org> > > --- > > arch/x86/kvm/mmu/mmu.c | 52 +++++++++++++++++++++++---------- > > arch/x86/kvm/mmu/mmu_internal.h | 1 + > > arch/x86/kvm/mmu/paging_tmpl.h | 8 +++-- > > arch/x86/kvm/mmu/spte.c | 4 ++- > > arch/x86/kvm/mmu/spte.h | 12 +++++++- > > arch/x86/kvm/mmu/tdp_mmu.c | 22 ++++++++------ > > include/linux/kvm_host.h | 3 ++ > > virt/kvm/kvm_main.c | 6 ++-- > > 8 files changed, 76 insertions(+), 32 deletions(-) > > Could you please tell which kernel tree you used for the base of this > series? This patch #6 doesn't apply cleanly to stable/mainline/next/kvm > > error: sha1 information is lacking or useless (arch/x86/kvm/mmu/mmu.c). > error: could not build fake ancestor This series is based on the kvm next branch (i.e. https://git.kernel.org/pub/scm/virt/kvm/kvm.git/log/?h=next). The specific hash is d011151616e73de20c139580b73fa4c7042bd861. -David
On Mon, Sep 18, 2023 at 8:19 PM Dmitry Osipenko <dmitry.osipenko@collabora.com> wrote: > > On 9/18/23 12:58, Dmitry Osipenko wrote: > > On 9/11/23 05:16, David Stevens wrote: > >> From: David Stevens <stevensd@chromium.org> > >> > >> Handle non-refcounted pages in __kvm_faultin_pfn. This allows the host > >> to map memory into the guest that is backed by non-refcounted struct > >> pages - for example, the tail pages of higher order non-compound pages > >> allocated by the amdgpu driver via ttm_pool_alloc_page. > >> > >> The bulk of this change is tracking the is_refcounted_page flag so that > >> non-refcounted pages don't trigger page_count() == 0 warnings. This is > >> done by storing the flag in an unused bit in the sptes. There are no > >> bits available in PAE SPTEs, so non-refcounted pages can only be handled > >> on TDP and x86-64. > >> > >> Signed-off-by: David Stevens <stevensd@chromium.org> > >> --- > >> arch/x86/kvm/mmu/mmu.c | 52 +++++++++++++++++++++++---------- > >> arch/x86/kvm/mmu/mmu_internal.h | 1 + > >> arch/x86/kvm/mmu/paging_tmpl.h | 8 +++-- > >> arch/x86/kvm/mmu/spte.c | 4 ++- > >> arch/x86/kvm/mmu/spte.h | 12 +++++++- > >> arch/x86/kvm/mmu/tdp_mmu.c | 22 ++++++++------ > >> include/linux/kvm_host.h | 3 ++ > >> virt/kvm/kvm_main.c | 6 ++-- > >> 8 files changed, 76 insertions(+), 32 deletions(-) > > > > Could you please tell which kernel tree you used for the base of this > > series? This patch #6 doesn't apply cleanly to stable/mainline/next/kvm > > > > error: sha1 information is lacking or useless (arch/x86/kvm/mmu/mmu.c). > > error: could not build fake ancestor > > I applied the patch manually to v6.5.2 and tested Venus using Intel TGL iGPU, the intel driver is crashing: > > BUG: kernel NULL pointer dereference, address: 0000000000000058 > #PF: supervisor read access in kernel mode > #PF: error_code(0x0000) - not-present page > PGD 0 P4D 0 > Oops: 0000 [#1] PREEMPT SMP > CPU: 1 PID: 5926 Comm: qemu-system-x86 Not tainted 6.5.2+ #114 > Hardware name: LENOVO 20VE/LNVNB161216, BIOS F8CN43WW(V2.06) 08/12/2021 > RIP: 0010:gen8_ppgtt_insert+0x50b/0x8f0 > Code: 00 00 f7 c2 00 00 20 00 74 15 f7 c3 ff ff 1f 00 75 0d 41 81 fc ff ff 1f 00 0f 87 0e 02 00 00 48 8b 74 24 08 44 89 c0 45 85 ed <48> 8b 4e 58 48 8b 04 c1 0f 85 0b 02 00 00 81 e2 00 00 01 00 0f 84 > RSP: 0018:ffffafc085afb820 EFLAGS: 00010246 > RAX: 0000000000000000 RBX: 00000000e9604000 RCX: 000000000000001b > RDX: 0000000000211000 RSI: 0000000000000000 RDI: ffff9513d44c1000 > RBP: ffff951106f8dfc0 R08: 0000000000000000 R09: 0000000000000003 > R10: 0000000000000fff R11: 00000000e9800000 R12: 00000000001fc000 > R13: 0000000000000000 R14: 0000000000001000 R15: 0000ffff00000000 > FS: 00007f2a5bcced80(0000) GS:ffff951a87a40000(0000) knlGS:0000000000000000 > CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > CR2: 0000000000000058 CR3: 0000000116f16006 CR4: 0000000000772ee0 > PKRU: 55555554 > Call Trace: > <TASK> > ? __die+0x1f/0x60 > ? page_fault_oops+0x14d/0x420 > ? exc_page_fault+0x3d7/0x880 > ? lock_acquire+0xc9/0x290 > ? asm_exc_page_fault+0x22/0x30 > ? gen8_ppgtt_insert+0x50b/0x8f0 > ppgtt_bind_vma+0x4f/0x60 > fence_work+0x1b/0x70 > fence_notify+0x8f/0x130 > __i915_sw_fence_complete+0x58/0x230 > i915_vma_pin_ww+0x513/0xa80 > eb_validate_vmas+0x17e/0x9e0 > ? eb_pin_engine+0x2bb/0x340 > i915_gem_do_execbuffer+0xc85/0x2bf0 > ? __lock_acquire+0x3b6/0x21c0 > i915_gem_execbuffer2_ioctl+0xee/0x240 > ? i915_gem_do_execbuffer+0x2bf0/0x2bf0 > drm_ioctl_kernel+0x9d/0x140 > drm_ioctl+0x1dd/0x410 > ? i915_gem_do_execbuffer+0x2bf0/0x2bf0 > ? __fget_files+0xc5/0x170 > __x64_sys_ioctl+0x8c/0xc0 > do_syscall_64+0x34/0x80 > entry_SYSCALL_64_after_hwframe+0x46/0xb0 > RIP: 0033:0x7f2a60b0c9df > > > $ ./scripts/faddr2line ./vmlinux gen8_ppgtt_insert+0x50b/0x8f0 > gen8_ppgtt_insert+0x50b/0x8f0: > i915_pt_entry at drivers/gpu/drm/i915/gt/intel_gtt.h:557 > (inlined by) gen8_ppgtt_insert_huge at drivers/gpu/drm/i915/gt/gen8_ppgtt.c:641 > (inlined by) gen8_ppgtt_insert at drivers/gpu/drm/i915/gt/gen8_ppgtt.c:743 > > It's likely should be the i915 driver issue that is crashes with the NULL deref, but the origin of the bug should be the kvm page fault handling. > > David, could you please tell what tests you've run and post a link to yours kernel tree? Maybe I made obscure mistake while applied the patch manually. For tests, I ran the kvm selftests and then various ChromeOS virtualization tests. Two things to note about the ChromeOS virtualization tests are that they use crosvm instead of qemu, and they use virtio-gpu+virgl for graphics in the guest. I tested on an AMD device (since the goal of this series is to fix a compatibility issue with the amdgpu driver), and on a TGL device. I don't have an easy way to share my kernel tree, but it's based on v6.5-r3. The series I sent out is rebased onto the kvm next branch, but there were only minor conflicts. -David
On 9/19/23 05:31, David Stevens wrote: > On Mon, Sep 18, 2023 at 6:58 PM Dmitry Osipenko > <dmitry.osipenko@collabora.com> wrote: >> >> On 9/11/23 05:16, David Stevens wrote: >>> From: David Stevens <stevensd@chromium.org> >>> >>> Handle non-refcounted pages in __kvm_faultin_pfn. This allows the host >>> to map memory into the guest that is backed by non-refcounted struct >>> pages - for example, the tail pages of higher order non-compound pages >>> allocated by the amdgpu driver via ttm_pool_alloc_page. >>> >>> The bulk of this change is tracking the is_refcounted_page flag so that >>> non-refcounted pages don't trigger page_count() == 0 warnings. This is >>> done by storing the flag in an unused bit in the sptes. There are no >>> bits available in PAE SPTEs, so non-refcounted pages can only be handled >>> on TDP and x86-64. >>> >>> Signed-off-by: David Stevens <stevensd@chromium.org> >>> --- >>> arch/x86/kvm/mmu/mmu.c | 52 +++++++++++++++++++++++---------- >>> arch/x86/kvm/mmu/mmu_internal.h | 1 + >>> arch/x86/kvm/mmu/paging_tmpl.h | 8 +++-- >>> arch/x86/kvm/mmu/spte.c | 4 ++- >>> arch/x86/kvm/mmu/spte.h | 12 +++++++- >>> arch/x86/kvm/mmu/tdp_mmu.c | 22 ++++++++------ >>> include/linux/kvm_host.h | 3 ++ >>> virt/kvm/kvm_main.c | 6 ++-- >>> 8 files changed, 76 insertions(+), 32 deletions(-) >> >> Could you please tell which kernel tree you used for the base of this >> series? This patch #6 doesn't apply cleanly to stable/mainline/next/kvm >> >> error: sha1 information is lacking or useless (arch/x86/kvm/mmu/mmu.c). >> error: could not build fake ancestor > > This series is based on the kvm next branch (i.e. > https://git.kernel.org/pub/scm/virt/kvm/kvm.git/log/?h=next). The > specific hash is d011151616e73de20c139580b73fa4c7042bd861. Thanks, this tag works
On 9/19/23 05:59, David Stevens wrote: > On Mon, Sep 18, 2023 at 8:19 PM Dmitry Osipenko > <dmitry.osipenko@collabora.com> wrote: >> >> On 9/18/23 12:58, Dmitry Osipenko wrote: >>> On 9/11/23 05:16, David Stevens wrote: >>>> From: David Stevens <stevensd@chromium.org> >>>> >>>> Handle non-refcounted pages in __kvm_faultin_pfn. This allows the host >>>> to map memory into the guest that is backed by non-refcounted struct >>>> pages - for example, the tail pages of higher order non-compound pages >>>> allocated by the amdgpu driver via ttm_pool_alloc_page. >>>> >>>> The bulk of this change is tracking the is_refcounted_page flag so that >>>> non-refcounted pages don't trigger page_count() == 0 warnings. This is >>>> done by storing the flag in an unused bit in the sptes. There are no >>>> bits available in PAE SPTEs, so non-refcounted pages can only be handled >>>> on TDP and x86-64. >>>> >>>> Signed-off-by: David Stevens <stevensd@chromium.org> >>>> --- >>>> arch/x86/kvm/mmu/mmu.c | 52 +++++++++++++++++++++++---------- >>>> arch/x86/kvm/mmu/mmu_internal.h | 1 + >>>> arch/x86/kvm/mmu/paging_tmpl.h | 8 +++-- >>>> arch/x86/kvm/mmu/spte.c | 4 ++- >>>> arch/x86/kvm/mmu/spte.h | 12 +++++++- >>>> arch/x86/kvm/mmu/tdp_mmu.c | 22 ++++++++------ >>>> include/linux/kvm_host.h | 3 ++ >>>> virt/kvm/kvm_main.c | 6 ++-- >>>> 8 files changed, 76 insertions(+), 32 deletions(-) >>> >>> Could you please tell which kernel tree you used for the base of this >>> series? This patch #6 doesn't apply cleanly to stable/mainline/next/kvm >>> >>> error: sha1 information is lacking or useless (arch/x86/kvm/mmu/mmu.c). >>> error: could not build fake ancestor >> >> I applied the patch manually to v6.5.2 and tested Venus using Intel TGL iGPU, the intel driver is crashing: >> >> BUG: kernel NULL pointer dereference, address: 0000000000000058 >> #PF: supervisor read access in kernel mode >> #PF: error_code(0x0000) - not-present page >> PGD 0 P4D 0 >> Oops: 0000 [#1] PREEMPT SMP >> CPU: 1 PID: 5926 Comm: qemu-system-x86 Not tainted 6.5.2+ #114 >> Hardware name: LENOVO 20VE/LNVNB161216, BIOS F8CN43WW(V2.06) 08/12/2021 >> RIP: 0010:gen8_ppgtt_insert+0x50b/0x8f0 >> Code: 00 00 f7 c2 00 00 20 00 74 15 f7 c3 ff ff 1f 00 75 0d 41 81 fc ff ff 1f 00 0f 87 0e 02 00 00 48 8b 74 24 08 44 89 c0 45 85 ed <48> 8b 4e 58 48 8b 04 c1 0f 85 0b 02 00 00 81 e2 00 00 01 00 0f 84 >> RSP: 0018:ffffafc085afb820 EFLAGS: 00010246 >> RAX: 0000000000000000 RBX: 00000000e9604000 RCX: 000000000000001b >> RDX: 0000000000211000 RSI: 0000000000000000 RDI: ffff9513d44c1000 >> RBP: ffff951106f8dfc0 R08: 0000000000000000 R09: 0000000000000003 >> R10: 0000000000000fff R11: 00000000e9800000 R12: 00000000001fc000 >> R13: 0000000000000000 R14: 0000000000001000 R15: 0000ffff00000000 >> FS: 00007f2a5bcced80(0000) GS:ffff951a87a40000(0000) knlGS:0000000000000000 >> CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 >> CR2: 0000000000000058 CR3: 0000000116f16006 CR4: 0000000000772ee0 >> PKRU: 55555554 >> Call Trace: >> <TASK> >> ? __die+0x1f/0x60 >> ? page_fault_oops+0x14d/0x420 >> ? exc_page_fault+0x3d7/0x880 >> ? lock_acquire+0xc9/0x290 >> ? asm_exc_page_fault+0x22/0x30 >> ? gen8_ppgtt_insert+0x50b/0x8f0 >> ppgtt_bind_vma+0x4f/0x60 >> fence_work+0x1b/0x70 >> fence_notify+0x8f/0x130 >> __i915_sw_fence_complete+0x58/0x230 >> i915_vma_pin_ww+0x513/0xa80 >> eb_validate_vmas+0x17e/0x9e0 >> ? eb_pin_engine+0x2bb/0x340 >> i915_gem_do_execbuffer+0xc85/0x2bf0 >> ? __lock_acquire+0x3b6/0x21c0 >> i915_gem_execbuffer2_ioctl+0xee/0x240 >> ? i915_gem_do_execbuffer+0x2bf0/0x2bf0 >> drm_ioctl_kernel+0x9d/0x140 >> drm_ioctl+0x1dd/0x410 >> ? i915_gem_do_execbuffer+0x2bf0/0x2bf0 >> ? __fget_files+0xc5/0x170 >> __x64_sys_ioctl+0x8c/0xc0 >> do_syscall_64+0x34/0x80 >> entry_SYSCALL_64_after_hwframe+0x46/0xb0 >> RIP: 0033:0x7f2a60b0c9df >> >> >> $ ./scripts/faddr2line ./vmlinux gen8_ppgtt_insert+0x50b/0x8f0 >> gen8_ppgtt_insert+0x50b/0x8f0: >> i915_pt_entry at drivers/gpu/drm/i915/gt/intel_gtt.h:557 >> (inlined by) gen8_ppgtt_insert_huge at drivers/gpu/drm/i915/gt/gen8_ppgtt.c:641 >> (inlined by) gen8_ppgtt_insert at drivers/gpu/drm/i915/gt/gen8_ppgtt.c:743 >> >> It's likely should be the i915 driver issue that is crashes with the NULL deref, but the origin of the bug should be the kvm page fault handling. >> >> David, could you please tell what tests you've run and post a link to yours kernel tree? Maybe I made obscure mistake while applied the patch manually. > > For tests, I ran the kvm selftests and then various ChromeOS > virtualization tests. Two things to note about the ChromeOS > virtualization tests are that they use crosvm instead of qemu, and > they use virtio-gpu+virgl for graphics in the guest. I tested on an > AMD device (since the goal of this series is to fix a compatibility > issue with the amdgpu driver), and on a TGL device. > > I don't have an easy way to share my kernel tree, but it's based on > v6.5-r3. The series I sent out is rebased onto the kvm next branch, > but there were only minor conflicts. It's good that you mentioned crosvm because I only tested qemu. Interestingly, the problem doesn't present using crosvm. I made few more tests using these options: 1. yours latest version of the patch based on v6.5.4 2. yours latest version of the patch based on the kvm tree 3. by forward-porting yours older version of the kvm patchset to v6.5 that works well based on v6.4 kernel Neither of the options work with qemu and v6.5 kernel, but crosvm works. At first I felt confident that it must be i915 bug of the v6.5 kernel. But given that crosvm works, now I'm not sure. Will try to bisect 6.5 kernel, at minimum the i915 driver changes.
On 9/21/23 23:06, Dmitry Osipenko wrote: > On 9/19/23 05:59, David Stevens wrote: >> On Mon, Sep 18, 2023 at 8:19 PM Dmitry Osipenko >> <dmitry.osipenko@collabora.com> wrote: >>> >>> On 9/18/23 12:58, Dmitry Osipenko wrote: >>>> On 9/11/23 05:16, David Stevens wrote: >>>>> From: David Stevens <stevensd@chromium.org> >>>>> >>>>> Handle non-refcounted pages in __kvm_faultin_pfn. This allows the host >>>>> to map memory into the guest that is backed by non-refcounted struct >>>>> pages - for example, the tail pages of higher order non-compound pages >>>>> allocated by the amdgpu driver via ttm_pool_alloc_page. >>>>> >>>>> The bulk of this change is tracking the is_refcounted_page flag so that >>>>> non-refcounted pages don't trigger page_count() == 0 warnings. This is >>>>> done by storing the flag in an unused bit in the sptes. There are no >>>>> bits available in PAE SPTEs, so non-refcounted pages can only be handled >>>>> on TDP and x86-64. >>>>> >>>>> Signed-off-by: David Stevens <stevensd@chromium.org> >>>>> --- >>>>> arch/x86/kvm/mmu/mmu.c | 52 +++++++++++++++++++++++---------- >>>>> arch/x86/kvm/mmu/mmu_internal.h | 1 + >>>>> arch/x86/kvm/mmu/paging_tmpl.h | 8 +++-- >>>>> arch/x86/kvm/mmu/spte.c | 4 ++- >>>>> arch/x86/kvm/mmu/spte.h | 12 +++++++- >>>>> arch/x86/kvm/mmu/tdp_mmu.c | 22 ++++++++------ >>>>> include/linux/kvm_host.h | 3 ++ >>>>> virt/kvm/kvm_main.c | 6 ++-- >>>>> 8 files changed, 76 insertions(+), 32 deletions(-) >>>> >>>> Could you please tell which kernel tree you used for the base of this >>>> series? This patch #6 doesn't apply cleanly to stable/mainline/next/kvm >>>> >>>> error: sha1 information is lacking or useless (arch/x86/kvm/mmu/mmu.c). >>>> error: could not build fake ancestor >>> >>> I applied the patch manually to v6.5.2 and tested Venus using Intel TGL iGPU, the intel driver is crashing: >>> >>> BUG: kernel NULL pointer dereference, address: 0000000000000058 >>> #PF: supervisor read access in kernel mode >>> #PF: error_code(0x0000) - not-present page >>> PGD 0 P4D 0 >>> Oops: 0000 [#1] PREEMPT SMP >>> CPU: 1 PID: 5926 Comm: qemu-system-x86 Not tainted 6.5.2+ #114 >>> Hardware name: LENOVO 20VE/LNVNB161216, BIOS F8CN43WW(V2.06) 08/12/2021 >>> RIP: 0010:gen8_ppgtt_insert+0x50b/0x8f0 >>> Code: 00 00 f7 c2 00 00 20 00 74 15 f7 c3 ff ff 1f 00 75 0d 41 81 fc ff ff 1f 00 0f 87 0e 02 00 00 48 8b 74 24 08 44 89 c0 45 85 ed <48> 8b 4e 58 48 8b 04 c1 0f 85 0b 02 00 00 81 e2 00 00 01 00 0f 84 >>> RSP: 0018:ffffafc085afb820 EFLAGS: 00010246 >>> RAX: 0000000000000000 RBX: 00000000e9604000 RCX: 000000000000001b >>> RDX: 0000000000211000 RSI: 0000000000000000 RDI: ffff9513d44c1000 >>> RBP: ffff951106f8dfc0 R08: 0000000000000000 R09: 0000000000000003 >>> R10: 0000000000000fff R11: 00000000e9800000 R12: 00000000001fc000 >>> R13: 0000000000000000 R14: 0000000000001000 R15: 0000ffff00000000 >>> FS: 00007f2a5bcced80(0000) GS:ffff951a87a40000(0000) knlGS:0000000000000000 >>> CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 >>> CR2: 0000000000000058 CR3: 0000000116f16006 CR4: 0000000000772ee0 >>> PKRU: 55555554 >>> Call Trace: >>> <TASK> >>> ? __die+0x1f/0x60 >>> ? page_fault_oops+0x14d/0x420 >>> ? exc_page_fault+0x3d7/0x880 >>> ? lock_acquire+0xc9/0x290 >>> ? asm_exc_page_fault+0x22/0x30 >>> ? gen8_ppgtt_insert+0x50b/0x8f0 >>> ppgtt_bind_vma+0x4f/0x60 >>> fence_work+0x1b/0x70 >>> fence_notify+0x8f/0x130 >>> __i915_sw_fence_complete+0x58/0x230 >>> i915_vma_pin_ww+0x513/0xa80 >>> eb_validate_vmas+0x17e/0x9e0 >>> ? eb_pin_engine+0x2bb/0x340 >>> i915_gem_do_execbuffer+0xc85/0x2bf0 >>> ? __lock_acquire+0x3b6/0x21c0 >>> i915_gem_execbuffer2_ioctl+0xee/0x240 >>> ? i915_gem_do_execbuffer+0x2bf0/0x2bf0 >>> drm_ioctl_kernel+0x9d/0x140 >>> drm_ioctl+0x1dd/0x410 >>> ? i915_gem_do_execbuffer+0x2bf0/0x2bf0 >>> ? __fget_files+0xc5/0x170 >>> __x64_sys_ioctl+0x8c/0xc0 >>> do_syscall_64+0x34/0x80 >>> entry_SYSCALL_64_after_hwframe+0x46/0xb0 >>> RIP: 0033:0x7f2a60b0c9df >>> >>> >>> $ ./scripts/faddr2line ./vmlinux gen8_ppgtt_insert+0x50b/0x8f0 >>> gen8_ppgtt_insert+0x50b/0x8f0: >>> i915_pt_entry at drivers/gpu/drm/i915/gt/intel_gtt.h:557 >>> (inlined by) gen8_ppgtt_insert_huge at drivers/gpu/drm/i915/gt/gen8_ppgtt.c:641 >>> (inlined by) gen8_ppgtt_insert at drivers/gpu/drm/i915/gt/gen8_ppgtt.c:743 >>> >>> It's likely should be the i915 driver issue that is crashes with the NULL deref, but the origin of the bug should be the kvm page fault handling. >>> >>> David, could you please tell what tests you've run and post a link to yours kernel tree? Maybe I made obscure mistake while applied the patch manually. >> >> For tests, I ran the kvm selftests and then various ChromeOS >> virtualization tests. Two things to note about the ChromeOS >> virtualization tests are that they use crosvm instead of qemu, and >> they use virtio-gpu+virgl for graphics in the guest. I tested on an >> AMD device (since the goal of this series is to fix a compatibility >> issue with the amdgpu driver), and on a TGL device. >> >> I don't have an easy way to share my kernel tree, but it's based on >> v6.5-r3. The series I sent out is rebased onto the kvm next branch, >> but there were only minor conflicts. > > It's good that you mentioned crosvm because I only tested qemu. > Interestingly, the problem doesn't present using crosvm. > > I made few more tests using these options: > > 1. yours latest version of the patch based on v6.5.4 > 2. yours latest version of the patch based on the kvm tree > 3. by forward-porting yours older version of the kvm patchset to v6.5 > that works well based on v6.4 kernel > > Neither of the options work with qemu and v6.5 kernel, but crosvm works. > > At first I felt confident that it must be i915 bug of the v6.5 kernel. > But given that crosvm works, now I'm not sure. > > Will try to bisect 6.5 kernel, at minimum the i915 driver changes. The bisection said that it's not i915 bug. Good news, the bug was fixed in a recent linux-next. Don't know what is the exact fix, suppose it's something in mm/
У пн, 2023-09-11 у 11:16 +0900, David Stevens пише: > From: David Stevens <stevensd@chromium.org> > > Handle non-refcounted pages in __kvm_faultin_pfn. This allows the host > to map memory into the guest that is backed by non-refcounted struct > pages - for example, the tail pages of higher order non-compound pages > allocated by the amdgpu driver via ttm_pool_alloc_page. > > The bulk of this change is tracking the is_refcounted_page flag so that > non-refcounted pages don't trigger page_count() == 0 warnings. This is > done by storing the flag in an unused bit in the sptes. There are no > bits available in PAE SPTEs, so non-refcounted pages can only be handled > on TDP and x86-64. > > Signed-off-by: David Stevens <stevensd@chromium.org> > --- > arch/x86/kvm/mmu/mmu.c | 52 +++++++++++++++++++++++---------- > arch/x86/kvm/mmu/mmu_internal.h | 1 + > arch/x86/kvm/mmu/paging_tmpl.h | 8 +++-- > arch/x86/kvm/mmu/spte.c | 4 ++- > arch/x86/kvm/mmu/spte.h | 12 +++++++- > arch/x86/kvm/mmu/tdp_mmu.c | 22 ++++++++------ > include/linux/kvm_host.h | 3 ++ > virt/kvm/kvm_main.c | 6 ++-- > 8 files changed, 76 insertions(+), 32 deletions(-) > > diff --git a/arch/x86/kvm/mmu/mmu.c b/arch/x86/kvm/mmu/mmu.c > index e1eca26215e2..b8168cc4cc96 100644 > --- a/arch/x86/kvm/mmu/mmu.c > +++ b/arch/x86/kvm/mmu/mmu.c > @@ -545,12 +545,14 @@ static bool mmu_spte_update(u64 *sptep, u64 new_spte) > > if (is_accessed_spte(old_spte) && !is_accessed_spte(new_spte)) { > flush = true; > - kvm_set_pfn_accessed(spte_to_pfn(old_spte)); > + if (is_refcounted_page_pte(old_spte)) > + kvm_set_page_accessed(pfn_to_page(spte_to_pfn(old_spte))); > } > > if (is_dirty_spte(old_spte) && !is_dirty_spte(new_spte)) { > flush = true; > - kvm_set_pfn_dirty(spte_to_pfn(old_spte)); > + if (is_refcounted_page_pte(old_spte)) > + kvm_set_page_dirty(pfn_to_page(spte_to_pfn(old_spte))); > } > > return flush; > @@ -588,14 +590,18 @@ static u64 mmu_spte_clear_track_bits(struct kvm *kvm, u64 *sptep) > * before they are reclaimed. Sanity check that, if the pfn is backed > * by a refcounted page, the refcount is elevated. > */ > - page = kvm_pfn_to_refcounted_page(pfn); > - WARN_ON_ONCE(page && !page_count(page)); > + if (is_refcounted_page_pte(old_spte)) { > + page = kvm_pfn_to_refcounted_page(pfn); > + WARN_ON_ONCE(!page || !page_count(page)); > + } > > - if (is_accessed_spte(old_spte)) > - kvm_set_pfn_accessed(pfn); > + if (is_refcounted_page_pte(old_spte)) { > + if (is_accessed_spte(old_spte)) > + kvm_set_page_accessed(pfn_to_page(pfn)); > > - if (is_dirty_spte(old_spte)) > - kvm_set_pfn_dirty(pfn); > + if (is_dirty_spte(old_spte)) > + kvm_set_page_dirty(pfn_to_page(pfn)); > + } > > return old_spte; > } > @@ -631,8 +637,8 @@ static bool mmu_spte_age(u64 *sptep) > * Capture the dirty status of the page, so that it doesn't get > * lost when the SPTE is marked for access tracking. > */ > - if (is_writable_pte(spte)) > - kvm_set_pfn_dirty(spte_to_pfn(spte)); > + if (is_writable_pte(spte) && is_refcounted_page_pte(spte)) > + kvm_set_page_dirty(pfn_to_page(spte_to_pfn(spte))); > > spte = mark_spte_for_access_track(spte); > mmu_spte_update_no_track(sptep, spte); > @@ -1261,8 +1267,8 @@ static bool spte_wrprot_for_clear_dirty(u64 *sptep) > { > bool was_writable = test_and_clear_bit(PT_WRITABLE_SHIFT, > (unsigned long *)sptep); > - if (was_writable && !spte_ad_enabled(*sptep)) > - kvm_set_pfn_dirty(spte_to_pfn(*sptep)); > + if (was_writable && !spte_ad_enabled(*sptep) && is_refcounted_page_pte(*sptep)) > + kvm_set_page_dirty(pfn_to_page(spte_to_pfn(*sptep))); > > return was_writable; > } > @@ -2913,6 +2919,11 @@ static int mmu_set_spte(struct kvm_vcpu *vcpu, struct kvm_memory_slot *slot, > bool host_writable = !fault || fault->map_writable; > bool prefetch = !fault || fault->prefetch; > bool write_fault = fault && fault->write; > + /* > + * Prefetching uses gfn_to_page_many_atomic, which never gets > + * non-refcounted pages. > + */ > + bool is_refcounted = !fault || fault->is_refcounted_page; A WARN_ON_ONCE for a future bug that will make this condition false might be a good idea. > > if (unlikely(is_noslot_pfn(pfn))) { > vcpu->stat.pf_mmio_spte_created++; > @@ -2940,7 +2951,7 @@ static int mmu_set_spte(struct kvm_vcpu *vcpu, struct kvm_memory_slot *slot, > } > > wrprot = make_spte(vcpu, sp, slot, pte_access, gfn, pfn, *sptep, prefetch, > - true, host_writable, &spte); > + true, host_writable, is_refcounted, &spte); > > if (*sptep == spte) { > ret = RET_PF_SPURIOUS; > @@ -4254,13 +4265,18 @@ void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu, struct kvm_async_pf *work) > static int __kvm_faultin_pfn(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault) > { > struct kvm_memory_slot *slot = fault->slot; > + /* > + * There are no extra bits for tracking non-refcounted pages in > + * PAE SPTEs, so reject non-refcounted struct pages in that case. > + */ > + bool has_spte_refcount_bit = tdp_enabled && IS_ENABLED(CONFIG_X86_64); I think that the tdp_enabled condition is needed because shadow paging uses the same paging mode as the guest and it can use PAE, thus there will be no reserved bits. Is this true? If true, can you write a comment about this? I haven't worked with shadow paging for a long time, I no longer remember some of the details. > struct kvm_follow_pfn foll = { > .slot = slot, > .gfn = fault->gfn, > .flags = fault->write ? FOLL_WRITE : 0, > .try_map_writable = true, > .guarded_by_mmu_notifier = true, > - .allow_non_refcounted_struct_page = false, > + .allow_non_refcounted_struct_page = has_spte_refcount_bit, > }; > > /* > @@ -4277,6 +4293,7 @@ static int __kvm_faultin_pfn(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault > fault->slot = NULL; > fault->pfn = KVM_PFN_NOSLOT; > fault->map_writable = false; > + fault->is_refcounted_page = false; > return RET_PF_CONTINUE; > } > /* > @@ -4332,6 +4349,7 @@ static int __kvm_faultin_pfn(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault > success: > fault->hva = foll.hva; > fault->map_writable = foll.writable; > + fault->is_refcounted_page = foll.is_refcounted_page; > return RET_PF_CONTINUE; > } > > @@ -4420,8 +4438,9 @@ static int direct_page_fault(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault > r = direct_map(vcpu, fault); > > out_unlock: > + if (fault->is_refcounted_page) > + kvm_set_page_accessed(pfn_to_page(fault->pfn)); > write_unlock(&vcpu->kvm->mmu_lock); > - kvm_release_pfn_clean(fault->pfn); > return r; > } > > @@ -4496,8 +4515,9 @@ static int kvm_tdp_mmu_page_fault(struct kvm_vcpu *vcpu, > r = kvm_tdp_mmu_map(vcpu, fault); > > out_unlock: > + if (fault->is_refcounted_page) > + kvm_set_page_accessed(pfn_to_page(fault->pfn)); > read_unlock(&vcpu->kvm->mmu_lock); > - kvm_release_pfn_clean(fault->pfn); > return r; > } > #endif > diff --git a/arch/x86/kvm/mmu/mmu_internal.h b/arch/x86/kvm/mmu/mmu_internal.h > index b102014e2c60..7f73bc2a552e 100644 > --- a/arch/x86/kvm/mmu/mmu_internal.h > +++ b/arch/x86/kvm/mmu/mmu_internal.h > @@ -239,6 +239,7 @@ struct kvm_page_fault { > kvm_pfn_t pfn; > hva_t hva; > bool map_writable; > + bool is_refcounted_page; > > /* > * Indicates the guest is trying to write a gfn that contains one or > diff --git a/arch/x86/kvm/mmu/paging_tmpl.h b/arch/x86/kvm/mmu/paging_tmpl.h > index c85255073f67..0ac4a4e5870c 100644 > --- a/arch/x86/kvm/mmu/paging_tmpl.h > +++ b/arch/x86/kvm/mmu/paging_tmpl.h > @@ -848,7 +848,8 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault > > out_unlock: > write_unlock(&vcpu->kvm->mmu_lock); > - kvm_release_pfn_clean(fault->pfn); > + if (fault->is_refcounted_page) > + kvm_set_page_accessed(pfn_to_page(fault->pfn)); > return r; > } > > @@ -902,7 +903,7 @@ static gpa_t FNAME(gva_to_gpa)(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, > */ > static int FNAME(sync_spte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, int i) > { > - bool host_writable; > + bool host_writable, is_refcounted; > gpa_t first_pte_gpa; > u64 *sptep, spte; > struct kvm_memory_slot *slot; > @@ -959,10 +960,11 @@ static int FNAME(sync_spte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, int > sptep = &sp->spt[i]; > spte = *sptep; > host_writable = spte & shadow_host_writable_mask; > + is_refcounted = spte & SPTE_MMU_PAGE_REFCOUNTED; What will happen if this function is run on 32 bit kernel and/or without tdp enabled (that is when SPTE_MMU_PAGE_REFCOUNTED is not defined)? > slot = kvm_vcpu_gfn_to_memslot(vcpu, gfn); > make_spte(vcpu, sp, slot, pte_access, gfn, > spte_to_pfn(spte), spte, true, false, > - host_writable, &spte); > + host_writable, is_refcounted, &spte); > > return mmu_spte_update(sptep, spte); > } > diff --git a/arch/x86/kvm/mmu/spte.c b/arch/x86/kvm/mmu/spte.c > index 4a599130e9c9..ce495819061f 100644 > --- a/arch/x86/kvm/mmu/spte.c > +++ b/arch/x86/kvm/mmu/spte.c > @@ -138,7 +138,7 @@ bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, > const struct kvm_memory_slot *slot, > unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, > u64 old_spte, bool prefetch, bool can_unsync, > - bool host_writable, u64 *new_spte) > + bool host_writable, bool is_refcounted, u64 *new_spte) > { > int level = sp->role.level; > u64 spte = SPTE_MMU_PRESENT_MASK; > @@ -188,6 +188,8 @@ bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, > > if (level > PG_LEVEL_4K) > spte |= PT_PAGE_SIZE_MASK; > + if (is_refcounted) > + spte |= SPTE_MMU_PAGE_REFCOUNTED; Same here, if make_spte is used in these modes won't it set a bit it shouldn't? > > if (shadow_memtype_mask) > spte |= static_call(kvm_x86_get_mt_mask)(vcpu, gfn, > diff --git a/arch/x86/kvm/mmu/spte.h b/arch/x86/kvm/mmu/spte.h > index a129951c9a88..4bf4a535c23d 100644 > --- a/arch/x86/kvm/mmu/spte.h > +++ b/arch/x86/kvm/mmu/spte.h > @@ -96,6 +96,11 @@ static_assert(!(EPT_SPTE_MMU_WRITABLE & SHADOW_ACC_TRACK_SAVED_MASK)); > /* Defined only to keep the above static asserts readable. */ > #undef SHADOW_ACC_TRACK_SAVED_MASK > > +/* > + * Indicates that the SPTE refers to a page with a valid refcount. > + */ > +#define SPTE_MMU_PAGE_REFCOUNTED BIT_ULL(59) > + > /* > * Due to limited space in PTEs, the MMIO generation is a 19 bit subset of > * the memslots generation and is derived as follows: > @@ -345,6 +350,11 @@ static inline bool is_dirty_spte(u64 spte) > return dirty_mask ? spte & dirty_mask : spte & PT_WRITABLE_MASK; > } > > +static inline bool is_refcounted_page_pte(u64 spte) > +{ > + return spte & SPTE_MMU_PAGE_REFCOUNTED; Here also: if the bit is not supported, we need to assume that all sptes are refcounted I think. > +} > + > static inline u64 get_rsvd_bits(struct rsvd_bits_validate *rsvd_check, u64 pte, > int level) > { > @@ -475,7 +485,7 @@ bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, > const struct kvm_memory_slot *slot, > unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, > u64 old_spte, bool prefetch, bool can_unsync, > - bool host_writable, u64 *new_spte); > + bool host_writable, bool is_refcounted, u64 *new_spte); > u64 make_huge_page_split_spte(struct kvm *kvm, u64 huge_spte, > union kvm_mmu_page_role role, int index); > u64 make_nonleaf_spte(u64 *child_pt, bool ad_disabled); > diff --git a/arch/x86/kvm/mmu/tdp_mmu.c b/arch/x86/kvm/mmu/tdp_mmu.c > index 6c63f2d1675f..185f3c666c2b 100644 > --- a/arch/x86/kvm/mmu/tdp_mmu.c > +++ b/arch/x86/kvm/mmu/tdp_mmu.c > @@ -474,6 +474,7 @@ static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, > bool was_leaf = was_present && is_last_spte(old_spte, level); > bool is_leaf = is_present && is_last_spte(new_spte, level); > bool pfn_changed = spte_to_pfn(old_spte) != spte_to_pfn(new_spte); > + bool is_refcounted = is_refcounted_page_pte(old_spte); > > WARN_ON_ONCE(level > PT64_ROOT_MAX_LEVEL); > WARN_ON_ONCE(level < PG_LEVEL_4K); > @@ -538,9 +539,9 @@ static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, > if (is_leaf != was_leaf) > kvm_update_page_stats(kvm, level, is_leaf ? 1 : -1); > > - if (was_leaf && is_dirty_spte(old_spte) && > + if (was_leaf && is_dirty_spte(old_spte) && is_refcounted && > (!is_present || !is_dirty_spte(new_spte) || pfn_changed)) > - kvm_set_pfn_dirty(spte_to_pfn(old_spte)); > + kvm_set_page_dirty(pfn_to_page(spte_to_pfn(old_spte))); > > /* > * Recursively handle child PTs if the change removed a subtree from > @@ -552,9 +553,9 @@ static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, > (is_leaf || !is_present || WARN_ON_ONCE(pfn_changed))) > handle_removed_pt(kvm, spte_to_child_pt(old_spte, level), shared); > > - if (was_leaf && is_accessed_spte(old_spte) && > + if (was_leaf && is_accessed_spte(old_spte) && is_refcounted && > (!is_present || !is_accessed_spte(new_spte) || pfn_changed)) > - kvm_set_pfn_accessed(spte_to_pfn(old_spte)); > + kvm_set_page_accessed(pfn_to_page(spte_to_pfn(old_spte))); > } > > /* > @@ -988,8 +989,9 @@ static int tdp_mmu_map_handle_target_level(struct kvm_vcpu *vcpu, > new_spte = make_mmio_spte(vcpu, iter->gfn, ACC_ALL); > else > wrprot = make_spte(vcpu, sp, fault->slot, ACC_ALL, iter->gfn, > - fault->pfn, iter->old_spte, fault->prefetch, true, > - fault->map_writable, &new_spte); > + fault->pfn, iter->old_spte, fault->prefetch, true, > + fault->map_writable, fault->is_refcounted_page, > + &new_spte); > > if (new_spte == iter->old_spte) > ret = RET_PF_SPURIOUS; > @@ -1205,8 +1207,9 @@ static bool age_gfn_range(struct kvm *kvm, struct tdp_iter *iter, > * Capture the dirty status of the page, so that it doesn't get > * lost when the SPTE is marked for access tracking. > */ > - if (is_writable_pte(iter->old_spte)) > - kvm_set_pfn_dirty(spte_to_pfn(iter->old_spte)); > + if (is_writable_pte(iter->old_spte) && > + is_refcounted_page_pte(iter->old_spte)) > + kvm_set_page_dirty(pfn_to_page(spte_to_pfn(iter->old_spte))); > > new_spte = mark_spte_for_access_track(iter->old_spte); > iter->old_spte = kvm_tdp_mmu_write_spte(iter->sptep, > @@ -1628,7 +1631,8 @@ static void clear_dirty_pt_masked(struct kvm *kvm, struct kvm_mmu_page *root, > trace_kvm_tdp_mmu_spte_changed(iter.as_id, iter.gfn, iter.level, > iter.old_spte, > iter.old_spte & ~dbit); > - kvm_set_pfn_dirty(spte_to_pfn(iter.old_spte)); > + if (is_refcounted_page_pte(iter.old_spte)) > + kvm_set_page_dirty(pfn_to_page(spte_to_pfn(iter.old_spte))); > } > > rcu_read_unlock(); > diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h > index b95c79b7833b..6696925f01f1 100644 > --- a/include/linux/kvm_host.h > +++ b/include/linux/kvm_host.h > @@ -1179,6 +1179,9 @@ unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn, > void kvm_release_page_clean(struct page *page); > void kvm_release_page_dirty(struct page *page); > > +void kvm_set_page_accessed(struct page *page); > +void kvm_set_page_dirty(struct page *page); > + > struct kvm_follow_pfn { > const struct kvm_memory_slot *slot; > gfn_t gfn; > diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c > index 913de4e86d9d..4d8538cdb690 100644 > --- a/virt/kvm/kvm_main.c > +++ b/virt/kvm/kvm_main.c > @@ -2979,17 +2979,19 @@ static bool kvm_is_ad_tracked_page(struct page *page) > return !PageReserved(page); > } > > -static void kvm_set_page_dirty(struct page *page) > +void kvm_set_page_dirty(struct page *page) > { > if (kvm_is_ad_tracked_page(page)) > SetPageDirty(page); > } > +EXPORT_SYMBOL_GPL(kvm_set_page_dirty); > > -static void kvm_set_page_accessed(struct page *page) > +void kvm_set_page_accessed(struct page *page) > { > if (kvm_is_ad_tracked_page(page)) > mark_page_accessed(page); > } > +EXPORT_SYMBOL_GPL(kvm_set_page_accessed); > > void kvm_release_page_clean(struct page *page) > { Best regards, Maxim Levitsky
On Tue, Sep 19, 2023, David Stevens wrote: > On Mon, Sep 18, 2023 at 6:58 PM Dmitry Osipenko > <dmitry.osipenko@collabora.com> wrote: > > > > On 9/11/23 05:16, David Stevens wrote: > > > From: David Stevens <stevensd@chromium.org> > > > > > > Handle non-refcounted pages in __kvm_faultin_pfn. This allows the host > > > to map memory into the guest that is backed by non-refcounted struct > > > pages - for example, the tail pages of higher order non-compound pages > > > allocated by the amdgpu driver via ttm_pool_alloc_page. > > > > > > The bulk of this change is tracking the is_refcounted_page flag so that > > > non-refcounted pages don't trigger page_count() == 0 warnings. This is > > > done by storing the flag in an unused bit in the sptes. There are no > > > bits available in PAE SPTEs, so non-refcounted pages can only be handled > > > on TDP and x86-64. > > > > > > Signed-off-by: David Stevens <stevensd@chromium.org> > > > --- > > > arch/x86/kvm/mmu/mmu.c | 52 +++++++++++++++++++++++---------- > > > arch/x86/kvm/mmu/mmu_internal.h | 1 + > > > arch/x86/kvm/mmu/paging_tmpl.h | 8 +++-- > > > arch/x86/kvm/mmu/spte.c | 4 ++- > > > arch/x86/kvm/mmu/spte.h | 12 +++++++- > > > arch/x86/kvm/mmu/tdp_mmu.c | 22 ++++++++------ > > > include/linux/kvm_host.h | 3 ++ > > > virt/kvm/kvm_main.c | 6 ++-- > > > 8 files changed, 76 insertions(+), 32 deletions(-) > > > > Could you please tell which kernel tree you used for the base of this > > series? This patch #6 doesn't apply cleanly to stable/mainline/next/kvm > > > > error: sha1 information is lacking or useless (arch/x86/kvm/mmu/mmu.c). > > error: could not build fake ancestor > > This series is based on the kvm next branch (i.e. > https://git.kernel.org/pub/scm/virt/kvm/kvm.git/log/?h=next). The > specific hash is d011151616e73de20c139580b73fa4c7042bd861. Drat, found this too late. Please use --base so that git appends the exact base commit. From Documentation/process/maintainer-kvm-x86.rst: Git Base ~~~~~~~~ If you are using git version 2.9.0 or later (Googlers, this is all of you!), use ``git format-patch`` with the ``--base`` flag to automatically include the base tree information in the generated patches. Note, ``--base=auto`` works as expected if and only if a branch's upstream is set to the base topic branch, e.g. it will do the wrong thing if your upstream is set to your personal repository for backup purposes. An alternative "auto" solution is to derive the names of your development branches based on their KVM x86 topic, and feed that into ``--base``. E.g. ``x86/pmu/my_branch_name``, and then write a small wrapper to extract ``pmu`` from the current branch name to yield ``--base=x/pmu``, where ``x`` is whatever name your repository uses to track the KVM x86 remote.
On Mon, Sep 11, 2023, David Stevens wrote: > From: David Stevens <stevensd@chromium.org> > > Handle non-refcounted pages in __kvm_faultin_pfn. This allows the host > to map memory into the guest that is backed by non-refcounted struct > pages - for example, the tail pages of higher order non-compound pages > allocated by the amdgpu driver via ttm_pool_alloc_page. > > The bulk of this change is tracking the is_refcounted_page flag so that > non-refcounted pages don't trigger page_count() == 0 warnings. This is > done by storing the flag in an unused bit in the sptes. There are no > bits available in PAE SPTEs, so non-refcounted pages can only be handled > on TDP and x86-64. Can you split this into two patches? One to add all of the SPTE tracking, and then one final patch to allow faulting in non-refcounted pages. I want to isolate the latter as much as possible, both for review purposes and in case something goes awry and needs to be reverted. > @@ -4254,13 +4265,18 @@ void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu, struct kvm_async_pf *work) > static int __kvm_faultin_pfn(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault) > { > struct kvm_memory_slot *slot = fault->slot; > + /* > + * There are no extra bits for tracking non-refcounted pages in > + * PAE SPTEs, so reject non-refcounted struct pages in that case. > + */ > + bool has_spte_refcount_bit = tdp_enabled && IS_ENABLED(CONFIG_X86_64); Eh, just drop the local variable and do .allow_non_refcounted_struct_page = tdp_enabled && IS_ENABLED(CONFIG_X86_64); (but keep the comment)
diff --git a/arch/x86/kvm/mmu/mmu.c b/arch/x86/kvm/mmu/mmu.c index e1eca26215e2..b8168cc4cc96 100644 --- a/arch/x86/kvm/mmu/mmu.c +++ b/arch/x86/kvm/mmu/mmu.c @@ -545,12 +545,14 @@ static bool mmu_spte_update(u64 *sptep, u64 new_spte) if (is_accessed_spte(old_spte) && !is_accessed_spte(new_spte)) { flush = true; - kvm_set_pfn_accessed(spte_to_pfn(old_spte)); + if (is_refcounted_page_pte(old_spte)) + kvm_set_page_accessed(pfn_to_page(spte_to_pfn(old_spte))); } if (is_dirty_spte(old_spte) && !is_dirty_spte(new_spte)) { flush = true; - kvm_set_pfn_dirty(spte_to_pfn(old_spte)); + if (is_refcounted_page_pte(old_spte)) + kvm_set_page_dirty(pfn_to_page(spte_to_pfn(old_spte))); } return flush; @@ -588,14 +590,18 @@ static u64 mmu_spte_clear_track_bits(struct kvm *kvm, u64 *sptep) * before they are reclaimed. Sanity check that, if the pfn is backed * by a refcounted page, the refcount is elevated. */ - page = kvm_pfn_to_refcounted_page(pfn); - WARN_ON_ONCE(page && !page_count(page)); + if (is_refcounted_page_pte(old_spte)) { + page = kvm_pfn_to_refcounted_page(pfn); + WARN_ON_ONCE(!page || !page_count(page)); + } - if (is_accessed_spte(old_spte)) - kvm_set_pfn_accessed(pfn); + if (is_refcounted_page_pte(old_spte)) { + if (is_accessed_spte(old_spte)) + kvm_set_page_accessed(pfn_to_page(pfn)); - if (is_dirty_spte(old_spte)) - kvm_set_pfn_dirty(pfn); + if (is_dirty_spte(old_spte)) + kvm_set_page_dirty(pfn_to_page(pfn)); + } return old_spte; } @@ -631,8 +637,8 @@ static bool mmu_spte_age(u64 *sptep) * Capture the dirty status of the page, so that it doesn't get * lost when the SPTE is marked for access tracking. */ - if (is_writable_pte(spte)) - kvm_set_pfn_dirty(spte_to_pfn(spte)); + if (is_writable_pte(spte) && is_refcounted_page_pte(spte)) + kvm_set_page_dirty(pfn_to_page(spte_to_pfn(spte))); spte = mark_spte_for_access_track(spte); mmu_spte_update_no_track(sptep, spte); @@ -1261,8 +1267,8 @@ static bool spte_wrprot_for_clear_dirty(u64 *sptep) { bool was_writable = test_and_clear_bit(PT_WRITABLE_SHIFT, (unsigned long *)sptep); - if (was_writable && !spte_ad_enabled(*sptep)) - kvm_set_pfn_dirty(spte_to_pfn(*sptep)); + if (was_writable && !spte_ad_enabled(*sptep) && is_refcounted_page_pte(*sptep)) + kvm_set_page_dirty(pfn_to_page(spte_to_pfn(*sptep))); return was_writable; } @@ -2913,6 +2919,11 @@ static int mmu_set_spte(struct kvm_vcpu *vcpu, struct kvm_memory_slot *slot, bool host_writable = !fault || fault->map_writable; bool prefetch = !fault || fault->prefetch; bool write_fault = fault && fault->write; + /* + * Prefetching uses gfn_to_page_many_atomic, which never gets + * non-refcounted pages. + */ + bool is_refcounted = !fault || fault->is_refcounted_page; if (unlikely(is_noslot_pfn(pfn))) { vcpu->stat.pf_mmio_spte_created++; @@ -2940,7 +2951,7 @@ static int mmu_set_spte(struct kvm_vcpu *vcpu, struct kvm_memory_slot *slot, } wrprot = make_spte(vcpu, sp, slot, pte_access, gfn, pfn, *sptep, prefetch, - true, host_writable, &spte); + true, host_writable, is_refcounted, &spte); if (*sptep == spte) { ret = RET_PF_SPURIOUS; @@ -4254,13 +4265,18 @@ void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu, struct kvm_async_pf *work) static int __kvm_faultin_pfn(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault) { struct kvm_memory_slot *slot = fault->slot; + /* + * There are no extra bits for tracking non-refcounted pages in + * PAE SPTEs, so reject non-refcounted struct pages in that case. + */ + bool has_spte_refcount_bit = tdp_enabled && IS_ENABLED(CONFIG_X86_64); struct kvm_follow_pfn foll = { .slot = slot, .gfn = fault->gfn, .flags = fault->write ? FOLL_WRITE : 0, .try_map_writable = true, .guarded_by_mmu_notifier = true, - .allow_non_refcounted_struct_page = false, + .allow_non_refcounted_struct_page = has_spte_refcount_bit, }; /* @@ -4277,6 +4293,7 @@ static int __kvm_faultin_pfn(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault fault->slot = NULL; fault->pfn = KVM_PFN_NOSLOT; fault->map_writable = false; + fault->is_refcounted_page = false; return RET_PF_CONTINUE; } /* @@ -4332,6 +4349,7 @@ static int __kvm_faultin_pfn(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault success: fault->hva = foll.hva; fault->map_writable = foll.writable; + fault->is_refcounted_page = foll.is_refcounted_page; return RET_PF_CONTINUE; } @@ -4420,8 +4438,9 @@ static int direct_page_fault(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault r = direct_map(vcpu, fault); out_unlock: + if (fault->is_refcounted_page) + kvm_set_page_accessed(pfn_to_page(fault->pfn)); write_unlock(&vcpu->kvm->mmu_lock); - kvm_release_pfn_clean(fault->pfn); return r; } @@ -4496,8 +4515,9 @@ static int kvm_tdp_mmu_page_fault(struct kvm_vcpu *vcpu, r = kvm_tdp_mmu_map(vcpu, fault); out_unlock: + if (fault->is_refcounted_page) + kvm_set_page_accessed(pfn_to_page(fault->pfn)); read_unlock(&vcpu->kvm->mmu_lock); - kvm_release_pfn_clean(fault->pfn); return r; } #endif diff --git a/arch/x86/kvm/mmu/mmu_internal.h b/arch/x86/kvm/mmu/mmu_internal.h index b102014e2c60..7f73bc2a552e 100644 --- a/arch/x86/kvm/mmu/mmu_internal.h +++ b/arch/x86/kvm/mmu/mmu_internal.h @@ -239,6 +239,7 @@ struct kvm_page_fault { kvm_pfn_t pfn; hva_t hva; bool map_writable; + bool is_refcounted_page; /* * Indicates the guest is trying to write a gfn that contains one or diff --git a/arch/x86/kvm/mmu/paging_tmpl.h b/arch/x86/kvm/mmu/paging_tmpl.h index c85255073f67..0ac4a4e5870c 100644 --- a/arch/x86/kvm/mmu/paging_tmpl.h +++ b/arch/x86/kvm/mmu/paging_tmpl.h @@ -848,7 +848,8 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault out_unlock: write_unlock(&vcpu->kvm->mmu_lock); - kvm_release_pfn_clean(fault->pfn); + if (fault->is_refcounted_page) + kvm_set_page_accessed(pfn_to_page(fault->pfn)); return r; } @@ -902,7 +903,7 @@ static gpa_t FNAME(gva_to_gpa)(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, */ static int FNAME(sync_spte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, int i) { - bool host_writable; + bool host_writable, is_refcounted; gpa_t first_pte_gpa; u64 *sptep, spte; struct kvm_memory_slot *slot; @@ -959,10 +960,11 @@ static int FNAME(sync_spte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, int sptep = &sp->spt[i]; spte = *sptep; host_writable = spte & shadow_host_writable_mask; + is_refcounted = spte & SPTE_MMU_PAGE_REFCOUNTED; slot = kvm_vcpu_gfn_to_memslot(vcpu, gfn); make_spte(vcpu, sp, slot, pte_access, gfn, spte_to_pfn(spte), spte, true, false, - host_writable, &spte); + host_writable, is_refcounted, &spte); return mmu_spte_update(sptep, spte); } diff --git a/arch/x86/kvm/mmu/spte.c b/arch/x86/kvm/mmu/spte.c index 4a599130e9c9..ce495819061f 100644 --- a/arch/x86/kvm/mmu/spte.c +++ b/arch/x86/kvm/mmu/spte.c @@ -138,7 +138,7 @@ bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, const struct kvm_memory_slot *slot, unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, u64 old_spte, bool prefetch, bool can_unsync, - bool host_writable, u64 *new_spte) + bool host_writable, bool is_refcounted, u64 *new_spte) { int level = sp->role.level; u64 spte = SPTE_MMU_PRESENT_MASK; @@ -188,6 +188,8 @@ bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, if (level > PG_LEVEL_4K) spte |= PT_PAGE_SIZE_MASK; + if (is_refcounted) + spte |= SPTE_MMU_PAGE_REFCOUNTED; if (shadow_memtype_mask) spte |= static_call(kvm_x86_get_mt_mask)(vcpu, gfn, diff --git a/arch/x86/kvm/mmu/spte.h b/arch/x86/kvm/mmu/spte.h index a129951c9a88..4bf4a535c23d 100644 --- a/arch/x86/kvm/mmu/spte.h +++ b/arch/x86/kvm/mmu/spte.h @@ -96,6 +96,11 @@ static_assert(!(EPT_SPTE_MMU_WRITABLE & SHADOW_ACC_TRACK_SAVED_MASK)); /* Defined only to keep the above static asserts readable. */ #undef SHADOW_ACC_TRACK_SAVED_MASK +/* + * Indicates that the SPTE refers to a page with a valid refcount. + */ +#define SPTE_MMU_PAGE_REFCOUNTED BIT_ULL(59) + /* * Due to limited space in PTEs, the MMIO generation is a 19 bit subset of * the memslots generation and is derived as follows: @@ -345,6 +350,11 @@ static inline bool is_dirty_spte(u64 spte) return dirty_mask ? spte & dirty_mask : spte & PT_WRITABLE_MASK; } +static inline bool is_refcounted_page_pte(u64 spte) +{ + return spte & SPTE_MMU_PAGE_REFCOUNTED; +} + static inline u64 get_rsvd_bits(struct rsvd_bits_validate *rsvd_check, u64 pte, int level) { @@ -475,7 +485,7 @@ bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, const struct kvm_memory_slot *slot, unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, u64 old_spte, bool prefetch, bool can_unsync, - bool host_writable, u64 *new_spte); + bool host_writable, bool is_refcounted, u64 *new_spte); u64 make_huge_page_split_spte(struct kvm *kvm, u64 huge_spte, union kvm_mmu_page_role role, int index); u64 make_nonleaf_spte(u64 *child_pt, bool ad_disabled); diff --git a/arch/x86/kvm/mmu/tdp_mmu.c b/arch/x86/kvm/mmu/tdp_mmu.c index 6c63f2d1675f..185f3c666c2b 100644 --- a/arch/x86/kvm/mmu/tdp_mmu.c +++ b/arch/x86/kvm/mmu/tdp_mmu.c @@ -474,6 +474,7 @@ static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, bool was_leaf = was_present && is_last_spte(old_spte, level); bool is_leaf = is_present && is_last_spte(new_spte, level); bool pfn_changed = spte_to_pfn(old_spte) != spte_to_pfn(new_spte); + bool is_refcounted = is_refcounted_page_pte(old_spte); WARN_ON_ONCE(level > PT64_ROOT_MAX_LEVEL); WARN_ON_ONCE(level < PG_LEVEL_4K); @@ -538,9 +539,9 @@ static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, if (is_leaf != was_leaf) kvm_update_page_stats(kvm, level, is_leaf ? 1 : -1); - if (was_leaf && is_dirty_spte(old_spte) && + if (was_leaf && is_dirty_spte(old_spte) && is_refcounted && (!is_present || !is_dirty_spte(new_spte) || pfn_changed)) - kvm_set_pfn_dirty(spte_to_pfn(old_spte)); + kvm_set_page_dirty(pfn_to_page(spte_to_pfn(old_spte))); /* * Recursively handle child PTs if the change removed a subtree from @@ -552,9 +553,9 @@ static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, (is_leaf || !is_present || WARN_ON_ONCE(pfn_changed))) handle_removed_pt(kvm, spte_to_child_pt(old_spte, level), shared); - if (was_leaf && is_accessed_spte(old_spte) && + if (was_leaf && is_accessed_spte(old_spte) && is_refcounted && (!is_present || !is_accessed_spte(new_spte) || pfn_changed)) - kvm_set_pfn_accessed(spte_to_pfn(old_spte)); + kvm_set_page_accessed(pfn_to_page(spte_to_pfn(old_spte))); } /* @@ -988,8 +989,9 @@ static int tdp_mmu_map_handle_target_level(struct kvm_vcpu *vcpu, new_spte = make_mmio_spte(vcpu, iter->gfn, ACC_ALL); else wrprot = make_spte(vcpu, sp, fault->slot, ACC_ALL, iter->gfn, - fault->pfn, iter->old_spte, fault->prefetch, true, - fault->map_writable, &new_spte); + fault->pfn, iter->old_spte, fault->prefetch, true, + fault->map_writable, fault->is_refcounted_page, + &new_spte); if (new_spte == iter->old_spte) ret = RET_PF_SPURIOUS; @@ -1205,8 +1207,9 @@ static bool age_gfn_range(struct kvm *kvm, struct tdp_iter *iter, * Capture the dirty status of the page, so that it doesn't get * lost when the SPTE is marked for access tracking. */ - if (is_writable_pte(iter->old_spte)) - kvm_set_pfn_dirty(spte_to_pfn(iter->old_spte)); + if (is_writable_pte(iter->old_spte) && + is_refcounted_page_pte(iter->old_spte)) + kvm_set_page_dirty(pfn_to_page(spte_to_pfn(iter->old_spte))); new_spte = mark_spte_for_access_track(iter->old_spte); iter->old_spte = kvm_tdp_mmu_write_spte(iter->sptep, @@ -1628,7 +1631,8 @@ static void clear_dirty_pt_masked(struct kvm *kvm, struct kvm_mmu_page *root, trace_kvm_tdp_mmu_spte_changed(iter.as_id, iter.gfn, iter.level, iter.old_spte, iter.old_spte & ~dbit); - kvm_set_pfn_dirty(spte_to_pfn(iter.old_spte)); + if (is_refcounted_page_pte(iter.old_spte)) + kvm_set_page_dirty(pfn_to_page(spte_to_pfn(iter.old_spte))); } rcu_read_unlock(); diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h index b95c79b7833b..6696925f01f1 100644 --- a/include/linux/kvm_host.h +++ b/include/linux/kvm_host.h @@ -1179,6 +1179,9 @@ unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn, void kvm_release_page_clean(struct page *page); void kvm_release_page_dirty(struct page *page); +void kvm_set_page_accessed(struct page *page); +void kvm_set_page_dirty(struct page *page); + struct kvm_follow_pfn { const struct kvm_memory_slot *slot; gfn_t gfn; diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c index 913de4e86d9d..4d8538cdb690 100644 --- a/virt/kvm/kvm_main.c +++ b/virt/kvm/kvm_main.c @@ -2979,17 +2979,19 @@ static bool kvm_is_ad_tracked_page(struct page *page) return !PageReserved(page); } -static void kvm_set_page_dirty(struct page *page) +void kvm_set_page_dirty(struct page *page) { if (kvm_is_ad_tracked_page(page)) SetPageDirty(page); } +EXPORT_SYMBOL_GPL(kvm_set_page_dirty); -static void kvm_set_page_accessed(struct page *page) +void kvm_set_page_accessed(struct page *page) { if (kvm_is_ad_tracked_page(page)) mark_page_accessed(page); } +EXPORT_SYMBOL_GPL(kvm_set_page_accessed); void kvm_release_page_clean(struct page *page) {