Message ID | 1453361977-19589-7-git-send-email-aik@ozlabs.ru (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Hi Alexey, [auto build test ERROR on kvm/linux-next] [also build test ERROR on v4.4 next-20160121] [if your patch is applied to the wrong git tree, please drop us a note to help improving the system] url: https://github.com/0day-ci/linux/commits/Alexey-Kardashevskiy/KVM-PPC-Add-in-kernel-multitce-handling/20160121-154336 base: https://git.kernel.org/pub/scm/virt/kvm/kvm.git linux-next config: powerpc-allyesconfig (attached as .config) reproduce: wget https://git.kernel.org/cgit/linux/kernel/git/wfg/lkp-tests.git/plain/sbin/make.cross -O ~/bin/make.cross chmod +x ~/bin/make.cross # save the attached .config to linux build tree make.cross ARCH=powerpc All error/warnings (new ones prefixed by >>): arch/powerpc/kvm/book3s_64_vio_hv.c: In function 'kvmppc_find_table': arch/powerpc/kvm/book3s_64_vio_hv.c:58:2: error: implicit declaration of function 'list_for_each_entry_lockless' [-Werror=implicit-function-declaration] list_for_each_entry_lockless(stt, &kvm->arch.spapr_tce_tables, list) ^ arch/powerpc/kvm/book3s_64_vio_hv.c:58:65: error: 'list' undeclared (first use in this function) list_for_each_entry_lockless(stt, &kvm->arch.spapr_tce_tables, list) ^ arch/powerpc/kvm/book3s_64_vio_hv.c:58:65: note: each undeclared identifier is reported only once for each function it appears in arch/powerpc/kvm/book3s_64_vio_hv.c:59:3: error: expected ';' before 'if' if (stt->liobn == liobn) ^ arch/powerpc/kvm/book3s_64_vio_hv.c: In function 'kvmppc_rm_h_put_tce_indirect': >> arch/powerpc/kvm/book3s_64_vio_hv.c:263:18: error: implicit declaration of function 'vmalloc_to_phys' [-Werror=implicit-function-declaration] rmap = (void *) vmalloc_to_phys(rmap); ^ >> arch/powerpc/kvm/book3s_64_vio_hv.c:263:9: warning: cast to pointer from integer of different size [-Wint-to-pointer-cast] rmap = (void *) vmalloc_to_phys(rmap); ^ cc1: some warnings being treated as errors vim +/vmalloc_to_phys +263 arch/powerpc/kvm/book3s_64_vio_hv.c 257 if (ret != H_SUCCESS) 258 return ret; 259 260 if (kvmppc_gpa_to_ua(vcpu->kvm, tce_list, &ua, &rmap)) 261 return H_TOO_HARD; 262 > 263 rmap = (void *) vmalloc_to_phys(rmap); 264 265 lock_rmap(rmap); 266 if (kvmppc_rm_ua_to_hpa(vcpu, ua, &tces)) { --- 0-DAY kernel test infrastructure Open Source Technology Center https://lists.01.org/pipermail/kbuild-all Intel Corporation
Right, this also depends on 69b907297f4e list: Add lockless list traversal primitives On 01/21/2016 06:56 PM, kbuild test robot wrote: > Hi Alexey, > > [auto build test ERROR on kvm/linux-next] > [also build test ERROR on v4.4 next-20160121] > [if your patch is applied to the wrong git tree, please drop us a note to help improving the system] > > url: https://github.com/0day-ci/linux/commits/Alexey-Kardashevskiy/KVM-PPC-Add-in-kernel-multitce-handling/20160121-154336 > base: https://git.kernel.org/pub/scm/virt/kvm/kvm.git linux-next > config: powerpc-allyesconfig (attached as .config) > reproduce: > wget https://git.kernel.org/cgit/linux/kernel/git/wfg/lkp-tests.git/plain/sbin/make.cross -O ~/bin/make.cross > chmod +x ~/bin/make.cross > # save the attached .config to linux build tree > make.cross ARCH=powerpc > > All error/warnings (new ones prefixed by >>): > > arch/powerpc/kvm/book3s_64_vio_hv.c: In function 'kvmppc_find_table': > arch/powerpc/kvm/book3s_64_vio_hv.c:58:2: error: implicit declaration of function 'list_for_each_entry_lockless' [-Werror=implicit-function-declaration] > list_for_each_entry_lockless(stt, &kvm->arch.spapr_tce_tables, list) > ^ > arch/powerpc/kvm/book3s_64_vio_hv.c:58:65: error: 'list' undeclared (first use in this function) > list_for_each_entry_lockless(stt, &kvm->arch.spapr_tce_tables, list) > ^ > arch/powerpc/kvm/book3s_64_vio_hv.c:58:65: note: each undeclared identifier is reported only once for each function it appears in > arch/powerpc/kvm/book3s_64_vio_hv.c:59:3: error: expected ';' before 'if' > if (stt->liobn == liobn) > ^ > arch/powerpc/kvm/book3s_64_vio_hv.c: In function 'kvmppc_rm_h_put_tce_indirect': >>> arch/powerpc/kvm/book3s_64_vio_hv.c:263:18: error: implicit declaration of function 'vmalloc_to_phys' [-Werror=implicit-function-declaration] > rmap = (void *) vmalloc_to_phys(rmap); > ^ >>> arch/powerpc/kvm/book3s_64_vio_hv.c:263:9: warning: cast to pointer from integer of different size [-Wint-to-pointer-cast] > rmap = (void *) vmalloc_to_phys(rmap); > ^ > cc1: some warnings being treated as errors > > vim +/vmalloc_to_phys +263 arch/powerpc/kvm/book3s_64_vio_hv.c > > 257 if (ret != H_SUCCESS) > 258 return ret; > 259 > 260 if (kvmppc_gpa_to_ua(vcpu->kvm, tce_list, &ua, &rmap)) > 261 return H_TOO_HARD; > 262 > > 263 rmap = (void *) vmalloc_to_phys(rmap); > 264 > 265 lock_rmap(rmap); > 266 if (kvmppc_rm_ua_to_hpa(vcpu, ua, &tces)) { > > --- > 0-DAY kernel test infrastructure Open Source Technology Center > https://lists.01.org/pipermail/kbuild-all Intel Corporation >
On Thu, Jan 21, 2016 at 06:39:37PM +1100, Alexey Kardashevskiy wrote: > This adds real and virtual mode handlers for the H_PUT_TCE_INDIRECT and > H_STUFF_TCE hypercalls for user space emulated devices such as IBMVIO > devices or emulated PCI. These calls allow adding multiple entries > (up to 512) into the TCE table in one call which saves time on > transition between kernel and user space. > > This implements the KVM_CAP_PPC_MULTITCE capability. When present, > the kernel will try handling H_PUT_TCE_INDIRECT and H_STUFF_TCE. > If they can not be handled by the kernel, they are passed on to > the user space. The user space still has to have an implementation > for these. > > Both HV and PR-syle KVM are supported. > > Signed-off-by: Alexey Kardashevskiy <aik@ozlabs.ru> > --- > Changes: > v2: > * compare @ret with H_SUCCESS instead of assuming H_SUCCESS is zero > * s/~IOMMU_PAGE_MASK_4K/SZ_4K-1/ when testing @tce_list > --- > Documentation/virtual/kvm/api.txt | 25 ++++++ > arch/powerpc/include/asm/kvm_ppc.h | 12 +++ > arch/powerpc/kvm/book3s_64_vio.c | 110 +++++++++++++++++++++++- > arch/powerpc/kvm/book3s_64_vio_hv.c | 145 ++++++++++++++++++++++++++++++-- > arch/powerpc/kvm/book3s_hv.c | 26 +++++- > arch/powerpc/kvm/book3s_hv_rmhandlers.S | 6 +- > arch/powerpc/kvm/book3s_pr_papr.c | 35 ++++++++ > arch/powerpc/kvm/powerpc.c | 3 + > 8 files changed, 349 insertions(+), 13 deletions(-) > > diff --git a/Documentation/virtual/kvm/api.txt b/Documentation/virtual/kvm/api.txt > index 07e4cdf..da39435 100644 > --- a/Documentation/virtual/kvm/api.txt > +++ b/Documentation/virtual/kvm/api.txt > @@ -3035,6 +3035,31 @@ Returns: 0 on success, -1 on error > > Queues an SMI on the thread's vcpu. > > +4.97 KVM_CAP_PPC_MULTITCE > + > +Capability: KVM_CAP_PPC_MULTITCE > +Architectures: ppc > +Type: vm > + > +This capability means the kernel is capable of handling hypercalls > +H_PUT_TCE_INDIRECT and H_STUFF_TCE without passing those into the user > +space. This significantly accelerates DMA operations for PPC KVM guests. > +User space should expect that its handlers for these hypercalls > +are not going to be called if user space previously registered LIOBN > +in KVM (via KVM_CREATE_SPAPR_TCE or similar calls). > + > +In order to enable H_PUT_TCE_INDIRECT and H_STUFF_TCE use in the guest, > +user space might have to advertise it for the guest. For example, > +IBM pSeries (sPAPR) guest starts using them if "hcall-multi-tce" is > +present in the "ibm,hypertas-functions" device-tree property. > + > +The hypercalls mentioned above may or may not be processed successfully > +in the kernel based fast path. If they can not be handled by the kernel, > +they will get passed on to user space. So user space still has to have > +an implementation for these despite the in kernel acceleration. > + > +This capability is always enabled. > + > 5. The kvm_run structure > ------------------------ > > diff --git a/arch/powerpc/include/asm/kvm_ppc.h b/arch/powerpc/include/asm/kvm_ppc.h > index 9513911..4cadee5 100644 > --- a/arch/powerpc/include/asm/kvm_ppc.h > +++ b/arch/powerpc/include/asm/kvm_ppc.h > @@ -166,12 +166,24 @@ extern int kvmppc_pseries_do_hcall(struct kvm_vcpu *vcpu); > > extern long kvm_vm_ioctl_create_spapr_tce(struct kvm *kvm, > struct kvm_create_spapr_tce *args); > +extern struct kvmppc_spapr_tce_table *kvmppc_find_table( > + struct kvm_vcpu *vcpu, unsigned long liobn); > extern long kvmppc_ioba_validate(struct kvmppc_spapr_tce_table *stt, > unsigned long ioba, unsigned long npages); > extern long kvmppc_tce_validate(struct kvmppc_spapr_tce_table *tt, > unsigned long tce); > +extern long kvmppc_gpa_to_ua(struct kvm *kvm, unsigned long gpa, > + unsigned long *ua, unsigned long **prmap); Putting a userspace address into an unsigned long is pretty nasty: it should be a something __user *. > +extern void kvmppc_tce_put(struct kvmppc_spapr_tce_table *tt, > + unsigned long idx, unsigned long tce); > extern long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, > unsigned long ioba, unsigned long tce); > +extern long kvmppc_h_put_tce_indirect(struct kvm_vcpu *vcpu, > + unsigned long liobn, unsigned long ioba, > + unsigned long tce_list, unsigned long npages); > +extern long kvmppc_h_stuff_tce(struct kvm_vcpu *vcpu, > + unsigned long liobn, unsigned long ioba, > + unsigned long tce_value, unsigned long npages); > extern long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn, > unsigned long ioba); > extern struct page *kvm_alloc_hpt(unsigned long nr_pages); > diff --git a/arch/powerpc/kvm/book3s_64_vio.c b/arch/powerpc/kvm/book3s_64_vio.c > index 975f0ab..987f406 100644 > --- a/arch/powerpc/kvm/book3s_64_vio.c > +++ b/arch/powerpc/kvm/book3s_64_vio.c > @@ -14,6 +14,7 @@ > * > * Copyright 2010 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> > * Copyright 2011 David Gibson, IBM Corporation <dwg@au1.ibm.com> > + * Copyright 2016 Alexey Kardashevskiy, IBM Corporation <aik@au1.ibm.com> > */ > > #include <linux/types.h> > @@ -37,8 +38,7 @@ > #include <asm/kvm_host.h> > #include <asm/udbg.h> > #include <asm/iommu.h> > - > -#define TCES_PER_PAGE (PAGE_SIZE / sizeof(u64)) > +#include <asm/tce.h> > > static unsigned long kvmppc_stt_npages(unsigned long window_size) > { > @@ -200,3 +200,109 @@ fail: > } > return ret; > } > + > +long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, > + unsigned long liobn, unsigned long ioba, > + unsigned long tce) > +{ > + long ret; > + struct kvmppc_spapr_tce_table *stt; > + > + stt = kvmppc_find_table(vcpu, liobn); > + if (!stt) > + return H_TOO_HARD; > + > + ret = kvmppc_ioba_validate(stt, ioba, 1); > + if (ret != H_SUCCESS) > + return ret; > + > + ret = kvmppc_tce_validate(stt, tce); > + if (ret != H_SUCCESS) > + return ret; > + > + kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce); > + > + return H_SUCCESS; > +} > +EXPORT_SYMBOL_GPL(kvmppc_h_put_tce); > + > +long kvmppc_h_put_tce_indirect(struct kvm_vcpu *vcpu, > + unsigned long liobn, unsigned long ioba, > + unsigned long tce_list, unsigned long npages) > +{ > + struct kvmppc_spapr_tce_table *stt; > + long i, ret = H_SUCCESS, idx; > + unsigned long entry, ua = 0; > + u64 __user *tces, tce; > + > + stt = kvmppc_find_table(vcpu, liobn); > + if (!stt) > + return H_TOO_HARD; > + > + entry = ioba >> IOMMU_PAGE_SHIFT_4K; > + /* > + * SPAPR spec says that the maximum size of the list is 512 TCEs > + * so the whole table fits in 4K page > + */ > + if (npages > 512) > + return H_PARAMETER; > + > + if (tce_list & (SZ_4K - 1)) > + return H_PARAMETER; > + > + ret = kvmppc_ioba_validate(stt, ioba, npages); > + if (ret != H_SUCCESS) > + return ret; > + > + idx = srcu_read_lock(&vcpu->kvm->srcu); > + if (kvmppc_gpa_to_ua(vcpu->kvm, tce_list, &ua, NULL)) { > + ret = H_TOO_HARD; > + goto unlock_exit; > + } > + tces = (u64 __user *) ua; > + > + for (i = 0; i < npages; ++i) { > + if (get_user(tce, tces + i)) { > + ret = H_PARAMETER; > + goto unlock_exit; > + } > + tce = be64_to_cpu(tce); > + > + ret = kvmppc_tce_validate(stt, tce); > + if (ret != H_SUCCESS) > + goto unlock_exit; > + > + kvmppc_tce_put(stt, entry + i, tce); > + } > + > +unlock_exit: > + srcu_read_unlock(&vcpu->kvm->srcu, idx); > + > + return ret; > +} > +EXPORT_SYMBOL_GPL(kvmppc_h_put_tce_indirect); > + > +long kvmppc_h_stuff_tce(struct kvm_vcpu *vcpu, > + unsigned long liobn, unsigned long ioba, > + unsigned long tce_value, unsigned long npages) > +{ > + struct kvmppc_spapr_tce_table *stt; > + long i, ret; > + > + stt = kvmppc_find_table(vcpu, liobn); > + if (!stt) > + return H_TOO_HARD; > + > + ret = kvmppc_ioba_validate(stt, ioba, npages); > + if (ret != H_SUCCESS) > + return ret; > + > + if (tce_value & (TCE_PCI_WRITE | TCE_PCI_READ)) > + return H_PARAMETER; Do we really need to allow no-permission but non-zero TCEs? > + > + for (i = 0; i < npages; ++i, ioba += IOMMU_PAGE_SIZE_4K) > + kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce_value); > + > + return H_SUCCESS; > +} > +EXPORT_SYMBOL_GPL(kvmppc_h_stuff_tce); > diff --git a/arch/powerpc/kvm/book3s_64_vio_hv.c b/arch/powerpc/kvm/book3s_64_vio_hv.c > index 8cd3a95..58c63ed 100644 > --- a/arch/powerpc/kvm/book3s_64_vio_hv.c > +++ b/arch/powerpc/kvm/book3s_64_vio_hv.c > @@ -14,6 +14,7 @@ > * > * Copyright 2010 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> > * Copyright 2011 David Gibson, IBM Corporation <dwg@au1.ibm.com> > + * Copyright 2016 Alexey Kardashevskiy, IBM Corporation <aik@au1.ibm.com> > */ > > #include <linux/types.h> > @@ -30,6 +31,7 @@ > #include <asm/kvm_ppc.h> > #include <asm/kvm_book3s.h> > #include <asm/mmu-hash64.h> > +#include <asm/mmu_context.h> > #include <asm/hvcall.h> > #include <asm/synch.h> > #include <asm/ppc-opcode.h> > @@ -37,6 +39,7 @@ > #include <asm/udbg.h> > #include <asm/iommu.h> > #include <asm/tce.h> > +#include <asm/iommu.h> > > #define TCES_PER_PAGE (PAGE_SIZE / sizeof(u64)) > > @@ -46,7 +49,7 @@ > * WARNING: This will be called in real or virtual mode on HV KVM and virtual > * mode on PR KVM > */ > -static struct kvmppc_spapr_tce_table *kvmppc_find_table(struct kvm_vcpu *vcpu, > +struct kvmppc_spapr_tce_table *kvmppc_find_table(struct kvm_vcpu *vcpu, > unsigned long liobn) > { > struct kvm *kvm = vcpu->kvm; > @@ -58,6 +61,7 @@ static struct kvmppc_spapr_tce_table *kvmppc_find_table(struct kvm_vcpu *vcpu, > > return NULL; > } > +EXPORT_SYMBOL_GPL(kvmppc_find_table); > > /* > * Validates IO address. > @@ -150,11 +154,31 @@ void kvmppc_tce_put(struct kvmppc_spapr_tce_table *stt, > } > EXPORT_SYMBOL_GPL(kvmppc_tce_put); > > -/* WARNING: This will be called in real-mode on HV KVM and virtual > - * mode on PR KVM > - */ > -long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, > - unsigned long ioba, unsigned long tce) > +long kvmppc_gpa_to_ua(struct kvm *kvm, unsigned long gpa, > + unsigned long *ua, unsigned long **prmap) > +{ > + unsigned long gfn = gpa >> PAGE_SHIFT; > + struct kvm_memory_slot *memslot; > + > + memslot = search_memslots(kvm_memslots(kvm), gfn); > + if (!memslot) > + return -EINVAL; > + > + *ua = __gfn_to_hva_memslot(memslot, gfn) | > + (gpa & ~(PAGE_MASK | TCE_PCI_READ | TCE_PCI_WRITE)); > + > +#ifdef CONFIG_KVM_BOOK3S_HV_POSSIBLE It's a bit odd to see a test for HV_POSSIBLE in a file named book3s_64_vio_hv.c > + if (prmap) > + *prmap = &memslot->arch.rmap[gfn - memslot->base_gfn]; > +#endif > + > + return 0; > +} > +EXPORT_SYMBOL_GPL(kvmppc_gpa_to_ua); > + > +#ifdef CONFIG_KVM_BOOK3S_HV_POSSIBLE > +long kvmppc_rm_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, > + unsigned long ioba, unsigned long tce) > { > struct kvmppc_spapr_tce_table *stt = kvmppc_find_table(vcpu, liobn); > long ret; > @@ -177,7 +201,112 @@ long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, > > return H_SUCCESS; > } > -EXPORT_SYMBOL_GPL(kvmppc_h_put_tce); > + > +static long kvmppc_rm_ua_to_hpa(struct kvm_vcpu *vcpu, > + unsigned long ua, unsigned long *phpa) ua should be a something __user * rather than an unsigned long. And come to that hpa should be a something * rather than an unsigned long. > +{ > + pte_t *ptep, pte; > + unsigned shift = 0; > + > + ptep = __find_linux_pte_or_hugepte(vcpu->arch.pgdir, ua, NULL, &shift); > + if (!ptep || !pte_present(*ptep)) > + return -ENXIO; > + pte = *ptep; > + > + if (!shift) > + shift = PAGE_SHIFT; > + > + /* Avoid handling anything potentially complicated in realmode */ > + if (shift > PAGE_SHIFT) > + return -EAGAIN; > + > + if (!pte_young(pte)) > + return -EAGAIN; > + > + *phpa = (pte_pfn(pte) << PAGE_SHIFT) | (ua & ((1ULL << shift) - 1)) | > + (ua & ~PAGE_MASK); > + > + return 0; > +} > + > +long kvmppc_rm_h_put_tce_indirect(struct kvm_vcpu *vcpu, > + unsigned long liobn, unsigned long ioba, > + unsigned long tce_list, unsigned long npages) > +{ > + struct kvmppc_spapr_tce_table *stt; > + long i, ret = H_SUCCESS; > + unsigned long tces, entry, ua = 0; > + unsigned long *rmap = NULL; > + > + stt = kvmppc_find_table(vcpu, liobn); > + if (!stt) > + return H_TOO_HARD; > + > + entry = ioba >> IOMMU_PAGE_SHIFT_4K; > + /* > + * The spec says that the maximum size of the list is 512 TCEs > + * so the whole table addressed resides in 4K page > + */ > + if (npages > 512) > + return H_PARAMETER; > + > + if (tce_list & (SZ_4K - 1)) > + return H_PARAMETER; > + > + ret = kvmppc_ioba_validate(stt, ioba, npages); > + if (ret != H_SUCCESS) > + return ret; > + > + if (kvmppc_gpa_to_ua(vcpu->kvm, tce_list, &ua, &rmap)) > + return H_TOO_HARD; > + > + rmap = (void *) vmalloc_to_phys(rmap); > + > + lock_rmap(rmap); > + if (kvmppc_rm_ua_to_hpa(vcpu, ua, &tces)) { > + ret = H_TOO_HARD; > + goto unlock_exit; > + } > + > + for (i = 0; i < npages; ++i) { > + unsigned long tce = be64_to_cpu(((u64 *)tces)[i]); > + > + ret = kvmppc_tce_validate(stt, tce); > + if (ret != H_SUCCESS) > + goto unlock_exit; > + > + kvmppc_tce_put(stt, entry + i, tce); > + } > + > +unlock_exit: > + unlock_rmap(rmap); > + > + return ret; > +} > + > +long kvmppc_rm_h_stuff_tce(struct kvm_vcpu *vcpu, > + unsigned long liobn, unsigned long ioba, > + unsigned long tce_value, unsigned long npages) Unlike put_indirect, this code appears to be identical to the non realmode code - can you combine them? > +{ > + struct kvmppc_spapr_tce_table *stt; > + long i, ret; > + > + stt = kvmppc_find_table(vcpu, liobn); > + if (!stt) > + return H_TOO_HARD; > + > + ret = kvmppc_ioba_validate(stt, ioba, npages); > + if (ret != H_SUCCESS) > + return ret; > + > + if (tce_value & (TCE_PCI_WRITE | TCE_PCI_READ)) > + return H_PARAMETER; > + > + for (i = 0; i < npages; ++i, ioba += IOMMU_PAGE_SIZE_4K) > + kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce_value); > + > + return H_SUCCESS; > +} > > long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn, > unsigned long ioba) > @@ -204,3 +333,5 @@ long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn, > return H_SUCCESS; > } > EXPORT_SYMBOL_GPL(kvmppc_h_get_tce); > + > +#endif /* KVM_BOOK3S_HV_POSSIBLE */ > diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c > index cff207b..df3fbae 100644 > --- a/arch/powerpc/kvm/book3s_hv.c > +++ b/arch/powerpc/kvm/book3s_hv.c > @@ -768,7 +768,31 @@ int kvmppc_pseries_do_hcall(struct kvm_vcpu *vcpu) > if (kvmppc_xics_enabled(vcpu)) { > ret = kvmppc_xics_hcall(vcpu, req); > break; > - } /* fallthrough */ > + } > + return RESUME_HOST; > + case H_PUT_TCE: > + ret = kvmppc_h_put_tce(vcpu, kvmppc_get_gpr(vcpu, 4), > + kvmppc_get_gpr(vcpu, 5), > + kvmppc_get_gpr(vcpu, 6)); > + if (ret == H_TOO_HARD) > + return RESUME_HOST; > + break; > + case H_PUT_TCE_INDIRECT: > + ret = kvmppc_h_put_tce_indirect(vcpu, kvmppc_get_gpr(vcpu, 4), > + kvmppc_get_gpr(vcpu, 5), > + kvmppc_get_gpr(vcpu, 6), > + kvmppc_get_gpr(vcpu, 7)); > + if (ret == H_TOO_HARD) > + return RESUME_HOST; > + break; > + case H_STUFF_TCE: > + ret = kvmppc_h_stuff_tce(vcpu, kvmppc_get_gpr(vcpu, 4), > + kvmppc_get_gpr(vcpu, 5), > + kvmppc_get_gpr(vcpu, 6), > + kvmppc_get_gpr(vcpu, 7)); > + if (ret == H_TOO_HARD) > + return RESUME_HOST; > + break; > default: > return RESUME_HOST; > } > diff --git a/arch/powerpc/kvm/book3s_hv_rmhandlers.S b/arch/powerpc/kvm/book3s_hv_rmhandlers.S > index 3c6badc..3bf6e72 100644 > --- a/arch/powerpc/kvm/book3s_hv_rmhandlers.S > +++ b/arch/powerpc/kvm/book3s_hv_rmhandlers.S > @@ -1928,7 +1928,7 @@ hcall_real_table: > .long DOTSYM(kvmppc_h_clear_ref) - hcall_real_table > .long DOTSYM(kvmppc_h_protect) - hcall_real_table > .long DOTSYM(kvmppc_h_get_tce) - hcall_real_table > - .long DOTSYM(kvmppc_h_put_tce) - hcall_real_table > + .long DOTSYM(kvmppc_rm_h_put_tce) - hcall_real_table > .long 0 /* 0x24 - H_SET_SPRG0 */ > .long DOTSYM(kvmppc_h_set_dabr) - hcall_real_table > .long 0 /* 0x2c */ > @@ -2006,8 +2006,8 @@ hcall_real_table: > .long 0 /* 0x12c */ > .long 0 /* 0x130 */ > .long DOTSYM(kvmppc_h_set_xdabr) - hcall_real_table > - .long 0 /* 0x138 */ > - .long 0 /* 0x13c */ > + .long DOTSYM(kvmppc_rm_h_stuff_tce) - hcall_real_table > + .long DOTSYM(kvmppc_rm_h_put_tce_indirect) - hcall_real_table > .long 0 /* 0x140 */ > .long 0 /* 0x144 */ > .long 0 /* 0x148 */ > diff --git a/arch/powerpc/kvm/book3s_pr_papr.c b/arch/powerpc/kvm/book3s_pr_papr.c > index f2c75a1..02176fd 100644 > --- a/arch/powerpc/kvm/book3s_pr_papr.c > +++ b/arch/powerpc/kvm/book3s_pr_papr.c > @@ -280,6 +280,37 @@ static int kvmppc_h_pr_logical_ci_store(struct kvm_vcpu *vcpu) > return EMULATE_DONE; > } > > +static int kvmppc_h_pr_put_tce_indirect(struct kvm_vcpu *vcpu) > +{ > + unsigned long liobn = kvmppc_get_gpr(vcpu, 4); > + unsigned long ioba = kvmppc_get_gpr(vcpu, 5); > + unsigned long tce = kvmppc_get_gpr(vcpu, 6); > + unsigned long npages = kvmppc_get_gpr(vcpu, 7); > + long rc; > + > + rc = kvmppc_h_put_tce_indirect(vcpu, liobn, ioba, > + tce, npages); > + if (rc == H_TOO_HARD) > + return EMULATE_FAIL; > + kvmppc_set_gpr(vcpu, 3, rc); > + return EMULATE_DONE; > +} > + > +static int kvmppc_h_pr_stuff_tce(struct kvm_vcpu *vcpu) > +{ > + unsigned long liobn = kvmppc_get_gpr(vcpu, 4); > + unsigned long ioba = kvmppc_get_gpr(vcpu, 5); > + unsigned long tce_value = kvmppc_get_gpr(vcpu, 6); > + unsigned long npages = kvmppc_get_gpr(vcpu, 7); > + long rc; > + > + rc = kvmppc_h_stuff_tce(vcpu, liobn, ioba, tce_value, npages); > + if (rc == H_TOO_HARD) > + return EMULATE_FAIL; > + kvmppc_set_gpr(vcpu, 3, rc); > + return EMULATE_DONE; > +} > + > static int kvmppc_h_pr_xics_hcall(struct kvm_vcpu *vcpu, u32 cmd) > { > long rc = kvmppc_xics_hcall(vcpu, cmd); > @@ -306,6 +337,10 @@ int kvmppc_h_pr(struct kvm_vcpu *vcpu, unsigned long cmd) > return kvmppc_h_pr_bulk_remove(vcpu); > case H_PUT_TCE: > return kvmppc_h_pr_put_tce(vcpu); > + case H_PUT_TCE_INDIRECT: > + return kvmppc_h_pr_put_tce_indirect(vcpu); > + case H_STUFF_TCE: > + return kvmppc_h_pr_stuff_tce(vcpu); > case H_CEDE: > kvmppc_set_msr_fast(vcpu, kvmppc_get_msr(vcpu) | MSR_EE); > kvm_vcpu_block(vcpu); > diff --git a/arch/powerpc/kvm/powerpc.c b/arch/powerpc/kvm/powerpc.c > index 6fd2405..164735c 100644 > --- a/arch/powerpc/kvm/powerpc.c > +++ b/arch/powerpc/kvm/powerpc.c > @@ -569,6 +569,9 @@ int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext) > case KVM_CAP_PPC_GET_SMMU_INFO: > r = 1; > break; > + case KVM_CAP_SPAPR_MULTITCE: > + r = 1; > + break; Hmm, usual practice has been not to enable new KVM hcalls, unless userspace (qemu) explicitly enables them with ENABLE_HCALL. I don't see an obvious way this extension could break, but it's probably safest to continue that pattern. > #endif > default: > r = 0;
On 01/25/2016 11:44 AM, David Gibson wrote: > On Thu, Jan 21, 2016 at 06:39:37PM +1100, Alexey Kardashevskiy wrote: >> This adds real and virtual mode handlers for the H_PUT_TCE_INDIRECT and >> H_STUFF_TCE hypercalls for user space emulated devices such as IBMVIO >> devices or emulated PCI. These calls allow adding multiple entries >> (up to 512) into the TCE table in one call which saves time on >> transition between kernel and user space. >> >> This implements the KVM_CAP_PPC_MULTITCE capability. When present, >> the kernel will try handling H_PUT_TCE_INDIRECT and H_STUFF_TCE. >> If they can not be handled by the kernel, they are passed on to >> the user space. The user space still has to have an implementation >> for these. >> >> Both HV and PR-syle KVM are supported. >> >> Signed-off-by: Alexey Kardashevskiy <aik@ozlabs.ru> >> --- >> Changes: >> v2: >> * compare @ret with H_SUCCESS instead of assuming H_SUCCESS is zero >> * s/~IOMMU_PAGE_MASK_4K/SZ_4K-1/ when testing @tce_list >> --- >> Documentation/virtual/kvm/api.txt | 25 ++++++ >> arch/powerpc/include/asm/kvm_ppc.h | 12 +++ >> arch/powerpc/kvm/book3s_64_vio.c | 110 +++++++++++++++++++++++- >> arch/powerpc/kvm/book3s_64_vio_hv.c | 145 ++++++++++++++++++++++++++++++-- >> arch/powerpc/kvm/book3s_hv.c | 26 +++++- >> arch/powerpc/kvm/book3s_hv_rmhandlers.S | 6 +- >> arch/powerpc/kvm/book3s_pr_papr.c | 35 ++++++++ >> arch/powerpc/kvm/powerpc.c | 3 + >> 8 files changed, 349 insertions(+), 13 deletions(-) >> >> diff --git a/Documentation/virtual/kvm/api.txt b/Documentation/virtual/kvm/api.txt >> index 07e4cdf..da39435 100644 >> --- a/Documentation/virtual/kvm/api.txt >> +++ b/Documentation/virtual/kvm/api.txt >> @@ -3035,6 +3035,31 @@ Returns: 0 on success, -1 on error >> >> Queues an SMI on the thread's vcpu. >> >> +4.97 KVM_CAP_PPC_MULTITCE >> + >> +Capability: KVM_CAP_PPC_MULTITCE >> +Architectures: ppc >> +Type: vm >> + >> +This capability means the kernel is capable of handling hypercalls >> +H_PUT_TCE_INDIRECT and H_STUFF_TCE without passing those into the user >> +space. This significantly accelerates DMA operations for PPC KVM guests. >> +User space should expect that its handlers for these hypercalls >> +are not going to be called if user space previously registered LIOBN >> +in KVM (via KVM_CREATE_SPAPR_TCE or similar calls). >> + >> +In order to enable H_PUT_TCE_INDIRECT and H_STUFF_TCE use in the guest, >> +user space might have to advertise it for the guest. For example, >> +IBM pSeries (sPAPR) guest starts using them if "hcall-multi-tce" is >> +present in the "ibm,hypertas-functions" device-tree property. >> + >> +The hypercalls mentioned above may or may not be processed successfully >> +in the kernel based fast path. If they can not be handled by the kernel, >> +they will get passed on to user space. So user space still has to have >> +an implementation for these despite the in kernel acceleration. >> + >> +This capability is always enabled. >> + >> 5. The kvm_run structure >> ------------------------ >> >> diff --git a/arch/powerpc/include/asm/kvm_ppc.h b/arch/powerpc/include/asm/kvm_ppc.h >> index 9513911..4cadee5 100644 >> --- a/arch/powerpc/include/asm/kvm_ppc.h >> +++ b/arch/powerpc/include/asm/kvm_ppc.h >> @@ -166,12 +166,24 @@ extern int kvmppc_pseries_do_hcall(struct kvm_vcpu *vcpu); >> >> extern long kvm_vm_ioctl_create_spapr_tce(struct kvm *kvm, >> struct kvm_create_spapr_tce *args); >> +extern struct kvmppc_spapr_tce_table *kvmppc_find_table( >> + struct kvm_vcpu *vcpu, unsigned long liobn); >> extern long kvmppc_ioba_validate(struct kvmppc_spapr_tce_table *stt, >> unsigned long ioba, unsigned long npages); >> extern long kvmppc_tce_validate(struct kvmppc_spapr_tce_table *tt, >> unsigned long tce); >> +extern long kvmppc_gpa_to_ua(struct kvm *kvm, unsigned long gpa, >> + unsigned long *ua, unsigned long **prmap); > > Putting a userspace address into an unsigned long is pretty nasty: it > should be a something __user *. > >> +extern void kvmppc_tce_put(struct kvmppc_spapr_tce_table *tt, >> + unsigned long idx, unsigned long tce); >> extern long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, >> unsigned long ioba, unsigned long tce); >> +extern long kvmppc_h_put_tce_indirect(struct kvm_vcpu *vcpu, >> + unsigned long liobn, unsigned long ioba, >> + unsigned long tce_list, unsigned long npages); >> +extern long kvmppc_h_stuff_tce(struct kvm_vcpu *vcpu, >> + unsigned long liobn, unsigned long ioba, >> + unsigned long tce_value, unsigned long npages); >> extern long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn, >> unsigned long ioba); >> extern struct page *kvm_alloc_hpt(unsigned long nr_pages); >> diff --git a/arch/powerpc/kvm/book3s_64_vio.c b/arch/powerpc/kvm/book3s_64_vio.c >> index 975f0ab..987f406 100644 >> --- a/arch/powerpc/kvm/book3s_64_vio.c >> +++ b/arch/powerpc/kvm/book3s_64_vio.c >> @@ -14,6 +14,7 @@ >> * >> * Copyright 2010 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> >> * Copyright 2011 David Gibson, IBM Corporation <dwg@au1.ibm.com> >> + * Copyright 2016 Alexey Kardashevskiy, IBM Corporation <aik@au1.ibm.com> >> */ >> >> #include <linux/types.h> >> @@ -37,8 +38,7 @@ >> #include <asm/kvm_host.h> >> #include <asm/udbg.h> >> #include <asm/iommu.h> >> - >> -#define TCES_PER_PAGE (PAGE_SIZE / sizeof(u64)) >> +#include <asm/tce.h> >> >> static unsigned long kvmppc_stt_npages(unsigned long window_size) >> { >> @@ -200,3 +200,109 @@ fail: >> } >> return ret; >> } >> + >> +long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, >> + unsigned long liobn, unsigned long ioba, >> + unsigned long tce) >> +{ >> + long ret; >> + struct kvmppc_spapr_tce_table *stt; >> + >> + stt = kvmppc_find_table(vcpu, liobn); >> + if (!stt) >> + return H_TOO_HARD; >> + >> + ret = kvmppc_ioba_validate(stt, ioba, 1); >> + if (ret != H_SUCCESS) >> + return ret; >> + >> + ret = kvmppc_tce_validate(stt, tce); >> + if (ret != H_SUCCESS) >> + return ret; >> + >> + kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce); >> + >> + return H_SUCCESS; >> +} >> +EXPORT_SYMBOL_GPL(kvmppc_h_put_tce); >> + >> +long kvmppc_h_put_tce_indirect(struct kvm_vcpu *vcpu, >> + unsigned long liobn, unsigned long ioba, >> + unsigned long tce_list, unsigned long npages) >> +{ >> + struct kvmppc_spapr_tce_table *stt; >> + long i, ret = H_SUCCESS, idx; >> + unsigned long entry, ua = 0; >> + u64 __user *tces, tce; >> + >> + stt = kvmppc_find_table(vcpu, liobn); >> + if (!stt) >> + return H_TOO_HARD; >> + >> + entry = ioba >> IOMMU_PAGE_SHIFT_4K; >> + /* >> + * SPAPR spec says that the maximum size of the list is 512 TCEs >> + * so the whole table fits in 4K page >> + */ >> + if (npages > 512) >> + return H_PARAMETER; >> + >> + if (tce_list & (SZ_4K - 1)) >> + return H_PARAMETER; >> + >> + ret = kvmppc_ioba_validate(stt, ioba, npages); >> + if (ret != H_SUCCESS) >> + return ret; >> + >> + idx = srcu_read_lock(&vcpu->kvm->srcu); >> + if (kvmppc_gpa_to_ua(vcpu->kvm, tce_list, &ua, NULL)) { >> + ret = H_TOO_HARD; >> + goto unlock_exit; >> + } >> + tces = (u64 __user *) ua; >> + >> + for (i = 0; i < npages; ++i) { >> + if (get_user(tce, tces + i)) { >> + ret = H_PARAMETER; >> + goto unlock_exit; >> + } >> + tce = be64_to_cpu(tce); >> + >> + ret = kvmppc_tce_validate(stt, tce); >> + if (ret != H_SUCCESS) >> + goto unlock_exit; >> + >> + kvmppc_tce_put(stt, entry + i, tce); >> + } >> + >> +unlock_exit: >> + srcu_read_unlock(&vcpu->kvm->srcu, idx); >> + >> + return ret; >> +} >> +EXPORT_SYMBOL_GPL(kvmppc_h_put_tce_indirect); >> + >> +long kvmppc_h_stuff_tce(struct kvm_vcpu *vcpu, >> + unsigned long liobn, unsigned long ioba, >> + unsigned long tce_value, unsigned long npages) >> +{ >> + struct kvmppc_spapr_tce_table *stt; >> + long i, ret; >> + >> + stt = kvmppc_find_table(vcpu, liobn); >> + if (!stt) >> + return H_TOO_HARD; >> + >> + ret = kvmppc_ioba_validate(stt, ioba, npages); >> + if (ret != H_SUCCESS) >> + return ret; >> + >> + if (tce_value & (TCE_PCI_WRITE | TCE_PCI_READ)) >> + return H_PARAMETER; > > Do we really need to allow no-permission but non-zero TCEs? Not sure, for debugging purposes one could want to poison the table. Totally useless? >> + >> + for (i = 0; i < npages; ++i, ioba += IOMMU_PAGE_SIZE_4K) >> + kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce_value); >> + >> + return H_SUCCESS; >> +} >> +EXPORT_SYMBOL_GPL(kvmppc_h_stuff_tce); >> diff --git a/arch/powerpc/kvm/book3s_64_vio_hv.c b/arch/powerpc/kvm/book3s_64_vio_hv.c >> index 8cd3a95..58c63ed 100644 >> --- a/arch/powerpc/kvm/book3s_64_vio_hv.c >> +++ b/arch/powerpc/kvm/book3s_64_vio_hv.c >> @@ -14,6 +14,7 @@ >> * >> * Copyright 2010 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> >> * Copyright 2011 David Gibson, IBM Corporation <dwg@au1.ibm.com> >> + * Copyright 2016 Alexey Kardashevskiy, IBM Corporation <aik@au1.ibm.com> >> */ >> >> #include <linux/types.h> >> @@ -30,6 +31,7 @@ >> #include <asm/kvm_ppc.h> >> #include <asm/kvm_book3s.h> >> #include <asm/mmu-hash64.h> >> +#include <asm/mmu_context.h> >> #include <asm/hvcall.h> >> #include <asm/synch.h> >> #include <asm/ppc-opcode.h> >> @@ -37,6 +39,7 @@ >> #include <asm/udbg.h> >> #include <asm/iommu.h> >> #include <asm/tce.h> >> +#include <asm/iommu.h> >> >> #define TCES_PER_PAGE (PAGE_SIZE / sizeof(u64)) >> >> @@ -46,7 +49,7 @@ >> * WARNING: This will be called in real or virtual mode on HV KVM and virtual >> * mode on PR KVM >> */ >> -static struct kvmppc_spapr_tce_table *kvmppc_find_table(struct kvm_vcpu *vcpu, >> +struct kvmppc_spapr_tce_table *kvmppc_find_table(struct kvm_vcpu *vcpu, >> unsigned long liobn) >> { >> struct kvm *kvm = vcpu->kvm; >> @@ -58,6 +61,7 @@ static struct kvmppc_spapr_tce_table *kvmppc_find_table(struct kvm_vcpu *vcpu, >> >> return NULL; >> } >> +EXPORT_SYMBOL_GPL(kvmppc_find_table); >> >> /* >> * Validates IO address. >> @@ -150,11 +154,31 @@ void kvmppc_tce_put(struct kvmppc_spapr_tce_table *stt, >> } >> EXPORT_SYMBOL_GPL(kvmppc_tce_put); >> >> -/* WARNING: This will be called in real-mode on HV KVM and virtual >> - * mode on PR KVM >> - */ >> -long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, >> - unsigned long ioba, unsigned long tce) >> +long kvmppc_gpa_to_ua(struct kvm *kvm, unsigned long gpa, >> + unsigned long *ua, unsigned long **prmap) >> +{ >> + unsigned long gfn = gpa >> PAGE_SHIFT; >> + struct kvm_memory_slot *memslot; >> + >> + memslot = search_memslots(kvm_memslots(kvm), gfn); >> + if (!memslot) >> + return -EINVAL; >> + >> + *ua = __gfn_to_hva_memslot(memslot, gfn) | >> + (gpa & ~(PAGE_MASK | TCE_PCI_READ | TCE_PCI_WRITE)); >> + >> +#ifdef CONFIG_KVM_BOOK3S_HV_POSSIBLE > > It's a bit odd to see a test for HV_POSSIBLE in a file named > book3s_64_vio_hv.c True, the file name should have probably been changed book3s_64_vio_rm.c. > >> + if (prmap) >> + *prmap = &memslot->arch.rmap[gfn - memslot->base_gfn]; >> +#endif >> + >> + return 0; >> +} >> +EXPORT_SYMBOL_GPL(kvmppc_gpa_to_ua); >> + >> +#ifdef CONFIG_KVM_BOOK3S_HV_POSSIBLE >> +long kvmppc_rm_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, >> + unsigned long ioba, unsigned long tce) >> { >> struct kvmppc_spapr_tce_table *stt = kvmppc_find_table(vcpu, liobn); >> long ret; >> @@ -177,7 +201,112 @@ long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, >> >> return H_SUCCESS; >> } >> -EXPORT_SYMBOL_GPL(kvmppc_h_put_tce); >> + >> +static long kvmppc_rm_ua_to_hpa(struct kvm_vcpu *vcpu, >> + unsigned long ua, unsigned long *phpa) > > ua should be a something __user * rather than an unsigned long. And > come to that hpa should be a something * rather than an unsigned long. @ua is the return type of __gfn_to_hva_memslot() so I kept it. Also, the only place where I actually read from this address is the virtualmode's H_PUT_TCE_INDIRECT handler, all other places just do translation with it so making it "unsigned long" saves some type convertions. It is also used in mm_iommu_ua_to_hpa() which is in upstream now (commit 15b244a88e1b289, part of DDW and preregistration patchset). Still need to change it? Regarding @phpa, the agreement here that we use "void *" for 0xC000.... type of addresses; and we use "unsigned long" if top 4 bits are not set as dereferencing such pointers will normally fail. >> +{ >> + pte_t *ptep, pte; >> + unsigned shift = 0; >> + >> + ptep = __find_linux_pte_or_hugepte(vcpu->arch.pgdir, ua, NULL, &shift); >> + if (!ptep || !pte_present(*ptep)) >> + return -ENXIO; >> + pte = *ptep; >> + >> + if (!shift) >> + shift = PAGE_SHIFT; >> + >> + /* Avoid handling anything potentially complicated in realmode */ >> + if (shift > PAGE_SHIFT) >> + return -EAGAIN; >> + >> + if (!pte_young(pte)) >> + return -EAGAIN; >> + >> + *phpa = (pte_pfn(pte) << PAGE_SHIFT) | (ua & ((1ULL << shift) - 1)) | >> + (ua & ~PAGE_MASK); >> + >> + return 0; >> +} >> + >> +long kvmppc_rm_h_put_tce_indirect(struct kvm_vcpu *vcpu, >> + unsigned long liobn, unsigned long ioba, >> + unsigned long tce_list, unsigned long npages) >> +{ >> + struct kvmppc_spapr_tce_table *stt; >> + long i, ret = H_SUCCESS; >> + unsigned long tces, entry, ua = 0; >> + unsigned long *rmap = NULL; >> + >> + stt = kvmppc_find_table(vcpu, liobn); >> + if (!stt) >> + return H_TOO_HARD; >> + >> + entry = ioba >> IOMMU_PAGE_SHIFT_4K; >> + /* >> + * The spec says that the maximum size of the list is 512 TCEs >> + * so the whole table addressed resides in 4K page >> + */ >> + if (npages > 512) >> + return H_PARAMETER; >> + >> + if (tce_list & (SZ_4K - 1)) >> + return H_PARAMETER; >> + >> + ret = kvmppc_ioba_validate(stt, ioba, npages); >> + if (ret != H_SUCCESS) >> + return ret; >> + >> + if (kvmppc_gpa_to_ua(vcpu->kvm, tce_list, &ua, &rmap)) >> + return H_TOO_HARD; >> + >> + rmap = (void *) vmalloc_to_phys(rmap); >> + >> + lock_rmap(rmap); >> + if (kvmppc_rm_ua_to_hpa(vcpu, ua, &tces)) { >> + ret = H_TOO_HARD; >> + goto unlock_exit; >> + } >> + >> + for (i = 0; i < npages; ++i) { >> + unsigned long tce = be64_to_cpu(((u64 *)tces)[i]); >> + >> + ret = kvmppc_tce_validate(stt, tce); >> + if (ret != H_SUCCESS) >> + goto unlock_exit; >> + >> + kvmppc_tce_put(stt, entry + i, tce); >> + } >> + >> +unlock_exit: >> + unlock_rmap(rmap); >> + >> + return ret; >> +} >> + >> +long kvmppc_rm_h_stuff_tce(struct kvm_vcpu *vcpu, >> + unsigned long liobn, unsigned long ioba, >> + unsigned long tce_value, unsigned long npages) > > Unlike put_indirect, this code appears to be identical to the non > realmode code - can you combine them? It is at this point but this will get different bits in "KVM: PPC: vfio kvm device: support spapr tce" later, may be sometime later I will manage to get to that part, eventually... >> +{ >> + struct kvmppc_spapr_tce_table *stt; >> + long i, ret; >> + >> + stt = kvmppc_find_table(vcpu, liobn); >> + if (!stt) >> + return H_TOO_HARD; >> + >> + ret = kvmppc_ioba_validate(stt, ioba, npages); >> + if (ret != H_SUCCESS) >> + return ret; >> + >> + if (tce_value & (TCE_PCI_WRITE | TCE_PCI_READ)) >> + return H_PARAMETER; >> + >> + for (i = 0; i < npages; ++i, ioba += IOMMU_PAGE_SIZE_4K) >> + kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce_value); >> + >> + return H_SUCCESS; >> +} >> >> long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn, >> unsigned long ioba) >> @@ -204,3 +333,5 @@ long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn, >> return H_SUCCESS; >> } >> EXPORT_SYMBOL_GPL(kvmppc_h_get_tce); >> + >> +#endif /* KVM_BOOK3S_HV_POSSIBLE */ >> diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c >> index cff207b..df3fbae 100644 >> --- a/arch/powerpc/kvm/book3s_hv.c >> +++ b/arch/powerpc/kvm/book3s_hv.c >> @@ -768,7 +768,31 @@ int kvmppc_pseries_do_hcall(struct kvm_vcpu *vcpu) >> if (kvmppc_xics_enabled(vcpu)) { >> ret = kvmppc_xics_hcall(vcpu, req); >> break; >> - } /* fallthrough */ >> + } >> + return RESUME_HOST; >> + case H_PUT_TCE: >> + ret = kvmppc_h_put_tce(vcpu, kvmppc_get_gpr(vcpu, 4), >> + kvmppc_get_gpr(vcpu, 5), >> + kvmppc_get_gpr(vcpu, 6)); >> + if (ret == H_TOO_HARD) >> + return RESUME_HOST; >> + break; >> + case H_PUT_TCE_INDIRECT: >> + ret = kvmppc_h_put_tce_indirect(vcpu, kvmppc_get_gpr(vcpu, 4), >> + kvmppc_get_gpr(vcpu, 5), >> + kvmppc_get_gpr(vcpu, 6), >> + kvmppc_get_gpr(vcpu, 7)); >> + if (ret == H_TOO_HARD) >> + return RESUME_HOST; >> + break; >> + case H_STUFF_TCE: >> + ret = kvmppc_h_stuff_tce(vcpu, kvmppc_get_gpr(vcpu, 4), >> + kvmppc_get_gpr(vcpu, 5), >> + kvmppc_get_gpr(vcpu, 6), >> + kvmppc_get_gpr(vcpu, 7)); >> + if (ret == H_TOO_HARD) >> + return RESUME_HOST; >> + break; >> default: >> return RESUME_HOST; >> } >> diff --git a/arch/powerpc/kvm/book3s_hv_rmhandlers.S b/arch/powerpc/kvm/book3s_hv_rmhandlers.S >> index 3c6badc..3bf6e72 100644 >> --- a/arch/powerpc/kvm/book3s_hv_rmhandlers.S >> +++ b/arch/powerpc/kvm/book3s_hv_rmhandlers.S >> @@ -1928,7 +1928,7 @@ hcall_real_table: >> .long DOTSYM(kvmppc_h_clear_ref) - hcall_real_table >> .long DOTSYM(kvmppc_h_protect) - hcall_real_table >> .long DOTSYM(kvmppc_h_get_tce) - hcall_real_table >> - .long DOTSYM(kvmppc_h_put_tce) - hcall_real_table >> + .long DOTSYM(kvmppc_rm_h_put_tce) - hcall_real_table >> .long 0 /* 0x24 - H_SET_SPRG0 */ >> .long DOTSYM(kvmppc_h_set_dabr) - hcall_real_table >> .long 0 /* 0x2c */ >> @@ -2006,8 +2006,8 @@ hcall_real_table: >> .long 0 /* 0x12c */ >> .long 0 /* 0x130 */ >> .long DOTSYM(kvmppc_h_set_xdabr) - hcall_real_table >> - .long 0 /* 0x138 */ >> - .long 0 /* 0x13c */ >> + .long DOTSYM(kvmppc_rm_h_stuff_tce) - hcall_real_table >> + .long DOTSYM(kvmppc_rm_h_put_tce_indirect) - hcall_real_table >> .long 0 /* 0x140 */ >> .long 0 /* 0x144 */ >> .long 0 /* 0x148 */ >> diff --git a/arch/powerpc/kvm/book3s_pr_papr.c b/arch/powerpc/kvm/book3s_pr_papr.c >> index f2c75a1..02176fd 100644 >> --- a/arch/powerpc/kvm/book3s_pr_papr.c >> +++ b/arch/powerpc/kvm/book3s_pr_papr.c >> @@ -280,6 +280,37 @@ static int kvmppc_h_pr_logical_ci_store(struct kvm_vcpu *vcpu) >> return EMULATE_DONE; >> } >> >> +static int kvmppc_h_pr_put_tce_indirect(struct kvm_vcpu *vcpu) >> +{ >> + unsigned long liobn = kvmppc_get_gpr(vcpu, 4); >> + unsigned long ioba = kvmppc_get_gpr(vcpu, 5); >> + unsigned long tce = kvmppc_get_gpr(vcpu, 6); >> + unsigned long npages = kvmppc_get_gpr(vcpu, 7); >> + long rc; >> + >> + rc = kvmppc_h_put_tce_indirect(vcpu, liobn, ioba, >> + tce, npages); >> + if (rc == H_TOO_HARD) >> + return EMULATE_FAIL; >> + kvmppc_set_gpr(vcpu, 3, rc); >> + return EMULATE_DONE; >> +} >> + >> +static int kvmppc_h_pr_stuff_tce(struct kvm_vcpu *vcpu) >> +{ >> + unsigned long liobn = kvmppc_get_gpr(vcpu, 4); >> + unsigned long ioba = kvmppc_get_gpr(vcpu, 5); >> + unsigned long tce_value = kvmppc_get_gpr(vcpu, 6); >> + unsigned long npages = kvmppc_get_gpr(vcpu, 7); >> + long rc; >> + >> + rc = kvmppc_h_stuff_tce(vcpu, liobn, ioba, tce_value, npages); >> + if (rc == H_TOO_HARD) >> + return EMULATE_FAIL; >> + kvmppc_set_gpr(vcpu, 3, rc); >> + return EMULATE_DONE; >> +} >> + >> static int kvmppc_h_pr_xics_hcall(struct kvm_vcpu *vcpu, u32 cmd) >> { >> long rc = kvmppc_xics_hcall(vcpu, cmd); >> @@ -306,6 +337,10 @@ int kvmppc_h_pr(struct kvm_vcpu *vcpu, unsigned long cmd) >> return kvmppc_h_pr_bulk_remove(vcpu); >> case H_PUT_TCE: >> return kvmppc_h_pr_put_tce(vcpu); >> + case H_PUT_TCE_INDIRECT: >> + return kvmppc_h_pr_put_tce_indirect(vcpu); >> + case H_STUFF_TCE: >> + return kvmppc_h_pr_stuff_tce(vcpu); >> case H_CEDE: >> kvmppc_set_msr_fast(vcpu, kvmppc_get_msr(vcpu) | MSR_EE); >> kvm_vcpu_block(vcpu); >> diff --git a/arch/powerpc/kvm/powerpc.c b/arch/powerpc/kvm/powerpc.c >> index 6fd2405..164735c 100644 >> --- a/arch/powerpc/kvm/powerpc.c >> +++ b/arch/powerpc/kvm/powerpc.c >> @@ -569,6 +569,9 @@ int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext) >> case KVM_CAP_PPC_GET_SMMU_INFO: >> r = 1; >> break; >> + case KVM_CAP_SPAPR_MULTITCE: >> + r = 1; >> + break; > > Hmm, usual practice has been not to enable new KVM hcalls, unless > userspace (qemu) explicitly enables them with ENABLE_HCALL. I don't > see an obvious way this extension could break, but it's probably > safest to continue that pattern. This advertises the capability but does not enable it, this is still required in QEMU: ret = kvm_vm_enable_cap(kvm_state, KVM_CAP_PPC_ENABLE_HCALL, 0, H_PUT_TCE_INDIRECT, 1); ret = kvm_vm_enable_cap(kvm_state, KVM_CAP_PPC_ENABLE_HCALL, 0, H_STUFF_TCE, 1); as multi-tce hcalls are not in the default_hcall_list list in arch/powerpc/kvm/book3s_hv.c. >> #endif >> default: >> r = 0; >
On Mon, Jan 25, 2016 at 12:24:29PM +1100, Alexey Kardashevskiy wrote: > On 01/25/2016 11:44 AM, David Gibson wrote: > >On Thu, Jan 21, 2016 at 06:39:37PM +1100, Alexey Kardashevskiy wrote: > >>This adds real and virtual mode handlers for the H_PUT_TCE_INDIRECT and > >>H_STUFF_TCE hypercalls for user space emulated devices such as IBMVIO > >>devices or emulated PCI. These calls allow adding multiple entries > >>(up to 512) into the TCE table in one call which saves time on > >>transition between kernel and user space. > >> > >>This implements the KVM_CAP_PPC_MULTITCE capability. When present, > >>the kernel will try handling H_PUT_TCE_INDIRECT and H_STUFF_TCE. > >>If they can not be handled by the kernel, they are passed on to > >>the user space. The user space still has to have an implementation > >>for these. > >> > >>Both HV and PR-syle KVM are supported. > >> > >>Signed-off-by: Alexey Kardashevskiy <aik@ozlabs.ru> > >>--- > >>Changes: > >>v2: > >>* compare @ret with H_SUCCESS instead of assuming H_SUCCESS is zero > >>* s/~IOMMU_PAGE_MASK_4K/SZ_4K-1/ when testing @tce_list > >>--- > >> Documentation/virtual/kvm/api.txt | 25 ++++++ > >> arch/powerpc/include/asm/kvm_ppc.h | 12 +++ > >> arch/powerpc/kvm/book3s_64_vio.c | 110 +++++++++++++++++++++++- > >> arch/powerpc/kvm/book3s_64_vio_hv.c | 145 ++++++++++++++++++++++++++++++-- > >> arch/powerpc/kvm/book3s_hv.c | 26 +++++- > >> arch/powerpc/kvm/book3s_hv_rmhandlers.S | 6 +- > >> arch/powerpc/kvm/book3s_pr_papr.c | 35 ++++++++ > >> arch/powerpc/kvm/powerpc.c | 3 + > >> 8 files changed, 349 insertions(+), 13 deletions(-) > >> > >>diff --git a/Documentation/virtual/kvm/api.txt b/Documentation/virtual/kvm/api.txt > >>index 07e4cdf..da39435 100644 > >>--- a/Documentation/virtual/kvm/api.txt > >>+++ b/Documentation/virtual/kvm/api.txt > >>@@ -3035,6 +3035,31 @@ Returns: 0 on success, -1 on error > >> > >> Queues an SMI on the thread's vcpu. > >> > >>+4.97 KVM_CAP_PPC_MULTITCE > >>+ > >>+Capability: KVM_CAP_PPC_MULTITCE > >>+Architectures: ppc > >>+Type: vm > >>+ > >>+This capability means the kernel is capable of handling hypercalls > >>+H_PUT_TCE_INDIRECT and H_STUFF_TCE without passing those into the user > >>+space. This significantly accelerates DMA operations for PPC KVM guests. > >>+User space should expect that its handlers for these hypercalls > >>+are not going to be called if user space previously registered LIOBN > >>+in KVM (via KVM_CREATE_SPAPR_TCE or similar calls). > >>+ > >>+In order to enable H_PUT_TCE_INDIRECT and H_STUFF_TCE use in the guest, > >>+user space might have to advertise it for the guest. For example, > >>+IBM pSeries (sPAPR) guest starts using them if "hcall-multi-tce" is > >>+present in the "ibm,hypertas-functions" device-tree property. > >>+ > >>+The hypercalls mentioned above may or may not be processed successfully > >>+in the kernel based fast path. If they can not be handled by the kernel, > >>+they will get passed on to user space. So user space still has to have > >>+an implementation for these despite the in kernel acceleration. > >>+ > >>+This capability is always enabled. > >>+ > >> 5. The kvm_run structure > >> ------------------------ > >> > >>diff --git a/arch/powerpc/include/asm/kvm_ppc.h b/arch/powerpc/include/asm/kvm_ppc.h > >>index 9513911..4cadee5 100644 > >>--- a/arch/powerpc/include/asm/kvm_ppc.h > >>+++ b/arch/powerpc/include/asm/kvm_ppc.h > >>@@ -166,12 +166,24 @@ extern int kvmppc_pseries_do_hcall(struct kvm_vcpu *vcpu); > >> > >> extern long kvm_vm_ioctl_create_spapr_tce(struct kvm *kvm, > >> struct kvm_create_spapr_tce *args); > >>+extern struct kvmppc_spapr_tce_table *kvmppc_find_table( > >>+ struct kvm_vcpu *vcpu, unsigned long liobn); > >> extern long kvmppc_ioba_validate(struct kvmppc_spapr_tce_table *stt, > >> unsigned long ioba, unsigned long npages); > >> extern long kvmppc_tce_validate(struct kvmppc_spapr_tce_table *tt, > >> unsigned long tce); > >>+extern long kvmppc_gpa_to_ua(struct kvm *kvm, unsigned long gpa, > >>+ unsigned long *ua, unsigned long **prmap); > > > >Putting a userspace address into an unsigned long is pretty nasty: it > >should be a something __user *. > > > >>+extern void kvmppc_tce_put(struct kvmppc_spapr_tce_table *tt, > >>+ unsigned long idx, unsigned long tce); > >> extern long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, > >> unsigned long ioba, unsigned long tce); > >>+extern long kvmppc_h_put_tce_indirect(struct kvm_vcpu *vcpu, > >>+ unsigned long liobn, unsigned long ioba, > >>+ unsigned long tce_list, unsigned long npages); > >>+extern long kvmppc_h_stuff_tce(struct kvm_vcpu *vcpu, > >>+ unsigned long liobn, unsigned long ioba, > >>+ unsigned long tce_value, unsigned long npages); > >> extern long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn, > >> unsigned long ioba); > >> extern struct page *kvm_alloc_hpt(unsigned long nr_pages); > >>diff --git a/arch/powerpc/kvm/book3s_64_vio.c b/arch/powerpc/kvm/book3s_64_vio.c > >>index 975f0ab..987f406 100644 > >>--- a/arch/powerpc/kvm/book3s_64_vio.c > >>+++ b/arch/powerpc/kvm/book3s_64_vio.c > >>@@ -14,6 +14,7 @@ > >> * > >> * Copyright 2010 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> > >> * Copyright 2011 David Gibson, IBM Corporation <dwg@au1.ibm.com> > >>+ * Copyright 2016 Alexey Kardashevskiy, IBM Corporation <aik@au1.ibm.com> > >> */ > >> > >> #include <linux/types.h> > >>@@ -37,8 +38,7 @@ > >> #include <asm/kvm_host.h> > >> #include <asm/udbg.h> > >> #include <asm/iommu.h> > >>- > >>-#define TCES_PER_PAGE (PAGE_SIZE / sizeof(u64)) > >>+#include <asm/tce.h> > >> > >> static unsigned long kvmppc_stt_npages(unsigned long window_size) > >> { > >>@@ -200,3 +200,109 @@ fail: > >> } > >> return ret; > >> } > >>+ > >>+long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, > >>+ unsigned long liobn, unsigned long ioba, > >>+ unsigned long tce) > >>+{ > >>+ long ret; > >>+ struct kvmppc_spapr_tce_table *stt; > >>+ > >>+ stt = kvmppc_find_table(vcpu, liobn); > >>+ if (!stt) > >>+ return H_TOO_HARD; > >>+ > >>+ ret = kvmppc_ioba_validate(stt, ioba, 1); > >>+ if (ret != H_SUCCESS) > >>+ return ret; > >>+ > >>+ ret = kvmppc_tce_validate(stt, tce); > >>+ if (ret != H_SUCCESS) > >>+ return ret; > >>+ > >>+ kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce); > >>+ > >>+ return H_SUCCESS; > >>+} > >>+EXPORT_SYMBOL_GPL(kvmppc_h_put_tce); > >>+ > >>+long kvmppc_h_put_tce_indirect(struct kvm_vcpu *vcpu, > >>+ unsigned long liobn, unsigned long ioba, > >>+ unsigned long tce_list, unsigned long npages) > >>+{ > >>+ struct kvmppc_spapr_tce_table *stt; > >>+ long i, ret = H_SUCCESS, idx; > >>+ unsigned long entry, ua = 0; > >>+ u64 __user *tces, tce; > >>+ > >>+ stt = kvmppc_find_table(vcpu, liobn); > >>+ if (!stt) > >>+ return H_TOO_HARD; > >>+ > >>+ entry = ioba >> IOMMU_PAGE_SHIFT_4K; > >>+ /* > >>+ * SPAPR spec says that the maximum size of the list is 512 TCEs > >>+ * so the whole table fits in 4K page > >>+ */ > >>+ if (npages > 512) > >>+ return H_PARAMETER; > >>+ > >>+ if (tce_list & (SZ_4K - 1)) > >>+ return H_PARAMETER; > >>+ > >>+ ret = kvmppc_ioba_validate(stt, ioba, npages); > >>+ if (ret != H_SUCCESS) > >>+ return ret; > >>+ > >>+ idx = srcu_read_lock(&vcpu->kvm->srcu); > >>+ if (kvmppc_gpa_to_ua(vcpu->kvm, tce_list, &ua, NULL)) { > >>+ ret = H_TOO_HARD; > >>+ goto unlock_exit; > >>+ } > >>+ tces = (u64 __user *) ua; > >>+ > >>+ for (i = 0; i < npages; ++i) { > >>+ if (get_user(tce, tces + i)) { > >>+ ret = H_PARAMETER; > >>+ goto unlock_exit; > >>+ } > >>+ tce = be64_to_cpu(tce); > >>+ > >>+ ret = kvmppc_tce_validate(stt, tce); > >>+ if (ret != H_SUCCESS) > >>+ goto unlock_exit; > >>+ > >>+ kvmppc_tce_put(stt, entry + i, tce); > >>+ } > >>+ > >>+unlock_exit: > >>+ srcu_read_unlock(&vcpu->kvm->srcu, idx); > >>+ > >>+ return ret; > >>+} > >>+EXPORT_SYMBOL_GPL(kvmppc_h_put_tce_indirect); > >>+ > >>+long kvmppc_h_stuff_tce(struct kvm_vcpu *vcpu, > >>+ unsigned long liobn, unsigned long ioba, > >>+ unsigned long tce_value, unsigned long npages) > >>+{ > >>+ struct kvmppc_spapr_tce_table *stt; > >>+ long i, ret; > >>+ > >>+ stt = kvmppc_find_table(vcpu, liobn); > >>+ if (!stt) > >>+ return H_TOO_HARD; > >>+ > >>+ ret = kvmppc_ioba_validate(stt, ioba, npages); > >>+ if (ret != H_SUCCESS) > >>+ return ret; > >>+ > >>+ if (tce_value & (TCE_PCI_WRITE | TCE_PCI_READ)) > >>+ return H_PARAMETER; > > > >Do we really need to allow no-permission but non-zero TCEs? > > Not sure, for debugging purposes one could want to poison the table. Totally > useless? Hmm, I guess. Ok, leave it as is. > >>+ > >>+ for (i = 0; i < npages; ++i, ioba += IOMMU_PAGE_SIZE_4K) > >>+ kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce_value); > >>+ > >>+ return H_SUCCESS; > >>+} > >>+EXPORT_SYMBOL_GPL(kvmppc_h_stuff_tce); > >>diff --git a/arch/powerpc/kvm/book3s_64_vio_hv.c b/arch/powerpc/kvm/book3s_64_vio_hv.c > >>index 8cd3a95..58c63ed 100644 > >>--- a/arch/powerpc/kvm/book3s_64_vio_hv.c > >>+++ b/arch/powerpc/kvm/book3s_64_vio_hv.c > >>@@ -14,6 +14,7 @@ > >> * > >> * Copyright 2010 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> > >> * Copyright 2011 David Gibson, IBM Corporation <dwg@au1.ibm.com> > >>+ * Copyright 2016 Alexey Kardashevskiy, IBM Corporation <aik@au1.ibm.com> > >> */ > >> > >> #include <linux/types.h> > >>@@ -30,6 +31,7 @@ > >> #include <asm/kvm_ppc.h> > >> #include <asm/kvm_book3s.h> > >> #include <asm/mmu-hash64.h> > >>+#include <asm/mmu_context.h> > >> #include <asm/hvcall.h> > >> #include <asm/synch.h> > >> #include <asm/ppc-opcode.h> > >>@@ -37,6 +39,7 @@ > >> #include <asm/udbg.h> > >> #include <asm/iommu.h> > >> #include <asm/tce.h> > >>+#include <asm/iommu.h> > >> > >> #define TCES_PER_PAGE (PAGE_SIZE / sizeof(u64)) > >> > >>@@ -46,7 +49,7 @@ > >> * WARNING: This will be called in real or virtual mode on HV KVM and virtual > >> * mode on PR KVM > >> */ > >>-static struct kvmppc_spapr_tce_table *kvmppc_find_table(struct kvm_vcpu *vcpu, > >>+struct kvmppc_spapr_tce_table *kvmppc_find_table(struct kvm_vcpu *vcpu, > >> unsigned long liobn) > >> { > >> struct kvm *kvm = vcpu->kvm; > >>@@ -58,6 +61,7 @@ static struct kvmppc_spapr_tce_table *kvmppc_find_table(struct kvm_vcpu *vcpu, > >> > >> return NULL; > >> } > >>+EXPORT_SYMBOL_GPL(kvmppc_find_table); > >> > >> /* > >> * Validates IO address. > >>@@ -150,11 +154,31 @@ void kvmppc_tce_put(struct kvmppc_spapr_tce_table *stt, > >> } > >> EXPORT_SYMBOL_GPL(kvmppc_tce_put); > >> > >>-/* WARNING: This will be called in real-mode on HV KVM and virtual > >>- * mode on PR KVM > >>- */ > >>-long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, > >>- unsigned long ioba, unsigned long tce) > >>+long kvmppc_gpa_to_ua(struct kvm *kvm, unsigned long gpa, > >>+ unsigned long *ua, unsigned long **prmap) > >>+{ > >>+ unsigned long gfn = gpa >> PAGE_SHIFT; > >>+ struct kvm_memory_slot *memslot; > >>+ > >>+ memslot = search_memslots(kvm_memslots(kvm), gfn); > >>+ if (!memslot) > >>+ return -EINVAL; > >>+ > >>+ *ua = __gfn_to_hva_memslot(memslot, gfn) | > >>+ (gpa & ~(PAGE_MASK | TCE_PCI_READ | TCE_PCI_WRITE)); > >>+ > >>+#ifdef CONFIG_KVM_BOOK3S_HV_POSSIBLE > > > >It's a bit odd to see a test for HV_POSSIBLE in a file named > >book3s_64_vio_hv.c > > > True, the file name should have probably been changed book3s_64_vio_rm.c. > > > > > >>+ if (prmap) > >>+ *prmap = &memslot->arch.rmap[gfn - memslot->base_gfn]; > >>+#endif > >>+ > >>+ return 0; > >>+} > >>+EXPORT_SYMBOL_GPL(kvmppc_gpa_to_ua); > >>+ > >>+#ifdef CONFIG_KVM_BOOK3S_HV_POSSIBLE > >>+long kvmppc_rm_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, > >>+ unsigned long ioba, unsigned long tce) > >> { > >> struct kvmppc_spapr_tce_table *stt = kvmppc_find_table(vcpu, liobn); > >> long ret; > >>@@ -177,7 +201,112 @@ long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, > >> > >> return H_SUCCESS; > >> } > >>-EXPORT_SYMBOL_GPL(kvmppc_h_put_tce); > >>+ > >>+static long kvmppc_rm_ua_to_hpa(struct kvm_vcpu *vcpu, > >>+ unsigned long ua, unsigned long *phpa) > > > >ua should be a something __user * rather than an unsigned long. And > >come to that hpa should be a something * rather than an unsigned long. > > > @ua is the return type of __gfn_to_hva_memslot() so I kept it. Also, the > only place where I actually read from this address is the virtualmode's > H_PUT_TCE_INDIRECT handler, all other places just do translation with it so > making it "unsigned long" saves some type convertions. It is also used in > mm_iommu_ua_to_hpa() which is in upstream now (commit 15b244a88e1b289, part > of DDW and preregistration patchset). Still need to change it? Hmm, I guess not. > Regarding @phpa, the agreement here that we use "void *" for 0xC000.... type > of addresses; and we use "unsigned long" if top 4 bits are not set as > dereferencing such pointers will normally fail. Ah, yes, sorry, forgot that it was an HV physical addr, not an HV virtual addr. > > > > >>+{ > >>+ pte_t *ptep, pte; > >>+ unsigned shift = 0; > >>+ > >>+ ptep = __find_linux_pte_or_hugepte(vcpu->arch.pgdir, ua, NULL, &shift); > >>+ if (!ptep || !pte_present(*ptep)) > >>+ return -ENXIO; > >>+ pte = *ptep; > >>+ > >>+ if (!shift) > >>+ shift = PAGE_SHIFT; > >>+ > >>+ /* Avoid handling anything potentially complicated in realmode */ > >>+ if (shift > PAGE_SHIFT) > >>+ return -EAGAIN; > >>+ > >>+ if (!pte_young(pte)) > >>+ return -EAGAIN; > >>+ > >>+ *phpa = (pte_pfn(pte) << PAGE_SHIFT) | (ua & ((1ULL << shift) - 1)) | > >>+ (ua & ~PAGE_MASK); > >>+ > >>+ return 0; > >>+} > >>+ > >>+long kvmppc_rm_h_put_tce_indirect(struct kvm_vcpu *vcpu, > >>+ unsigned long liobn, unsigned long ioba, > >>+ unsigned long tce_list, unsigned long npages) > >>+{ > >>+ struct kvmppc_spapr_tce_table *stt; > >>+ long i, ret = H_SUCCESS; > >>+ unsigned long tces, entry, ua = 0; > >>+ unsigned long *rmap = NULL; > >>+ > >>+ stt = kvmppc_find_table(vcpu, liobn); > >>+ if (!stt) > >>+ return H_TOO_HARD; > >>+ > >>+ entry = ioba >> IOMMU_PAGE_SHIFT_4K; > >>+ /* > >>+ * The spec says that the maximum size of the list is 512 TCEs > >>+ * so the whole table addressed resides in 4K page > >>+ */ > >>+ if (npages > 512) > >>+ return H_PARAMETER; > >>+ > >>+ if (tce_list & (SZ_4K - 1)) > >>+ return H_PARAMETER; > >>+ > >>+ ret = kvmppc_ioba_validate(stt, ioba, npages); > >>+ if (ret != H_SUCCESS) > >>+ return ret; > >>+ > >>+ if (kvmppc_gpa_to_ua(vcpu->kvm, tce_list, &ua, &rmap)) > >>+ return H_TOO_HARD; > >>+ > >>+ rmap = (void *) vmalloc_to_phys(rmap); > >>+ > >>+ lock_rmap(rmap); > >>+ if (kvmppc_rm_ua_to_hpa(vcpu, ua, &tces)) { > >>+ ret = H_TOO_HARD; > >>+ goto unlock_exit; > >>+ } > >>+ > >>+ for (i = 0; i < npages; ++i) { > >>+ unsigned long tce = be64_to_cpu(((u64 *)tces)[i]); > >>+ > >>+ ret = kvmppc_tce_validate(stt, tce); > >>+ if (ret != H_SUCCESS) > >>+ goto unlock_exit; > >>+ > >>+ kvmppc_tce_put(stt, entry + i, tce); > >>+ } > >>+ > >>+unlock_exit: > >>+ unlock_rmap(rmap); > >>+ > >>+ return ret; > >>+} > >>+ > >>+long kvmppc_rm_h_stuff_tce(struct kvm_vcpu *vcpu, > >>+ unsigned long liobn, unsigned long ioba, > >>+ unsigned long tce_value, unsigned long npages) > > > >Unlike put_indirect, this code appears to be identical to the non > >realmode code - can you combine them? > > > It is at this point but this will get different bits in "KVM: PPC: vfio kvm > device: support spapr tce" later, may be sometime later I will manage to get > to that part, eventually... I think I'd prefer to see them split only in the patch that actually makes them different. > > > >>+{ > >>+ struct kvmppc_spapr_tce_table *stt; > >>+ long i, ret; > >>+ > >>+ stt = kvmppc_find_table(vcpu, liobn); > >>+ if (!stt) > >>+ return H_TOO_HARD; > >>+ > >>+ ret = kvmppc_ioba_validate(stt, ioba, npages); > >>+ if (ret != H_SUCCESS) > >>+ return ret; > >>+ > >>+ if (tce_value & (TCE_PCI_WRITE | TCE_PCI_READ)) > >>+ return H_PARAMETER; > >>+ > >>+ for (i = 0; i < npages; ++i, ioba += IOMMU_PAGE_SIZE_4K) > >>+ kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce_value); > >>+ > >>+ return H_SUCCESS; > >>+} > >> > >> long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn, > >> unsigned long ioba) > >>@@ -204,3 +333,5 @@ long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn, > >> return H_SUCCESS; > >> } > >> EXPORT_SYMBOL_GPL(kvmppc_h_get_tce); > >>+ > >>+#endif /* KVM_BOOK3S_HV_POSSIBLE */ > >>diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c > >>index cff207b..df3fbae 100644 > >>--- a/arch/powerpc/kvm/book3s_hv.c > >>+++ b/arch/powerpc/kvm/book3s_hv.c > >>@@ -768,7 +768,31 @@ int kvmppc_pseries_do_hcall(struct kvm_vcpu *vcpu) > >> if (kvmppc_xics_enabled(vcpu)) { > >> ret = kvmppc_xics_hcall(vcpu, req); > >> break; > >>- } /* fallthrough */ > >>+ } > >>+ return RESUME_HOST; > >>+ case H_PUT_TCE: > >>+ ret = kvmppc_h_put_tce(vcpu, kvmppc_get_gpr(vcpu, 4), > >>+ kvmppc_get_gpr(vcpu, 5), > >>+ kvmppc_get_gpr(vcpu, 6)); > >>+ if (ret == H_TOO_HARD) > >>+ return RESUME_HOST; > >>+ break; > >>+ case H_PUT_TCE_INDIRECT: > >>+ ret = kvmppc_h_put_tce_indirect(vcpu, kvmppc_get_gpr(vcpu, 4), > >>+ kvmppc_get_gpr(vcpu, 5), > >>+ kvmppc_get_gpr(vcpu, 6), > >>+ kvmppc_get_gpr(vcpu, 7)); > >>+ if (ret == H_TOO_HARD) > >>+ return RESUME_HOST; > >>+ break; > >>+ case H_STUFF_TCE: > >>+ ret = kvmppc_h_stuff_tce(vcpu, kvmppc_get_gpr(vcpu, 4), > >>+ kvmppc_get_gpr(vcpu, 5), > >>+ kvmppc_get_gpr(vcpu, 6), > >>+ kvmppc_get_gpr(vcpu, 7)); > >>+ if (ret == H_TOO_HARD) > >>+ return RESUME_HOST; > >>+ break; > >> default: > >> return RESUME_HOST; > >> } > >>diff --git a/arch/powerpc/kvm/book3s_hv_rmhandlers.S b/arch/powerpc/kvm/book3s_hv_rmhandlers.S > >>index 3c6badc..3bf6e72 100644 > >>--- a/arch/powerpc/kvm/book3s_hv_rmhandlers.S > >>+++ b/arch/powerpc/kvm/book3s_hv_rmhandlers.S > >>@@ -1928,7 +1928,7 @@ hcall_real_table: > >> .long DOTSYM(kvmppc_h_clear_ref) - hcall_real_table > >> .long DOTSYM(kvmppc_h_protect) - hcall_real_table > >> .long DOTSYM(kvmppc_h_get_tce) - hcall_real_table > >>- .long DOTSYM(kvmppc_h_put_tce) - hcall_real_table > >>+ .long DOTSYM(kvmppc_rm_h_put_tce) - hcall_real_table > >> .long 0 /* 0x24 - H_SET_SPRG0 */ > >> .long DOTSYM(kvmppc_h_set_dabr) - hcall_real_table > >> .long 0 /* 0x2c */ > >>@@ -2006,8 +2006,8 @@ hcall_real_table: > >> .long 0 /* 0x12c */ > >> .long 0 /* 0x130 */ > >> .long DOTSYM(kvmppc_h_set_xdabr) - hcall_real_table > >>- .long 0 /* 0x138 */ > >>- .long 0 /* 0x13c */ > >>+ .long DOTSYM(kvmppc_rm_h_stuff_tce) - hcall_real_table > >>+ .long DOTSYM(kvmppc_rm_h_put_tce_indirect) - hcall_real_table > >> .long 0 /* 0x140 */ > >> .long 0 /* 0x144 */ > >> .long 0 /* 0x148 */ > >>diff --git a/arch/powerpc/kvm/book3s_pr_papr.c b/arch/powerpc/kvm/book3s_pr_papr.c > >>index f2c75a1..02176fd 100644 > >>--- a/arch/powerpc/kvm/book3s_pr_papr.c > >>+++ b/arch/powerpc/kvm/book3s_pr_papr.c > >>@@ -280,6 +280,37 @@ static int kvmppc_h_pr_logical_ci_store(struct kvm_vcpu *vcpu) > >> return EMULATE_DONE; > >> } > >> > >>+static int kvmppc_h_pr_put_tce_indirect(struct kvm_vcpu *vcpu) > >>+{ > >>+ unsigned long liobn = kvmppc_get_gpr(vcpu, 4); > >>+ unsigned long ioba = kvmppc_get_gpr(vcpu, 5); > >>+ unsigned long tce = kvmppc_get_gpr(vcpu, 6); > >>+ unsigned long npages = kvmppc_get_gpr(vcpu, 7); > >>+ long rc; > >>+ > >>+ rc = kvmppc_h_put_tce_indirect(vcpu, liobn, ioba, > >>+ tce, npages); > >>+ if (rc == H_TOO_HARD) > >>+ return EMULATE_FAIL; > >>+ kvmppc_set_gpr(vcpu, 3, rc); > >>+ return EMULATE_DONE; > >>+} > >>+ > >>+static int kvmppc_h_pr_stuff_tce(struct kvm_vcpu *vcpu) > >>+{ > >>+ unsigned long liobn = kvmppc_get_gpr(vcpu, 4); > >>+ unsigned long ioba = kvmppc_get_gpr(vcpu, 5); > >>+ unsigned long tce_value = kvmppc_get_gpr(vcpu, 6); > >>+ unsigned long npages = kvmppc_get_gpr(vcpu, 7); > >>+ long rc; > >>+ > >>+ rc = kvmppc_h_stuff_tce(vcpu, liobn, ioba, tce_value, npages); > >>+ if (rc == H_TOO_HARD) > >>+ return EMULATE_FAIL; > >>+ kvmppc_set_gpr(vcpu, 3, rc); > >>+ return EMULATE_DONE; > >>+} > >>+ > >> static int kvmppc_h_pr_xics_hcall(struct kvm_vcpu *vcpu, u32 cmd) > >> { > >> long rc = kvmppc_xics_hcall(vcpu, cmd); > >>@@ -306,6 +337,10 @@ int kvmppc_h_pr(struct kvm_vcpu *vcpu, unsigned long cmd) > >> return kvmppc_h_pr_bulk_remove(vcpu); > >> case H_PUT_TCE: > >> return kvmppc_h_pr_put_tce(vcpu); > >>+ case H_PUT_TCE_INDIRECT: > >>+ return kvmppc_h_pr_put_tce_indirect(vcpu); > >>+ case H_STUFF_TCE: > >>+ return kvmppc_h_pr_stuff_tce(vcpu); > >> case H_CEDE: > >> kvmppc_set_msr_fast(vcpu, kvmppc_get_msr(vcpu) | MSR_EE); > >> kvm_vcpu_block(vcpu); > >>diff --git a/arch/powerpc/kvm/powerpc.c b/arch/powerpc/kvm/powerpc.c > >>index 6fd2405..164735c 100644 > >>--- a/arch/powerpc/kvm/powerpc.c > >>+++ b/arch/powerpc/kvm/powerpc.c > >>@@ -569,6 +569,9 @@ int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext) > >> case KVM_CAP_PPC_GET_SMMU_INFO: > >> r = 1; > >> break; > >>+ case KVM_CAP_SPAPR_MULTITCE: > >>+ r = 1; > >>+ break; > > > >Hmm, usual practice has been not to enable new KVM hcalls, unless > >userspace (qemu) explicitly enables them with ENABLE_HCALL. I don't > >see an obvious way this extension could break, but it's probably > >safest to continue that pattern. > > > This advertises the capability but does not enable it, this is still > required in QEMU: > > ret = kvm_vm_enable_cap(kvm_state, KVM_CAP_PPC_ENABLE_HCALL, 0, > H_PUT_TCE_INDIRECT, 1); > ret = kvm_vm_enable_cap(kvm_state, KVM_CAP_PPC_ENABLE_HCALL, 0, > H_STUFF_TCE, 1); > > as multi-tce hcalls are not in the default_hcall_list list in > arch/powerpc/kvm/book3s_hv.c. Ah, right, sorry.
On Thu, Jan 21, 2016 at 06:39:37PM +1100, Alexey Kardashevskiy wrote: > This adds real and virtual mode handlers for the H_PUT_TCE_INDIRECT and > H_STUFF_TCE hypercalls for user space emulated devices such as IBMVIO > devices or emulated PCI. These calls allow adding multiple entries > (up to 512) into the TCE table in one call which saves time on > transition between kernel and user space. > > This implements the KVM_CAP_PPC_MULTITCE capability. When present, > the kernel will try handling H_PUT_TCE_INDIRECT and H_STUFF_TCE. > If they can not be handled by the kernel, they are passed on to > the user space. The user space still has to have an implementation > for these. > > Both HV and PR-syle KVM are supported. > > Signed-off-by: Alexey Kardashevskiy <aik@ozlabs.ru> [snip] > diff --git a/arch/powerpc/kvm/book3s_64_vio.c b/arch/powerpc/kvm/book3s_64_vio.c > index 975f0ab..987f406 100644 > --- a/arch/powerpc/kvm/book3s_64_vio.c > +++ b/arch/powerpc/kvm/book3s_64_vio.c > @@ -14,6 +14,7 @@ > * > * Copyright 2010 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> > * Copyright 2011 David Gibson, IBM Corporation <dwg@au1.ibm.com> > + * Copyright 2016 Alexey Kardashevskiy, IBM Corporation <aik@au1.ibm.com> > */ > > #include <linux/types.h> > @@ -37,8 +38,7 @@ > #include <asm/kvm_host.h> > #include <asm/udbg.h> > #include <asm/iommu.h> > - > -#define TCES_PER_PAGE (PAGE_SIZE / sizeof(u64)) > +#include <asm/tce.h> > > static unsigned long kvmppc_stt_npages(unsigned long window_size) > { > @@ -200,3 +200,109 @@ fail: > } > return ret; > } > + > +long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, > + unsigned long liobn, unsigned long ioba, > + unsigned long tce) > +{ > + long ret; > + struct kvmppc_spapr_tce_table *stt; > + > + stt = kvmppc_find_table(vcpu, liobn); > + if (!stt) > + return H_TOO_HARD; > + > + ret = kvmppc_ioba_validate(stt, ioba, 1); > + if (ret != H_SUCCESS) > + return ret; > + > + ret = kvmppc_tce_validate(stt, tce); > + if (ret != H_SUCCESS) > + return ret; > + > + kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce); > + > + return H_SUCCESS; > +} > +EXPORT_SYMBOL_GPL(kvmppc_h_put_tce); As far as I can see, this is functionally identical to the kvmppc_h_put_tce that we have in book3s_64_vio_hv.c, that gets renamed later on in this patch. It would be good to have an explanation in the commit message why we want two almost-identical functions (and similarly for kvmppc_h_stuff_tce). Is it because a future patch is going to make them different, for instance? > +long kvmppc_h_stuff_tce(struct kvm_vcpu *vcpu, > + unsigned long liobn, unsigned long ioba, > + unsigned long tce_value, unsigned long npages) > +{ > + struct kvmppc_spapr_tce_table *stt; > + long i, ret; > + > + stt = kvmppc_find_table(vcpu, liobn); > + if (!stt) > + return H_TOO_HARD; > + > + ret = kvmppc_ioba_validate(stt, ioba, npages); > + if (ret != H_SUCCESS) > + return ret; > + > + if (tce_value & (TCE_PCI_WRITE | TCE_PCI_READ)) > + return H_PARAMETER; > + > + for (i = 0; i < npages; ++i, ioba += IOMMU_PAGE_SIZE_4K) Looks like we need a bounds check on npages, presumably in kvmppc_ioba_validate(). > diff --git a/arch/powerpc/kvm/book3s_64_vio_hv.c b/arch/powerpc/kvm/book3s_64_vio_hv.c > index 8cd3a95..58c63ed 100644 > --- a/arch/powerpc/kvm/book3s_64_vio_hv.c > +++ b/arch/powerpc/kvm/book3s_64_vio_hv.c [...] > +long kvmppc_rm_h_put_tce_indirect(struct kvm_vcpu *vcpu, > + unsigned long liobn, unsigned long ioba, > + unsigned long tce_list, unsigned long npages) > +{ > + struct kvmppc_spapr_tce_table *stt; > + long i, ret = H_SUCCESS; > + unsigned long tces, entry, ua = 0; > + unsigned long *rmap = NULL; > + > + stt = kvmppc_find_table(vcpu, liobn); > + if (!stt) > + return H_TOO_HARD; > + > + entry = ioba >> IOMMU_PAGE_SHIFT_4K; > + /* > + * The spec says that the maximum size of the list is 512 TCEs > + * so the whole table addressed resides in 4K page > + */ > + if (npages > 512) > + return H_PARAMETER; > + > + if (tce_list & (SZ_4K - 1)) > + return H_PARAMETER; > + > + ret = kvmppc_ioba_validate(stt, ioba, npages); > + if (ret != H_SUCCESS) > + return ret; > + > + if (kvmppc_gpa_to_ua(vcpu->kvm, tce_list, &ua, &rmap)) > + return H_TOO_HARD; > + > + rmap = (void *) vmalloc_to_phys(rmap); > + > + lock_rmap(rmap); A comment here explaining why we lock the rmap and what that achieves would be useful for future generations. > + if (kvmppc_rm_ua_to_hpa(vcpu, ua, &tces)) { > + ret = H_TOO_HARD; > + goto unlock_exit; > + } > + > + for (i = 0; i < npages; ++i) { > + unsigned long tce = be64_to_cpu(((u64 *)tces)[i]); > + > + ret = kvmppc_tce_validate(stt, tce); > + if (ret != H_SUCCESS) > + goto unlock_exit; > + > + kvmppc_tce_put(stt, entry + i, tce); > + } > + > +unlock_exit: > + unlock_rmap(rmap); > + > + return ret; > +} Paul. -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On 02/11/2016 04:32 PM, Paul Mackerras wrote: > On Thu, Jan 21, 2016 at 06:39:37PM +1100, Alexey Kardashevskiy wrote: >> This adds real and virtual mode handlers for the H_PUT_TCE_INDIRECT and >> H_STUFF_TCE hypercalls for user space emulated devices such as IBMVIO >> devices or emulated PCI. These calls allow adding multiple entries >> (up to 512) into the TCE table in one call which saves time on >> transition between kernel and user space. >> >> This implements the KVM_CAP_PPC_MULTITCE capability. When present, >> the kernel will try handling H_PUT_TCE_INDIRECT and H_STUFF_TCE. >> If they can not be handled by the kernel, they are passed on to >> the user space. The user space still has to have an implementation >> for these. >> >> Both HV and PR-syle KVM are supported. >> >> Signed-off-by: Alexey Kardashevskiy <aik@ozlabs.ru> > > [snip] > >> diff --git a/arch/powerpc/kvm/book3s_64_vio.c b/arch/powerpc/kvm/book3s_64_vio.c >> index 975f0ab..987f406 100644 >> --- a/arch/powerpc/kvm/book3s_64_vio.c >> +++ b/arch/powerpc/kvm/book3s_64_vio.c >> @@ -14,6 +14,7 @@ >> * >> * Copyright 2010 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> >> * Copyright 2011 David Gibson, IBM Corporation <dwg@au1.ibm.com> >> + * Copyright 2016 Alexey Kardashevskiy, IBM Corporation <aik@au1.ibm.com> >> */ >> >> #include <linux/types.h> >> @@ -37,8 +38,7 @@ >> #include <asm/kvm_host.h> >> #include <asm/udbg.h> >> #include <asm/iommu.h> >> - >> -#define TCES_PER_PAGE (PAGE_SIZE / sizeof(u64)) >> +#include <asm/tce.h> >> >> static unsigned long kvmppc_stt_npages(unsigned long window_size) >> { >> @@ -200,3 +200,109 @@ fail: >> } >> return ret; >> } >> + >> +long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, >> + unsigned long liobn, unsigned long ioba, >> + unsigned long tce) >> +{ >> + long ret; >> + struct kvmppc_spapr_tce_table *stt; >> + >> + stt = kvmppc_find_table(vcpu, liobn); >> + if (!stt) >> + return H_TOO_HARD; >> + >> + ret = kvmppc_ioba_validate(stt, ioba, 1); >> + if (ret != H_SUCCESS) >> + return ret; >> + >> + ret = kvmppc_tce_validate(stt, tce); >> + if (ret != H_SUCCESS) >> + return ret; >> + >> + kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce); >> + >> + return H_SUCCESS; >> +} >> +EXPORT_SYMBOL_GPL(kvmppc_h_put_tce); > > As far as I can see, this is functionally identical to the > kvmppc_h_put_tce that we have in book3s_64_vio_hv.c, that gets renamed > later on in this patch. It would be good to have an explanation in > the commit message why we want two almost-identical functions (and > similarly for kvmppc_h_stuff_tce). Is it because a future patch is > going to make them different, for instance? > >> +long kvmppc_h_stuff_tce(struct kvm_vcpu *vcpu, >> + unsigned long liobn, unsigned long ioba, >> + unsigned long tce_value, unsigned long npages) >> +{ >> + struct kvmppc_spapr_tce_table *stt; >> + long i, ret; >> + >> + stt = kvmppc_find_table(vcpu, liobn); >> + if (!stt) >> + return H_TOO_HARD; >> + >> + ret = kvmppc_ioba_validate(stt, ioba, npages); >> + if (ret != H_SUCCESS) >> + return ret; >> + >> + if (tce_value & (TCE_PCI_WRITE | TCE_PCI_READ)) >> + return H_PARAMETER; >> + >> + for (i = 0; i < npages; ++i, ioba += IOMMU_PAGE_SIZE_4K) > > Looks like we need a bounds check on npages, presumably in > kvmppc_ioba_validate(). > >> diff --git a/arch/powerpc/kvm/book3s_64_vio_hv.c b/arch/powerpc/kvm/book3s_64_vio_hv.c >> index 8cd3a95..58c63ed 100644 >> --- a/arch/powerpc/kvm/book3s_64_vio_hv.c >> +++ b/arch/powerpc/kvm/book3s_64_vio_hv.c > [...] >> +long kvmppc_rm_h_put_tce_indirect(struct kvm_vcpu *vcpu, >> + unsigned long liobn, unsigned long ioba, >> + unsigned long tce_list, unsigned long npages) >> +{ >> + struct kvmppc_spapr_tce_table *stt; >> + long i, ret = H_SUCCESS; >> + unsigned long tces, entry, ua = 0; >> + unsigned long *rmap = NULL; >> + >> + stt = kvmppc_find_table(vcpu, liobn); >> + if (!stt) >> + return H_TOO_HARD; >> + >> + entry = ioba >> IOMMU_PAGE_SHIFT_4K; >> + /* >> + * The spec says that the maximum size of the list is 512 TCEs >> + * so the whole table addressed resides in 4K page >> + */ >> + if (npages > 512) >> + return H_PARAMETER; >> + >> + if (tce_list & (SZ_4K - 1)) >> + return H_PARAMETER; >> + >> + ret = kvmppc_ioba_validate(stt, ioba, npages); >> + if (ret != H_SUCCESS) >> + return ret; >> + >> + if (kvmppc_gpa_to_ua(vcpu->kvm, tce_list, &ua, &rmap)) >> + return H_TOO_HARD; >> + >> + rmap = (void *) vmalloc_to_phys(rmap); >> + >> + lock_rmap(rmap); > > A comment here explaining why we lock the rmap and what that achieves > would be useful for future generations. /* This protects the guest page with the TCE list from going away while we are reading TCE list */ ? By "going away" I mean H_ENTER/H_REMOVE executed on parallel CPUs, is this roughly correct? as I did grep for "lock_rmap()" and did not find a single comment next to it... > >> + if (kvmppc_rm_ua_to_hpa(vcpu, ua, &tces)) { >> + ret = H_TOO_HARD; >> + goto unlock_exit; >> + } >> + >> + for (i = 0; i < npages; ++i) { >> + unsigned long tce = be64_to_cpu(((u64 *)tces)[i]); >> + >> + ret = kvmppc_tce_validate(stt, tce); >> + if (ret != H_SUCCESS) >> + goto unlock_exit; >> + >> + kvmppc_tce_put(stt, entry + i, tce); >> + } >> + >> +unlock_exit: >> + unlock_rmap(rmap); >> + >> + return ret; >> +} > > Paul. >
On Fri, Feb 12, 2016 at 03:54:18PM +1100, Alexey Kardashevskiy wrote: > On 02/11/2016 04:32 PM, Paul Mackerras wrote: > >On Thu, Jan 21, 2016 at 06:39:37PM +1100, Alexey Kardashevskiy wrote: > >>+ rmap = (void *) vmalloc_to_phys(rmap); > >>+ > >>+ lock_rmap(rmap); > > > >A comment here explaining why we lock the rmap and what that achieves > >would be useful for future generations. > > > /* This protects the guest page with the TCE list from going away while we > are reading TCE list */ > > ? > > By "going away" I mean H_ENTER/H_REMOVE executed on parallel CPUs, is this > roughly correct? as I did grep for "lock_rmap()" and did not find a single > comment next to it... Actually, taking the rmap lock stops the guest real -> host real mapping from changing. For the comment, I suggest this: /* * Synchronize with the MMU notifier callbacks in * book3s_64_mmu_hv.c (kvm_unmap_hva_hv etc.). * While we have the rmap lock, code running on other CPUs * cannot finish unmapping the host real page that backs * this guest real page, so we are OK to access the host * real page. */ Paul. -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
diff --git a/Documentation/virtual/kvm/api.txt b/Documentation/virtual/kvm/api.txt index 07e4cdf..da39435 100644 --- a/Documentation/virtual/kvm/api.txt +++ b/Documentation/virtual/kvm/api.txt @@ -3035,6 +3035,31 @@ Returns: 0 on success, -1 on error Queues an SMI on the thread's vcpu. +4.97 KVM_CAP_PPC_MULTITCE + +Capability: KVM_CAP_PPC_MULTITCE +Architectures: ppc +Type: vm + +This capability means the kernel is capable of handling hypercalls +H_PUT_TCE_INDIRECT and H_STUFF_TCE without passing those into the user +space. This significantly accelerates DMA operations for PPC KVM guests. +User space should expect that its handlers for these hypercalls +are not going to be called if user space previously registered LIOBN +in KVM (via KVM_CREATE_SPAPR_TCE or similar calls). + +In order to enable H_PUT_TCE_INDIRECT and H_STUFF_TCE use in the guest, +user space might have to advertise it for the guest. For example, +IBM pSeries (sPAPR) guest starts using them if "hcall-multi-tce" is +present in the "ibm,hypertas-functions" device-tree property. + +The hypercalls mentioned above may or may not be processed successfully +in the kernel based fast path. If they can not be handled by the kernel, +they will get passed on to user space. So user space still has to have +an implementation for these despite the in kernel acceleration. + +This capability is always enabled. + 5. The kvm_run structure ------------------------ diff --git a/arch/powerpc/include/asm/kvm_ppc.h b/arch/powerpc/include/asm/kvm_ppc.h index 9513911..4cadee5 100644 --- a/arch/powerpc/include/asm/kvm_ppc.h +++ b/arch/powerpc/include/asm/kvm_ppc.h @@ -166,12 +166,24 @@ extern int kvmppc_pseries_do_hcall(struct kvm_vcpu *vcpu); extern long kvm_vm_ioctl_create_spapr_tce(struct kvm *kvm, struct kvm_create_spapr_tce *args); +extern struct kvmppc_spapr_tce_table *kvmppc_find_table( + struct kvm_vcpu *vcpu, unsigned long liobn); extern long kvmppc_ioba_validate(struct kvmppc_spapr_tce_table *stt, unsigned long ioba, unsigned long npages); extern long kvmppc_tce_validate(struct kvmppc_spapr_tce_table *tt, unsigned long tce); +extern long kvmppc_gpa_to_ua(struct kvm *kvm, unsigned long gpa, + unsigned long *ua, unsigned long **prmap); +extern void kvmppc_tce_put(struct kvmppc_spapr_tce_table *tt, + unsigned long idx, unsigned long tce); extern long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, unsigned long ioba, unsigned long tce); +extern long kvmppc_h_put_tce_indirect(struct kvm_vcpu *vcpu, + unsigned long liobn, unsigned long ioba, + unsigned long tce_list, unsigned long npages); +extern long kvmppc_h_stuff_tce(struct kvm_vcpu *vcpu, + unsigned long liobn, unsigned long ioba, + unsigned long tce_value, unsigned long npages); extern long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn, unsigned long ioba); extern struct page *kvm_alloc_hpt(unsigned long nr_pages); diff --git a/arch/powerpc/kvm/book3s_64_vio.c b/arch/powerpc/kvm/book3s_64_vio.c index 975f0ab..987f406 100644 --- a/arch/powerpc/kvm/book3s_64_vio.c +++ b/arch/powerpc/kvm/book3s_64_vio.c @@ -14,6 +14,7 @@ * * Copyright 2010 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> * Copyright 2011 David Gibson, IBM Corporation <dwg@au1.ibm.com> + * Copyright 2016 Alexey Kardashevskiy, IBM Corporation <aik@au1.ibm.com> */ #include <linux/types.h> @@ -37,8 +38,7 @@ #include <asm/kvm_host.h> #include <asm/udbg.h> #include <asm/iommu.h> - -#define TCES_PER_PAGE (PAGE_SIZE / sizeof(u64)) +#include <asm/tce.h> static unsigned long kvmppc_stt_npages(unsigned long window_size) { @@ -200,3 +200,109 @@ fail: } return ret; } + +long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, + unsigned long liobn, unsigned long ioba, + unsigned long tce) +{ + long ret; + struct kvmppc_spapr_tce_table *stt; + + stt = kvmppc_find_table(vcpu, liobn); + if (!stt) + return H_TOO_HARD; + + ret = kvmppc_ioba_validate(stt, ioba, 1); + if (ret != H_SUCCESS) + return ret; + + ret = kvmppc_tce_validate(stt, tce); + if (ret != H_SUCCESS) + return ret; + + kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce); + + return H_SUCCESS; +} +EXPORT_SYMBOL_GPL(kvmppc_h_put_tce); + +long kvmppc_h_put_tce_indirect(struct kvm_vcpu *vcpu, + unsigned long liobn, unsigned long ioba, + unsigned long tce_list, unsigned long npages) +{ + struct kvmppc_spapr_tce_table *stt; + long i, ret = H_SUCCESS, idx; + unsigned long entry, ua = 0; + u64 __user *tces, tce; + + stt = kvmppc_find_table(vcpu, liobn); + if (!stt) + return H_TOO_HARD; + + entry = ioba >> IOMMU_PAGE_SHIFT_4K; + /* + * SPAPR spec says that the maximum size of the list is 512 TCEs + * so the whole table fits in 4K page + */ + if (npages > 512) + return H_PARAMETER; + + if (tce_list & (SZ_4K - 1)) + return H_PARAMETER; + + ret = kvmppc_ioba_validate(stt, ioba, npages); + if (ret != H_SUCCESS) + return ret; + + idx = srcu_read_lock(&vcpu->kvm->srcu); + if (kvmppc_gpa_to_ua(vcpu->kvm, tce_list, &ua, NULL)) { + ret = H_TOO_HARD; + goto unlock_exit; + } + tces = (u64 __user *) ua; + + for (i = 0; i < npages; ++i) { + if (get_user(tce, tces + i)) { + ret = H_PARAMETER; + goto unlock_exit; + } + tce = be64_to_cpu(tce); + + ret = kvmppc_tce_validate(stt, tce); + if (ret != H_SUCCESS) + goto unlock_exit; + + kvmppc_tce_put(stt, entry + i, tce); + } + +unlock_exit: + srcu_read_unlock(&vcpu->kvm->srcu, idx); + + return ret; +} +EXPORT_SYMBOL_GPL(kvmppc_h_put_tce_indirect); + +long kvmppc_h_stuff_tce(struct kvm_vcpu *vcpu, + unsigned long liobn, unsigned long ioba, + unsigned long tce_value, unsigned long npages) +{ + struct kvmppc_spapr_tce_table *stt; + long i, ret; + + stt = kvmppc_find_table(vcpu, liobn); + if (!stt) + return H_TOO_HARD; + + ret = kvmppc_ioba_validate(stt, ioba, npages); + if (ret != H_SUCCESS) + return ret; + + if (tce_value & (TCE_PCI_WRITE | TCE_PCI_READ)) + return H_PARAMETER; + + for (i = 0; i < npages; ++i, ioba += IOMMU_PAGE_SIZE_4K) + kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce_value); + + return H_SUCCESS; +} +EXPORT_SYMBOL_GPL(kvmppc_h_stuff_tce); diff --git a/arch/powerpc/kvm/book3s_64_vio_hv.c b/arch/powerpc/kvm/book3s_64_vio_hv.c index 8cd3a95..58c63ed 100644 --- a/arch/powerpc/kvm/book3s_64_vio_hv.c +++ b/arch/powerpc/kvm/book3s_64_vio_hv.c @@ -14,6 +14,7 @@ * * Copyright 2010 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> * Copyright 2011 David Gibson, IBM Corporation <dwg@au1.ibm.com> + * Copyright 2016 Alexey Kardashevskiy, IBM Corporation <aik@au1.ibm.com> */ #include <linux/types.h> @@ -30,6 +31,7 @@ #include <asm/kvm_ppc.h> #include <asm/kvm_book3s.h> #include <asm/mmu-hash64.h> +#include <asm/mmu_context.h> #include <asm/hvcall.h> #include <asm/synch.h> #include <asm/ppc-opcode.h> @@ -37,6 +39,7 @@ #include <asm/udbg.h> #include <asm/iommu.h> #include <asm/tce.h> +#include <asm/iommu.h> #define TCES_PER_PAGE (PAGE_SIZE / sizeof(u64)) @@ -46,7 +49,7 @@ * WARNING: This will be called in real or virtual mode on HV KVM and virtual * mode on PR KVM */ -static struct kvmppc_spapr_tce_table *kvmppc_find_table(struct kvm_vcpu *vcpu, +struct kvmppc_spapr_tce_table *kvmppc_find_table(struct kvm_vcpu *vcpu, unsigned long liobn) { struct kvm *kvm = vcpu->kvm; @@ -58,6 +61,7 @@ static struct kvmppc_spapr_tce_table *kvmppc_find_table(struct kvm_vcpu *vcpu, return NULL; } +EXPORT_SYMBOL_GPL(kvmppc_find_table); /* * Validates IO address. @@ -150,11 +154,31 @@ void kvmppc_tce_put(struct kvmppc_spapr_tce_table *stt, } EXPORT_SYMBOL_GPL(kvmppc_tce_put); -/* WARNING: This will be called in real-mode on HV KVM and virtual - * mode on PR KVM - */ -long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, - unsigned long ioba, unsigned long tce) +long kvmppc_gpa_to_ua(struct kvm *kvm, unsigned long gpa, + unsigned long *ua, unsigned long **prmap) +{ + unsigned long gfn = gpa >> PAGE_SHIFT; + struct kvm_memory_slot *memslot; + + memslot = search_memslots(kvm_memslots(kvm), gfn); + if (!memslot) + return -EINVAL; + + *ua = __gfn_to_hva_memslot(memslot, gfn) | + (gpa & ~(PAGE_MASK | TCE_PCI_READ | TCE_PCI_WRITE)); + +#ifdef CONFIG_KVM_BOOK3S_HV_POSSIBLE + if (prmap) + *prmap = &memslot->arch.rmap[gfn - memslot->base_gfn]; +#endif + + return 0; +} +EXPORT_SYMBOL_GPL(kvmppc_gpa_to_ua); + +#ifdef CONFIG_KVM_BOOK3S_HV_POSSIBLE +long kvmppc_rm_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, + unsigned long ioba, unsigned long tce) { struct kvmppc_spapr_tce_table *stt = kvmppc_find_table(vcpu, liobn); long ret; @@ -177,7 +201,112 @@ long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, return H_SUCCESS; } -EXPORT_SYMBOL_GPL(kvmppc_h_put_tce); + +static long kvmppc_rm_ua_to_hpa(struct kvm_vcpu *vcpu, + unsigned long ua, unsigned long *phpa) +{ + pte_t *ptep, pte; + unsigned shift = 0; + + ptep = __find_linux_pte_or_hugepte(vcpu->arch.pgdir, ua, NULL, &shift); + if (!ptep || !pte_present(*ptep)) + return -ENXIO; + pte = *ptep; + + if (!shift) + shift = PAGE_SHIFT; + + /* Avoid handling anything potentially complicated in realmode */ + if (shift > PAGE_SHIFT) + return -EAGAIN; + + if (!pte_young(pte)) + return -EAGAIN; + + *phpa = (pte_pfn(pte) << PAGE_SHIFT) | (ua & ((1ULL << shift) - 1)) | + (ua & ~PAGE_MASK); + + return 0; +} + +long kvmppc_rm_h_put_tce_indirect(struct kvm_vcpu *vcpu, + unsigned long liobn, unsigned long ioba, + unsigned long tce_list, unsigned long npages) +{ + struct kvmppc_spapr_tce_table *stt; + long i, ret = H_SUCCESS; + unsigned long tces, entry, ua = 0; + unsigned long *rmap = NULL; + + stt = kvmppc_find_table(vcpu, liobn); + if (!stt) + return H_TOO_HARD; + + entry = ioba >> IOMMU_PAGE_SHIFT_4K; + /* + * The spec says that the maximum size of the list is 512 TCEs + * so the whole table addressed resides in 4K page + */ + if (npages > 512) + return H_PARAMETER; + + if (tce_list & (SZ_4K - 1)) + return H_PARAMETER; + + ret = kvmppc_ioba_validate(stt, ioba, npages); + if (ret != H_SUCCESS) + return ret; + + if (kvmppc_gpa_to_ua(vcpu->kvm, tce_list, &ua, &rmap)) + return H_TOO_HARD; + + rmap = (void *) vmalloc_to_phys(rmap); + + lock_rmap(rmap); + if (kvmppc_rm_ua_to_hpa(vcpu, ua, &tces)) { + ret = H_TOO_HARD; + goto unlock_exit; + } + + for (i = 0; i < npages; ++i) { + unsigned long tce = be64_to_cpu(((u64 *)tces)[i]); + + ret = kvmppc_tce_validate(stt, tce); + if (ret != H_SUCCESS) + goto unlock_exit; + + kvmppc_tce_put(stt, entry + i, tce); + } + +unlock_exit: + unlock_rmap(rmap); + + return ret; +} + +long kvmppc_rm_h_stuff_tce(struct kvm_vcpu *vcpu, + unsigned long liobn, unsigned long ioba, + unsigned long tce_value, unsigned long npages) +{ + struct kvmppc_spapr_tce_table *stt; + long i, ret; + + stt = kvmppc_find_table(vcpu, liobn); + if (!stt) + return H_TOO_HARD; + + ret = kvmppc_ioba_validate(stt, ioba, npages); + if (ret != H_SUCCESS) + return ret; + + if (tce_value & (TCE_PCI_WRITE | TCE_PCI_READ)) + return H_PARAMETER; + + for (i = 0; i < npages; ++i, ioba += IOMMU_PAGE_SIZE_4K) + kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce_value); + + return H_SUCCESS; +} long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn, unsigned long ioba) @@ -204,3 +333,5 @@ long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn, return H_SUCCESS; } EXPORT_SYMBOL_GPL(kvmppc_h_get_tce); + +#endif /* KVM_BOOK3S_HV_POSSIBLE */ diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c index cff207b..df3fbae 100644 --- a/arch/powerpc/kvm/book3s_hv.c +++ b/arch/powerpc/kvm/book3s_hv.c @@ -768,7 +768,31 @@ int kvmppc_pseries_do_hcall(struct kvm_vcpu *vcpu) if (kvmppc_xics_enabled(vcpu)) { ret = kvmppc_xics_hcall(vcpu, req); break; - } /* fallthrough */ + } + return RESUME_HOST; + case H_PUT_TCE: + ret = kvmppc_h_put_tce(vcpu, kvmppc_get_gpr(vcpu, 4), + kvmppc_get_gpr(vcpu, 5), + kvmppc_get_gpr(vcpu, 6)); + if (ret == H_TOO_HARD) + return RESUME_HOST; + break; + case H_PUT_TCE_INDIRECT: + ret = kvmppc_h_put_tce_indirect(vcpu, kvmppc_get_gpr(vcpu, 4), + kvmppc_get_gpr(vcpu, 5), + kvmppc_get_gpr(vcpu, 6), + kvmppc_get_gpr(vcpu, 7)); + if (ret == H_TOO_HARD) + return RESUME_HOST; + break; + case H_STUFF_TCE: + ret = kvmppc_h_stuff_tce(vcpu, kvmppc_get_gpr(vcpu, 4), + kvmppc_get_gpr(vcpu, 5), + kvmppc_get_gpr(vcpu, 6), + kvmppc_get_gpr(vcpu, 7)); + if (ret == H_TOO_HARD) + return RESUME_HOST; + break; default: return RESUME_HOST; } diff --git a/arch/powerpc/kvm/book3s_hv_rmhandlers.S b/arch/powerpc/kvm/book3s_hv_rmhandlers.S index 3c6badc..3bf6e72 100644 --- a/arch/powerpc/kvm/book3s_hv_rmhandlers.S +++ b/arch/powerpc/kvm/book3s_hv_rmhandlers.S @@ -1928,7 +1928,7 @@ hcall_real_table: .long DOTSYM(kvmppc_h_clear_ref) - hcall_real_table .long DOTSYM(kvmppc_h_protect) - hcall_real_table .long DOTSYM(kvmppc_h_get_tce) - hcall_real_table - .long DOTSYM(kvmppc_h_put_tce) - hcall_real_table + .long DOTSYM(kvmppc_rm_h_put_tce) - hcall_real_table .long 0 /* 0x24 - H_SET_SPRG0 */ .long DOTSYM(kvmppc_h_set_dabr) - hcall_real_table .long 0 /* 0x2c */ @@ -2006,8 +2006,8 @@ hcall_real_table: .long 0 /* 0x12c */ .long 0 /* 0x130 */ .long DOTSYM(kvmppc_h_set_xdabr) - hcall_real_table - .long 0 /* 0x138 */ - .long 0 /* 0x13c */ + .long DOTSYM(kvmppc_rm_h_stuff_tce) - hcall_real_table + .long DOTSYM(kvmppc_rm_h_put_tce_indirect) - hcall_real_table .long 0 /* 0x140 */ .long 0 /* 0x144 */ .long 0 /* 0x148 */ diff --git a/arch/powerpc/kvm/book3s_pr_papr.c b/arch/powerpc/kvm/book3s_pr_papr.c index f2c75a1..02176fd 100644 --- a/arch/powerpc/kvm/book3s_pr_papr.c +++ b/arch/powerpc/kvm/book3s_pr_papr.c @@ -280,6 +280,37 @@ static int kvmppc_h_pr_logical_ci_store(struct kvm_vcpu *vcpu) return EMULATE_DONE; } +static int kvmppc_h_pr_put_tce_indirect(struct kvm_vcpu *vcpu) +{ + unsigned long liobn = kvmppc_get_gpr(vcpu, 4); + unsigned long ioba = kvmppc_get_gpr(vcpu, 5); + unsigned long tce = kvmppc_get_gpr(vcpu, 6); + unsigned long npages = kvmppc_get_gpr(vcpu, 7); + long rc; + + rc = kvmppc_h_put_tce_indirect(vcpu, liobn, ioba, + tce, npages); + if (rc == H_TOO_HARD) + return EMULATE_FAIL; + kvmppc_set_gpr(vcpu, 3, rc); + return EMULATE_DONE; +} + +static int kvmppc_h_pr_stuff_tce(struct kvm_vcpu *vcpu) +{ + unsigned long liobn = kvmppc_get_gpr(vcpu, 4); + unsigned long ioba = kvmppc_get_gpr(vcpu, 5); + unsigned long tce_value = kvmppc_get_gpr(vcpu, 6); + unsigned long npages = kvmppc_get_gpr(vcpu, 7); + long rc; + + rc = kvmppc_h_stuff_tce(vcpu, liobn, ioba, tce_value, npages); + if (rc == H_TOO_HARD) + return EMULATE_FAIL; + kvmppc_set_gpr(vcpu, 3, rc); + return EMULATE_DONE; +} + static int kvmppc_h_pr_xics_hcall(struct kvm_vcpu *vcpu, u32 cmd) { long rc = kvmppc_xics_hcall(vcpu, cmd); @@ -306,6 +337,10 @@ int kvmppc_h_pr(struct kvm_vcpu *vcpu, unsigned long cmd) return kvmppc_h_pr_bulk_remove(vcpu); case H_PUT_TCE: return kvmppc_h_pr_put_tce(vcpu); + case H_PUT_TCE_INDIRECT: + return kvmppc_h_pr_put_tce_indirect(vcpu); + case H_STUFF_TCE: + return kvmppc_h_pr_stuff_tce(vcpu); case H_CEDE: kvmppc_set_msr_fast(vcpu, kvmppc_get_msr(vcpu) | MSR_EE); kvm_vcpu_block(vcpu); diff --git a/arch/powerpc/kvm/powerpc.c b/arch/powerpc/kvm/powerpc.c index 6fd2405..164735c 100644 --- a/arch/powerpc/kvm/powerpc.c +++ b/arch/powerpc/kvm/powerpc.c @@ -569,6 +569,9 @@ int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext) case KVM_CAP_PPC_GET_SMMU_INFO: r = 1; break; + case KVM_CAP_SPAPR_MULTITCE: + r = 1; + break; #endif default: r = 0;
This adds real and virtual mode handlers for the H_PUT_TCE_INDIRECT and H_STUFF_TCE hypercalls for user space emulated devices such as IBMVIO devices or emulated PCI. These calls allow adding multiple entries (up to 512) into the TCE table in one call which saves time on transition between kernel and user space. This implements the KVM_CAP_PPC_MULTITCE capability. When present, the kernel will try handling H_PUT_TCE_INDIRECT and H_STUFF_TCE. If they can not be handled by the kernel, they are passed on to the user space. The user space still has to have an implementation for these. Both HV and PR-syle KVM are supported. Signed-off-by: Alexey Kardashevskiy <aik@ozlabs.ru> --- Changes: v2: * compare @ret with H_SUCCESS instead of assuming H_SUCCESS is zero * s/~IOMMU_PAGE_MASK_4K/SZ_4K-1/ when testing @tce_list --- Documentation/virtual/kvm/api.txt | 25 ++++++ arch/powerpc/include/asm/kvm_ppc.h | 12 +++ arch/powerpc/kvm/book3s_64_vio.c | 110 +++++++++++++++++++++++- arch/powerpc/kvm/book3s_64_vio_hv.c | 145 ++++++++++++++++++++++++++++++-- arch/powerpc/kvm/book3s_hv.c | 26 +++++- arch/powerpc/kvm/book3s_hv_rmhandlers.S | 6 +- arch/powerpc/kvm/book3s_pr_papr.c | 35 ++++++++ arch/powerpc/kvm/powerpc.c | 3 + 8 files changed, 349 insertions(+), 13 deletions(-)