@@ -377,6 +377,10 @@ extern s64 kvm_nvhe_sym(hyp_physvirt_offset);
extern u64 kvm_nvhe_sym(hyp_cpu_logical_map)[NR_CPUS];
#define hyp_cpu_logical_map CHOOSE_NVHE_SYM(hyp_cpu_logical_map)
+enum kvm_iommu_driver {
+ KVM_IOMMU_DRIVER_NONE,
+};
+
struct vcpu_reset_state {
unsigned long pc;
unsigned long r0;
@@ -114,7 +114,8 @@ void __noreturn __hyp_do_panic(struct kvm_cpu_context *host_ctxt, u64 spsr,
void __pkvm_init_switch_pgd(phys_addr_t phys, unsigned long size,
phys_addr_t pgd, void *sp, void *cont_fn);
int __pkvm_init(phys_addr_t phys, unsigned long size, unsigned long nr_cpus,
- unsigned long *per_cpu_base, u32 hyp_va_bits);
+ unsigned long *per_cpu_base, u32 hyp_va_bits,
+ enum kvm_iommu_driver iommu_driver);
void __noreturn __host_enter(struct kvm_cpu_context *host_ctxt);
#endif
new file mode 100644
@@ -0,0 +1,11 @@
+/* SPDX-License-Identifier: GPL-2.0 */
+#ifndef __ARM64_KVM_NVHE_IOMMU_H__
+#define __ARM64_KVM_NVHE_IOMMU_H__
+
+struct kvm_iommu_ops {
+ int (*init)(void);
+};
+
+extern struct kvm_iommu_ops kvm_iommu_ops;
+
+#endif /* __ARM64_KVM_NVHE_IOMMU_H__ */
@@ -1899,6 +1899,15 @@ static bool init_psci_relay(void)
return true;
}
+static int init_stage2_iommu(void)
+{
+ return KVM_IOMMU_DRIVER_NONE;
+}
+
+static void remove_stage2_iommu(enum kvm_iommu_driver iommu)
+{
+}
+
static int init_subsystems(void)
{
int err = 0;
@@ -1957,7 +1966,7 @@ static void teardown_hyp_mode(void)
}
}
-static int do_pkvm_init(u32 hyp_va_bits)
+static int do_pkvm_init(u32 hyp_va_bits, enum kvm_iommu_driver iommu_driver)
{
void *per_cpu_base = kvm_ksym_ref(kvm_nvhe_sym(kvm_arm_hyp_percpu_base));
int ret;
@@ -1966,7 +1975,7 @@ static int do_pkvm_init(u32 hyp_va_bits)
cpu_hyp_init_context();
ret = kvm_call_hyp_nvhe(__pkvm_init, hyp_mem_base, hyp_mem_size,
num_possible_cpus(), kern_hyp_va(per_cpu_base),
- hyp_va_bits);
+ hyp_va_bits, iommu_driver);
cpu_hyp_init_features();
/*
@@ -1996,15 +2005,23 @@ static void kvm_hyp_init_symbols(void)
static int kvm_hyp_init_protection(u32 hyp_va_bits)
{
void *addr = phys_to_virt(hyp_mem_base);
+ enum kvm_iommu_driver iommu;
int ret;
ret = create_hyp_mappings(addr, addr + hyp_mem_size, PAGE_HYP);
if (ret)
return ret;
- ret = do_pkvm_init(hyp_va_bits);
- if (ret)
+ ret = init_stage2_iommu();
+ if (ret < 0)
return ret;
+ iommu = ret;
+
+ ret = do_pkvm_init(hyp_va_bits, iommu);
+ if (ret) {
+ remove_stage2_iommu(iommu);
+ return ret;
+ }
free_hyp_pgds();
@@ -14,6 +14,7 @@
#include <asm/kvm_host.h>
#include <asm/kvm_hyp.h>
+#include <nvhe/iommu.h>
#include <nvhe/mem_protect.h>
#include <nvhe/mm.h>
#include <nvhe/pkvm.h>
@@ -34,6 +35,8 @@ static DEFINE_PER_CPU(struct user_fpsimd_state, loaded_host_fpsimd_state);
DEFINE_PER_CPU(struct kvm_nvhe_init_params, kvm_init_params);
+struct kvm_iommu_ops kvm_iommu_ops;
+
void __kvm_hyp_host_forward_smc(struct kvm_cpu_context *host_ctxt);
typedef void (*hyp_entry_exit_handler_fn)(struct pkvm_hyp_vcpu *);
@@ -958,6 +961,7 @@ static void handle___pkvm_init(struct kvm_cpu_context *host_ctxt)
DECLARE_REG(unsigned long, nr_cpus, host_ctxt, 3);
DECLARE_REG(unsigned long *, per_cpu_base, host_ctxt, 4);
DECLARE_REG(u32, hyp_va_bits, host_ctxt, 5);
+ DECLARE_REG(enum kvm_iommu_driver, iommu_driver, host_ctxt, 6);
/*
* __pkvm_init() will return only if an error occurred, otherwise it
@@ -965,7 +969,7 @@ static void handle___pkvm_init(struct kvm_cpu_context *host_ctxt)
* with the host context directly.
*/
cpu_reg(host_ctxt, 1) = __pkvm_init(phys, size, nr_cpus, per_cpu_base,
- hyp_va_bits);
+ hyp_va_bits, iommu_driver);
}
static void handle___pkvm_cpu_set_vector(struct kvm_cpu_context *host_ctxt)
@@ -11,6 +11,7 @@
#include <nvhe/early_alloc.h>
#include <nvhe/gfp.h>
+#include <nvhe/iommu.h>
#include <nvhe/memory.h>
#include <nvhe/mem_protect.h>
#include <nvhe/mm.h>
@@ -288,6 +289,16 @@ static int fix_hyp_pgtable_refcnt(void)
&walker);
}
+static int select_iommu_ops(enum kvm_iommu_driver driver)
+{
+ switch (driver) {
+ case KVM_IOMMU_DRIVER_NONE:
+ return 0;
+ }
+
+ return -EINVAL;
+}
+
void __noreturn __pkvm_init_finalise(void)
{
struct kvm_host_data *host_data = this_cpu_ptr(&kvm_host_data);
@@ -321,6 +332,12 @@ void __noreturn __pkvm_init_finalise(void)
if (ret)
goto out;
+ if (kvm_iommu_ops.init) {
+ ret = kvm_iommu_ops.init();
+ if (ret)
+ goto out;
+ }
+
ret = fix_host_ownership();
if (ret)
goto out;
@@ -345,7 +362,8 @@ void __noreturn __pkvm_init_finalise(void)
}
int __pkvm_init(phys_addr_t phys, unsigned long size, unsigned long nr_cpus,
- unsigned long *per_cpu_base, u32 hyp_va_bits)
+ unsigned long *per_cpu_base, u32 hyp_va_bits,
+ enum kvm_iommu_driver iommu_driver)
{
struct kvm_nvhe_init_params *params;
void *virt = hyp_phys_to_virt(phys);
@@ -368,6 +386,10 @@ int __pkvm_init(phys_addr_t phys, unsigned long size, unsigned long nr_cpus,
if (ret)
return ret;
+ ret = select_iommu_ops(iommu_driver);
+ if (ret)
+ return ret;
+
update_nvhe_init_params();
/* Jump in the idmap page to switch to the new page-tables */