@@ -194,7 +194,7 @@ enum kvm_pgtable_walk_flags {
typedef int (*kvm_pgtable_visitor_fn_t)(u64 addr, u64 end, u32 level,
kvm_pte_t *ptep, kvm_pte_t *old,
enum kvm_pgtable_walk_flags flag,
- void * const arg);
+ void * const arg, bool shared);
/**
* struct kvm_pgtable_walker - Hook into a page-table walk.
@@ -490,6 +490,7 @@ int kvm_pgtable_stage2_flush(struct kvm_pgtable *pgt, u64 addr, u64 size);
* @addr: Input address for the start of the walk.
* @size: Size of the range to walk.
* @walker: Walker callback description.
+ * @shared: Indicates if the page table walk can be done in parallel
*
* The offset of @addr within a page is ignored and @size is rounded-up to
* the next page boundary.
@@ -506,7 +507,7 @@ int kvm_pgtable_stage2_flush(struct kvm_pgtable *pgt, u64 addr, u64 size);
* Return: 0 on success, negative error code on failure.
*/
int kvm_pgtable_walk(struct kvm_pgtable *pgt, u64 addr, u64 size,
- struct kvm_pgtable_walker *walker);
+ struct kvm_pgtable_walker *walker, bool shared);
/**
* kvm_pgtable_get_leaf() - Walk a page-table and retrieve the leaf entry
@@ -424,7 +424,7 @@ struct check_walk_data {
static int __check_page_state_visitor(u64 addr, u64 end, u32 level,
kvm_pte_t *ptep, kvm_pte_t *old,
enum kvm_pgtable_walk_flags flag,
- void * const arg)
+ void * const arg, bool shared)
{
struct check_walk_data *d = arg;
@@ -443,7 +443,7 @@ static int check_page_state_range(struct kvm_pgtable *pgt, u64 addr, u64 size,
.flags = KVM_PGTABLE_WALK_LEAF,
};
- return kvm_pgtable_walk(pgt, addr, size, &walker);
+ return kvm_pgtable_walk(pgt, addr, size, &walker, false);
}
static enum pkvm_page_state host_get_page_state(kvm_pte_t pte)
@@ -164,7 +164,7 @@ static void hpool_put_page(void *addr)
static int finalize_host_mappings_walker(u64 addr, u64 end, u32 level,
kvm_pte_t *ptep, kvm_pte_t *old,
enum kvm_pgtable_walk_flags flag,
- void * const arg)
+ void * const arg, bool shared)
{
struct kvm_pgtable_mm_ops *mm_ops = arg;
enum kvm_pgtable_prot prot;
@@ -224,7 +224,7 @@ static int finalize_host_mappings(void)
struct memblock_region *reg = &hyp_memory[i];
u64 start = (u64)hyp_phys_to_virt(reg->base);
- ret = kvm_pgtable_walk(&pkvm_pgtable, start, reg->size, &walker);
+ ret = kvm_pgtable_walk(&pkvm_pgtable, start, reg->size, &walker, false);
if (ret)
return ret;
}
@@ -198,17 +198,17 @@ static u8 kvm_invalid_pte_owner(kvm_pte_t pte)
static int kvm_pgtable_visitor_cb(struct kvm_pgtable_walk_data *data, u64 addr,
u32 level, kvm_pte_t *ptep, kvm_pte_t *old,
- enum kvm_pgtable_walk_flags flag)
+ enum kvm_pgtable_walk_flags flag, bool shared)
{
struct kvm_pgtable_walker *walker = data->walker;
- return walker->cb(addr, data->end, level, ptep, old, flag, walker->arg);
+ return walker->cb(addr, data->end, level, ptep, old, flag, walker->arg, shared);
}
static int __kvm_pgtable_walk(struct kvm_pgtable_walk_data *data,
- kvm_pte_t *pgtable, u32 level);
+ kvm_pte_t *pgtable, u32 level, bool shared);
static inline int __kvm_pgtable_visit(struct kvm_pgtable_walk_data *data,
- kvm_pte_t *ptep, u32 level)
+ kvm_pte_t *ptep, u32 level, bool shared)
{
int ret = 0;
u64 addr = data->addr;
@@ -218,12 +218,12 @@ static inline int __kvm_pgtable_visit(struct kvm_pgtable_walk_data *data,
if (table && (flags & KVM_PGTABLE_WALK_TABLE_PRE)) {
ret = kvm_pgtable_visitor_cb(data, addr, level, ptep, &pte,
- KVM_PGTABLE_WALK_TABLE_PRE);
+ KVM_PGTABLE_WALK_TABLE_PRE, shared);
}
if (!table && (flags & KVM_PGTABLE_WALK_LEAF)) {
ret = kvm_pgtable_visitor_cb(data, addr, level, ptep, &pte,
- KVM_PGTABLE_WALK_LEAF);
+ KVM_PGTABLE_WALK_LEAF, shared);
}
if (ret)
@@ -237,13 +237,13 @@ static inline int __kvm_pgtable_visit(struct kvm_pgtable_walk_data *data,
}
childp = kvm_pte_follow(pte, data->pgt->mm_ops);
- ret = __kvm_pgtable_walk(data, childp, level + 1);
+ ret = __kvm_pgtable_walk(data, childp, level + 1, shared);
if (ret)
goto out;
if (flags & KVM_PGTABLE_WALK_TABLE_POST) {
ret = kvm_pgtable_visitor_cb(data, addr, level, ptep, &pte,
- KVM_PGTABLE_WALK_TABLE_POST);
+ KVM_PGTABLE_WALK_TABLE_POST, shared);
}
out:
@@ -251,7 +251,7 @@ static inline int __kvm_pgtable_visit(struct kvm_pgtable_walk_data *data,
}
static int __kvm_pgtable_walk(struct kvm_pgtable_walk_data *data,
- kvm_pte_t *pgtable, u32 level)
+ kvm_pte_t *pgtable, u32 level, bool shared)
{
u32 idx;
int ret = 0;
@@ -265,7 +265,7 @@ static int __kvm_pgtable_walk(struct kvm_pgtable_walk_data *data,
if (data->addr >= data->end)
break;
- ret = __kvm_pgtable_visit(data, ptep, level);
+ ret = __kvm_pgtable_visit(data, ptep, level, shared);
if (ret)
break;
}
@@ -273,7 +273,7 @@ static int __kvm_pgtable_walk(struct kvm_pgtable_walk_data *data,
return ret;
}
-static int _kvm_pgtable_walk(struct kvm_pgtable_walk_data *data)
+static int _kvm_pgtable_walk(struct kvm_pgtable_walk_data *data, bool shared)
{
u32 idx;
int ret = 0;
@@ -289,7 +289,7 @@ static int _kvm_pgtable_walk(struct kvm_pgtable_walk_data *data)
for (idx = kvm_pgd_page_idx(data); data->addr < data->end; ++idx) {
kvm_pte_t *ptep = &pgt->pgd[idx * PTRS_PER_PTE];
- ret = __kvm_pgtable_walk(data, ptep, pgt->start_level);
+ ret = __kvm_pgtable_walk(data, ptep, pgt->start_level, shared);
if (ret)
break;
}
@@ -298,7 +298,7 @@ static int _kvm_pgtable_walk(struct kvm_pgtable_walk_data *data)
}
int kvm_pgtable_walk(struct kvm_pgtable *pgt, u64 addr, u64 size,
- struct kvm_pgtable_walker *walker)
+ struct kvm_pgtable_walker *walker, bool shared)
{
struct kvm_pgtable_walk_data walk_data = {
.pgt = pgt,
@@ -308,7 +308,7 @@ int kvm_pgtable_walk(struct kvm_pgtable *pgt, u64 addr, u64 size,
};
kvm_pgtable_walk_begin();
- return _kvm_pgtable_walk(&walk_data);
+ return _kvm_pgtable_walk(&walk_data, shared);
kvm_pgtable_walk_end();
}
@@ -318,7 +318,7 @@ struct leaf_walk_data {
};
static int leaf_walker(u64 addr, u64 end, u32 level, kvm_pte_t *ptep, kvm_pte_t *old,
- enum kvm_pgtable_walk_flags flag, void * const arg)
+ enum kvm_pgtable_walk_flags flag, void * const arg, bool shared)
{
struct leaf_walk_data *data = arg;
@@ -340,7 +340,7 @@ int kvm_pgtable_get_leaf(struct kvm_pgtable *pgt, u64 addr,
int ret;
ret = kvm_pgtable_walk(pgt, ALIGN_DOWN(addr, PAGE_SIZE),
- PAGE_SIZE, &walker);
+ PAGE_SIZE, &walker, false);
if (!ret) {
if (ptep)
*ptep = data.pte;
@@ -409,7 +409,7 @@ enum kvm_pgtable_prot kvm_pgtable_hyp_pte_prot(kvm_pte_t pte)
}
static bool hyp_map_walker_try_leaf(u64 addr, u64 end, u32 level, kvm_pte_t *ptep,
- kvm_pte_t old, struct hyp_map_data *data)
+ kvm_pte_t old, struct hyp_map_data *data, bool shared)
{
kvm_pte_t new;
u64 granule = kvm_granule_size(level), phys = data->phys;
@@ -431,13 +431,13 @@ static bool hyp_map_walker_try_leaf(u64 addr, u64 end, u32 level, kvm_pte_t *pte
}
static int hyp_map_walker(u64 addr, u64 end, u32 level, kvm_pte_t *ptep, kvm_pte_t *old,
- enum kvm_pgtable_walk_flags flag, void * const arg)
+ enum kvm_pgtable_walk_flags flag, void * const arg, bool shared)
{
kvm_pte_t *childp;
struct hyp_map_data *data = arg;
struct kvm_pgtable_mm_ops *mm_ops = data->mm_ops;
- if (hyp_map_walker_try_leaf(addr, end, level, ptep, *old, arg))
+ if (hyp_map_walker_try_leaf(addr, end, level, ptep, *old, arg, shared))
return 0;
if (WARN_ON(level == KVM_PGTABLE_MAX_LEVELS - 1))
@@ -471,7 +471,7 @@ int kvm_pgtable_hyp_map(struct kvm_pgtable *pgt, u64 addr, u64 size, u64 phys,
if (ret)
return ret;
- ret = kvm_pgtable_walk(pgt, addr, size, &walker);
+ ret = kvm_pgtable_walk(pgt, addr, size, &walker, false);
dsb(ishst);
isb();
return ret;
@@ -483,7 +483,7 @@ struct hyp_unmap_data {
};
static int hyp_unmap_walker(u64 addr, u64 end, u32 level, kvm_pte_t *ptep, kvm_pte_t *old,
- enum kvm_pgtable_walk_flags flag, void * const arg)
+ enum kvm_pgtable_walk_flags flag, void * const arg, bool shared)
{
kvm_pte_t *childp = NULL;
u64 granule = kvm_granule_size(level);
@@ -536,7 +536,7 @@ u64 kvm_pgtable_hyp_unmap(struct kvm_pgtable *pgt, u64 addr, u64 size)
if (!pgt->mm_ops->page_count)
return 0;
- kvm_pgtable_walk(pgt, addr, size, &walker);
+ kvm_pgtable_walk(pgt, addr, size, &walker, false);
return unmap_data.unmapped;
}
@@ -559,7 +559,7 @@ int kvm_pgtable_hyp_init(struct kvm_pgtable *pgt, u32 va_bits,
}
static int hyp_free_walker(u64 addr, u64 end, u32 level, kvm_pte_t *ptep, kvm_pte_t *old,
- enum kvm_pgtable_walk_flags flag, void * const arg)
+ enum kvm_pgtable_walk_flags flag, void * const arg, bool shared)
{
struct kvm_pgtable_mm_ops *mm_ops = arg;
@@ -582,7 +582,7 @@ void kvm_pgtable_hyp_destroy(struct kvm_pgtable *pgt)
.arg = pgt->mm_ops,
};
- WARN_ON(kvm_pgtable_walk(pgt, 0, BIT(pgt->ia_bits), &walker));
+ WARN_ON(kvm_pgtable_walk(pgt, 0, BIT(pgt->ia_bits), &walker, false));
pgt->mm_ops->put_page(pgt->pgd);
pgt->pgd = NULL;
}
@@ -744,7 +744,8 @@ static bool stage2_leaf_mapping_allowed(u64 addr, u64 end, u32 level,
static int stage2_map_walker_try_leaf(u64 addr, u64 end, u32 level,
kvm_pte_t *ptep, kvm_pte_t old,
- struct stage2_map_data *data)
+ struct stage2_map_data *data,
+ bool shared)
{
kvm_pte_t new;
u64 granule = kvm_granule_size(level), phys = data->phys;
@@ -790,7 +791,8 @@ static int stage2_map_walker_try_leaf(u64 addr, u64 end, u32 level,
static int stage2_map_walk_table_pre(u64 addr, u64 end, u32 level,
kvm_pte_t *ptep, kvm_pte_t *old,
- struct stage2_map_data *data)
+ struct stage2_map_data *data,
+ bool shared)
{
if (data->anchor)
return 0;
@@ -812,7 +814,7 @@ static int stage2_map_walk_table_pre(u64 addr, u64 end, u32 level,
}
static int stage2_map_walk_leaf(u64 addr, u64 end, u32 level, kvm_pte_t *ptep,
- kvm_pte_t *old, struct stage2_map_data *data)
+ kvm_pte_t *old, struct stage2_map_data *data, bool shared)
{
struct kvm_pgtable_mm_ops *mm_ops = data->mm_ops;
kvm_pte_t *childp;
@@ -825,7 +827,7 @@ static int stage2_map_walk_leaf(u64 addr, u64 end, u32 level, kvm_pte_t *ptep,
return 0;
}
- ret = stage2_map_walker_try_leaf(addr, end, level, ptep, *old, data);
+ ret = stage2_map_walker_try_leaf(addr, end, level, ptep, *old, data, shared);
if (ret != -E2BIG)
return ret;
@@ -855,7 +857,8 @@ static int stage2_map_walk_leaf(u64 addr, u64 end, u32 level, kvm_pte_t *ptep,
static int stage2_map_walk_table_post(u64 addr, u64 end, u32 level,
kvm_pte_t *ptep, kvm_pte_t *old,
- struct stage2_map_data *data)
+ struct stage2_map_data *data,
+ bool shared)
{
struct kvm_pgtable_mm_ops *mm_ops = data->mm_ops;
kvm_pte_t *childp;
@@ -868,7 +871,7 @@ static int stage2_map_walk_table_post(u64 addr, u64 end, u32 level,
childp = data->childp;
data->anchor = NULL;
data->childp = NULL;
- ret = stage2_map_walk_leaf(addr, end, level, ptep, old, data);
+ ret = stage2_map_walk_leaf(addr, end, level, ptep, old, data, shared);
} else {
childp = kvm_pte_follow(*old, mm_ops);
}
@@ -899,17 +902,17 @@ static int stage2_map_walk_table_post(u64 addr, u64 end, u32 level,
* pointer and clearing the anchor to NULL.
*/
static int stage2_map_walker(u64 addr, u64 end, u32 level, kvm_pte_t *ptep, kvm_pte_t *old,
- enum kvm_pgtable_walk_flags flag, void * const arg)
+ enum kvm_pgtable_walk_flags flag, void * const arg, bool shared)
{
struct stage2_map_data *data = arg;
switch (flag) {
case KVM_PGTABLE_WALK_TABLE_PRE:
- return stage2_map_walk_table_pre(addr, end, level, ptep, old, data);
+ return stage2_map_walk_table_pre(addr, end, level, ptep, old, data, shared);
case KVM_PGTABLE_WALK_LEAF:
- return stage2_map_walk_leaf(addr, end, level, ptep, old, data);
+ return stage2_map_walk_leaf(addr, end, level, ptep, old, data, shared);
case KVM_PGTABLE_WALK_TABLE_POST:
- return stage2_map_walk_table_post(addr, end, level, ptep, old, data);
+ return stage2_map_walk_table_post(addr, end, level, ptep, old, data, shared);
}
return -EINVAL;
@@ -942,7 +945,7 @@ int kvm_pgtable_stage2_map(struct kvm_pgtable *pgt, u64 addr, u64 size,
if (ret)
return ret;
- ret = kvm_pgtable_walk(pgt, addr, size, &walker);
+ ret = kvm_pgtable_walk(pgt, addr, size, &walker, false);
dsb(ishst);
return ret;
}
@@ -970,13 +973,13 @@ int kvm_pgtable_stage2_set_owner(struct kvm_pgtable *pgt, u64 addr, u64 size,
if (owner_id > KVM_MAX_OWNER_ID)
return -EINVAL;
- ret = kvm_pgtable_walk(pgt, addr, size, &walker);
+ ret = kvm_pgtable_walk(pgt, addr, size, &walker, false);
return ret;
}
static int stage2_unmap_walker(u64 addr, u64 end, u32 level, kvm_pte_t *ptep,
kvm_pte_t *old, enum kvm_pgtable_walk_flags flag,
- void * const arg)
+ void * const arg, bool shared)
{
struct kvm_pgtable *pgt = arg;
struct kvm_s2_mmu *mmu = pgt->mmu;
@@ -1026,7 +1029,7 @@ int kvm_pgtable_stage2_unmap(struct kvm_pgtable *pgt, u64 addr, u64 size)
.flags = KVM_PGTABLE_WALK_LEAF | KVM_PGTABLE_WALK_TABLE_POST,
};
- return kvm_pgtable_walk(pgt, addr, size, &walker);
+ return kvm_pgtable_walk(pgt, addr, size, &walker, false);
}
struct stage2_attr_data {
@@ -1039,7 +1042,7 @@ struct stage2_attr_data {
static int stage2_attr_walker(u64 addr, u64 end, u32 level, kvm_pte_t *ptep,
kvm_pte_t *old, enum kvm_pgtable_walk_flags flag,
- void * const arg)
+ void * const arg, bool shared)
{
kvm_pte_t pte = *old;
struct stage2_attr_data *data = arg;
@@ -1091,7 +1094,7 @@ static int stage2_update_leaf_attrs(struct kvm_pgtable *pgt, u64 addr,
.flags = KVM_PGTABLE_WALK_LEAF,
};
- ret = kvm_pgtable_walk(pgt, addr, size, &walker);
+ ret = kvm_pgtable_walk(pgt, addr, size, &walker, false);
if (ret)
return ret;
@@ -1167,7 +1170,7 @@ int kvm_pgtable_stage2_relax_perms(struct kvm_pgtable *pgt, u64 addr,
static int stage2_flush_walker(u64 addr, u64 end, u32 level, kvm_pte_t *ptep,
kvm_pte_t *old, enum kvm_pgtable_walk_flags flag,
- void * const arg)
+ void * const arg, bool shared)
{
struct kvm_pgtable *pgt = arg;
struct kvm_pgtable_mm_ops *mm_ops = pgt->mm_ops;
@@ -1192,7 +1195,7 @@ int kvm_pgtable_stage2_flush(struct kvm_pgtable *pgt, u64 addr, u64 size)
if (stage2_has_fwb(pgt))
return 0;
- return kvm_pgtable_walk(pgt, addr, size, &walker);
+ return kvm_pgtable_walk(pgt, addr, size, &walker, false);
}
@@ -1226,7 +1229,7 @@ int __kvm_pgtable_stage2_init(struct kvm_pgtable *pgt, struct kvm_s2_mmu *mmu,
static int stage2_free_walker(u64 addr, u64 end, u32 level, kvm_pte_t *ptep,
kvm_pte_t *old, enum kvm_pgtable_walk_flags flag,
- void * const arg)
+ void * const arg, bool shared)
{
struct kvm_pgtable_mm_ops *mm_ops = arg;
@@ -1251,7 +1254,7 @@ void kvm_pgtable_stage2_destroy(struct kvm_pgtable *pgt)
.arg = pgt->mm_ops,
};
- WARN_ON(kvm_pgtable_walk(pgt, 0, BIT(pgt->ia_bits), &walker));
+ WARN_ON(kvm_pgtable_walk(pgt, 0, BIT(pgt->ia_bits), &walker, false));
pgd_sz = kvm_pgd_pages(pgt->ia_bits, pgt->start_level) * PAGE_SIZE;
pgt->mm_ops->free_pages_exact(pgt->pgd, pgd_sz);
pgt->pgd = NULL;
It is desirable to reuse the same page walkers for serial and parallel faults. Take an argument to kvm_pgtable_walk() (and throughout) to indicate whether or not a walk might happen in parallel with another. No functional change intended. Signed-off-by: Oliver Upton <oupton@google.com> --- arch/arm64/include/asm/kvm_pgtable.h | 5 +- arch/arm64/kvm/hyp/nvhe/mem_protect.c | 4 +- arch/arm64/kvm/hyp/nvhe/setup.c | 4 +- arch/arm64/kvm/hyp/pgtable.c | 91 ++++++++++++++------------- 4 files changed, 54 insertions(+), 50 deletions(-)