@@ -542,7 +542,7 @@ static inline struct kvm_vm *vm_create(uint64_t nr_pages)
struct kvm_vm *__vm_create_with_vcpus(enum vm_guest_mode mode, uint32_t nr_vcpus,
uint64_t slot0_mem_pages, uint64_t extra_mem_pages,
uint32_t num_percpu_pages, void *guest_code,
- uint32_t vcpuids[], struct kvm_vcpu *vcpus[]);
+ struct kvm_vcpu *vcpus[]);
static inline struct kvm_vm *vm_create_with_vcpus(uint32_t nr_vcpus,
void *guest_code,
@@ -550,7 +550,7 @@ static inline struct kvm_vm *vm_create_with_vcpus(uint32_t nr_vcpus,
{
return __vm_create_with_vcpus(VM_MODE_DEFAULT, nr_vcpus,
DEFAULT_GUEST_PHY_PAGES, 0, 0,
- guest_code, NULL, vcpus);
+ guest_code, vcpus);
}
/*
@@ -269,7 +269,7 @@ static struct kvm_vm *pre_init_before_test(enum vm_guest_mode mode, void *arg)
/* Create a VM with enough guest pages */
guest_num_pages = test_mem_size / guest_page_size;
vm = __vm_create_with_vcpus(mode, nr_vcpus, DEFAULT_GUEST_PHY_PAGES,
- guest_num_pages, 0, guest_code, NULL, NULL);
+ guest_num_pages, 0, guest_code, NULL);
/* Align down GPA of the testing memslot */
if (!p->phys_offset)
@@ -303,7 +303,7 @@ struct kvm_vm *__vm_create(enum vm_guest_mode mode, uint64_t nr_pages)
struct kvm_vm *__vm_create_with_vcpus(enum vm_guest_mode mode, uint32_t nr_vcpus,
uint64_t slot0_mem_pages, uint64_t extra_mem_pages,
uint32_t num_percpu_pages, void *guest_code,
- uint32_t vcpuids[], struct kvm_vcpu *vcpus[])
+ struct kvm_vcpu *vcpus[])
{
uint64_t vcpu_pages, extra_pg_pages, pages;
struct kvm_vcpu *vcpu;
@@ -331,9 +331,7 @@ struct kvm_vm *__vm_create_with_vcpus(enum vm_guest_mode mode, uint32_t nr_vcpus
vm = __vm_create(mode, pages);
for (i = 0; i < nr_vcpus; ++i) {
- uint32_t vcpuid = vcpuids ? vcpuids[i] : i;
-
- vcpu = vm_vcpu_add(vm, vcpuid, guest_code);
+ vcpu = vm_vcpu_add(vm, i, guest_code);
if (vcpus)
vcpus[i] = vcpu;
}
@@ -349,7 +347,7 @@ struct kvm_vm *__vm_create_with_one_vcpu(struct kvm_vcpu **vcpu,
struct kvm_vm *vm;
vm = __vm_create_with_vcpus(VM_MODE_DEFAULT, 1, DEFAULT_GUEST_PHY_PAGES,
- extra_mem_pages, 0, guest_code, NULL, vcpus);
+ extra_mem_pages, 0, guest_code, vcpus);
*vcpu = vcpus[0];
return vm;
@@ -140,7 +140,7 @@ struct kvm_vm *perf_test_create_vm(enum vm_guest_mode mode, int vcpus,
* effect as KVM allows aliasing HVAs in meslots.
*/
vm = __vm_create_with_vcpus(mode, vcpus, DEFAULT_GUEST_PHY_PAGES,
- guest_num_pages, 0, guest_code, NULL, NULL);
+ guest_num_pages, 0, guest_code, NULL);
pta->vm = vm;
Drop the @vcpuids parameter from VM creators now that there are no users. Allowing tests to specify IDs was a gigantic mistake as it resulted in tests with arbitrary and ultimately meaningless IDs that differed only because the author used test X intead of test Y as the source for copy+paste (the de facto standard way to create a KVM selftest). Except for literally two tests, x86's set_boot_cpu_id and s390's resets, tests do not and should not care about the vCPU ID. Signed-off-by: Sean Christopherson <seanjc@google.com> --- tools/testing/selftests/kvm/include/kvm_util_base.h | 4 ++-- tools/testing/selftests/kvm/kvm_page_table_test.c | 2 +- tools/testing/selftests/kvm/lib/kvm_util.c | 8 +++----- tools/testing/selftests/kvm/lib/perf_test_util.c | 2 +- 4 files changed, 7 insertions(+), 9 deletions(-)