@@ -196,7 +196,7 @@ static bool nested_svm_vmrun_msrpm(struct vcpu_svm *svm)
return true;
}
-static bool svm_get_nested_state_pages(struct kvm_vcpu *vcpu)
+static bool nested_svm_get_pages(struct kvm_vcpu *vcpu)
{
struct vcpu_svm *svm = to_svm(vcpu);
if (!nested_svm_vmrun_msrpm(svm)) {
@@ -834,7 +834,7 @@ static int nested_svm_intercept(struct vcpu_svm *svm)
/*
* Host-intercepted exceptions have been checked already in
* nested_svm_exit_special. There is nothing to do here,
- * the vmexit is injected by svm_check_nested_events.
+ * the vmexit is injected by nested_svm_check_events.
*/
vmexit = NESTED_EXIT_DONE;
break;
@@ -965,7 +965,7 @@ static void nested_svm_init(struct vcpu_svm *svm)
}
-static int svm_check_nested_events(struct kvm_vcpu *vcpu)
+static int nested_svm_check_events(struct kvm_vcpu *vcpu)
{
struct vcpu_svm *svm = to_svm(vcpu);
bool block_nested_events =
@@ -1049,7 +1049,7 @@ int nested_svm_exit_special(struct vcpu_svm *svm)
return NESTED_EXIT_CONTINUE;
}
-static int svm_get_nested_state(struct kvm_vcpu *vcpu,
+static int nested_svm_get_state(struct kvm_vcpu *vcpu,
struct kvm_nested_state __user *user_kvm_nested_state,
u32 user_data_size)
{
@@ -1106,7 +1106,7 @@ static int svm_get_nested_state(struct kvm_vcpu *vcpu,
return kvm_state.size;
}
-static int svm_set_nested_state(struct kvm_vcpu *vcpu,
+static int nested_svm_set_state(struct kvm_vcpu *vcpu,
struct kvm_nested_state __user *user_kvm_nested_state,
struct kvm_nested_state *kvm_state)
{
@@ -1209,9 +1209,11 @@ static int svm_set_nested_state(struct kvm_vcpu *vcpu,
return ret;
}
+#define KVM_X86_NESTED_OP_NAME(name) .name = nested_svm_##name
+
struct kvm_x86_nested_ops svm_nested_ops = {
- .check_events = svm_check_nested_events,
- .get_pages = svm_get_nested_state_pages,
- .get_state = svm_get_nested_state,
- .set_state = svm_set_nested_state,
+ KVM_X86_NESTED_OP_NAME(check_events),
+ KVM_X86_NESTED_OP_NAME(get_pages),
+ KVM_X86_NESTED_OP_NAME(get_state),
+ KVM_X86_NESTED_OP_NAME(set_state),
};