Message ID | 200905081203.55484.mark.langsdorf@amd.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Mark Langsdorf wrote: > From 01813db8627e74018c8cec90df7e345839351f23 Mon Sep 17 00:00:00 2001 > From: Mark Langsdorf <mark.langsdorf@amd.com> > Date: Thu, 7 May 2009 09:44:10 -0500 > Subject: [PATCH] Add support for Pause Filtering to AMD SVM > What's the differences wrt retry 1? > This feature creates a new field in the VMCB called Pause > Filter Count. If Pause Filter Count is greater than 0 and > intercepting PAUSEs is enabled, the processor will increment > an internal counter when a PAUSE instruction occurs instead > of intercepting. When the internal counter reaches the > Pause Filter Count value, a PAUSE intercept will occur. > > This feature can be used to detect contended spinlocks, > especially when the lock holding VCPU is not scheduled. > Rescheduling another VCPU prevents the VCPU seeking the > lock from wasting its quantum by spinning idly. > > Experimental results show that most spinlocks are held > for less than 1000 PAUSE cycles or more than a few > thousand. Default the Pause Filter Counter to 3000 to > detect the contended spinlocks. > 3000. > Processor support for this feature is indicated by a CPUID > bit. > > On a 24 core system running 4 guests each with 16 VCPUs, > this patch improved overall performance of each guest's > 32 job kernbench by approximately 1%. Further performance > improvement may be possible with a more sophisticated > yield algorithm. > Like I mentioned earlier, I don't think schedule() does anything on CFS. Try sched_yield(), but set /proc/sys/kernel/sched_compat_yield. > + > + if (svm_has(SVM_FEATURE_PAUSE_FILTER)) { > + control->pause_filter_count = 5000; > + control->intercept |= (1ULL << INTERCEPT_PAUSE); > + } > + > Here, 5000? > } > > static int svm_vcpu_reset(struct kvm_vcpu *vcpu) > @@ -2087,6 +2094,15 @@ static int interrupt_window_interception(struct vcpu_svm *svm, > return 1; > } > > +static int pause_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run) > +{ > + /* Simple yield */ > + vcpu_put(&svm->vcpu); > + schedule(); > + vcpu_load(&svm->vcpu); > + return 1; > +} > + > You don't need to vcpu_put() and vcpu_load(). The scheduler will call them for you if/when it switches tasks.
> What's the differences wrt retry 1? I'm using git format-patch as you requested. > > This feature creates a new field in the VMCB called Pause > > Filter Count. If Pause Filter Count is greater than 0 and > > intercepting PAUSEs is enabled, the processor will increment > > an internal counter when a PAUSE instruction occurs instead > > of intercepting. When the internal counter reaches the > > Pause Filter Count value, a PAUSE intercept will occur. > > > > This feature can be used to detect contended spinlocks, > > especially when the lock holding VCPU is not scheduled. > > Rescheduling another VCPU prevents the VCPU seeking the > > lock from wasting its quantum by spinning idly. > > > > Experimental results show that most spinlocks are held > > for less than 1000 PAUSE cycles or more than a few > > thousand. Default the Pause Filter Counter to 3000 to > > detect the contended spinlocks. > > 3000. Thanks, I keep missing that. > > On a 24 core system running 4 guests each with 16 VCPUs, > > this patch improved overall performance of each guest's > > 32 job kernbench by approximately 1%. Further performance > > improvement may be possible with a more sophisticated > > yield algorithm. > > > > Like I mentioned earlier, I don't think schedule() does > anything on CFS. > > Try sched_yield(), but set /proc/sys/kernel/sched_compat_yield. Will do. > > +static int pause_interception(struct vcpu_svm *svm, struct > kvm_run *kvm_run) > > +{ > > + /* Simple yield */ > > + vcpu_put(&svm->vcpu); > > + schedule(); > > + vcpu_load(&svm->vcpu); > > + return 1; > > +} > > + > > > > You don't need to vcpu_put() and vcpu_load(). The scheduler > will call them for you if/when it switches tasks. I was waiting for feedback from Ingo on that issue, but I'll try sched_yield() instead. -Mark Langsdorf Operating System Research Center AMD -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
diff --git a/arch/x86/include/asm/svm.h b/arch/x86/include/asm/svm.h index 85574b7..1fecb7e 100644 --- a/arch/x86/include/asm/svm.h +++ b/arch/x86/include/asm/svm.h @@ -57,7 +57,8 @@ struct __attribute__ ((__packed__)) vmcb_control_area { u16 intercept_dr_write; u32 intercept_exceptions; u64 intercept; - u8 reserved_1[44]; + u8 reserved_1[42]; + u16 pause_filter_count; u64 iopm_base_pa; u64 msrpm_base_pa; u64 tsc_offset; diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c index ef43a18..4279141 100644 --- a/arch/x86/kvm/svm.c +++ b/arch/x86/kvm/svm.c @@ -45,6 +45,7 @@ MODULE_LICENSE("GPL"); #define SVM_FEATURE_NPT (1 << 0) #define SVM_FEATURE_LBRV (1 << 1) #define SVM_FEATURE_SVML (1 << 2) +#define SVM_FEATURE_PAUSE_FILTER (1 << 10) #define DEBUGCTL_RESERVED_BITS (~(0x3fULL)) @@ -575,6 +576,12 @@ static void init_vmcb(struct vcpu_svm *svm) svm->nested_vmcb = 0; svm->vcpu.arch.hflags = HF_GIF_MASK; + + if (svm_has(SVM_FEATURE_PAUSE_FILTER)) { + control->pause_filter_count = 5000; + control->intercept |= (1ULL << INTERCEPT_PAUSE); + } + } static int svm_vcpu_reset(struct kvm_vcpu *vcpu) @@ -2087,6 +2094,15 @@ static int interrupt_window_interception(struct vcpu_svm *svm, return 1; } +static int pause_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run) +{ + /* Simple yield */ + vcpu_put(&svm->vcpu); + schedule(); + vcpu_load(&svm->vcpu); + return 1; +} + static int (*svm_exit_handlers[])(struct vcpu_svm *svm, struct kvm_run *kvm_run) = { [SVM_EXIT_READ_CR0] = emulate_on_interception, @@ -2123,6 +2139,7 @@ static int (*svm_exit_handlers[])(struct vcpu_svm *svm, [SVM_EXIT_CPUID] = cpuid_interception, [SVM_EXIT_IRET] = iret_interception, [SVM_EXIT_INVD] = emulate_on_interception, + [SVM_EXIT_PAUSE] = pause_interception, [SVM_EXIT_HLT] = halt_interception, [SVM_EXIT_INVLPG] = invlpg_interception, [SVM_EXIT_INVLPGA] = invalid_op_interception, diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c index 2b73e19..e2b730d 100644 --- a/virt/kvm/kvm_main.c +++ b/virt/kvm/kvm_main.c @@ -710,6 +710,7 @@ void vcpu_load(struct kvm_vcpu *vcpu) kvm_arch_vcpu_load(vcpu, cpu); put_cpu(); } +EXPORT_SYMBOL_GPL(vcpu_load); void vcpu_put(struct kvm_vcpu *vcpu) { @@ -719,6 +720,7 @@ void vcpu_put(struct kvm_vcpu *vcpu) preempt_enable(); mutex_unlock(&vcpu->mutex); } +EXPORT_SYMBOL_GPL(vcpu_put); static void ack_flush(void *_completed) {