diff mbox series

[4/4] arm64: KVM: Enable support for :G/:H perf event modifiers

Message ID 1542286549-4501-5-git-send-email-andrew.murray@arm.com (mailing list archive)
State Superseded
Headers show
Series arm64: Support perf event modifiers :G and :H | expand

Commit Message

Andrew Murray Nov. 15, 2018, 12:55 p.m. UTC
Enable/disable event counters as appropriate when entering and exiting
the guest to enable support for guest or host only event counting.

For both VHE and non-VHE we switch the counters between host/guest at
EL2. EL2 is filtered out by the PMU when we are using the :G modifier.

Signed-off-by: Andrew Murray <andrew.murray@arm.com>
---
 arch/arm64/kvm/hyp/switch.c | 38 ++++++++++++++++++++++++++++++++++++++
 1 file changed, 38 insertions(+)

Comments

Julien Thierry Nov. 15, 2018, 2 p.m. UTC | #1
Hi Andrew,

On 15/11/18 12:55, Andrew Murray wrote:
> Enable/disable event counters as appropriate when entering and exiting
> the guest to enable support for guest or host only event counting.
> 
> For both VHE and non-VHE we switch the counters between host/guest at
> EL2. EL2 is filtered out by the PMU when we are using the :G modifier.
> 
> Signed-off-by: Andrew Murray <andrew.murray@arm.com>
> ---
>   arch/arm64/kvm/hyp/switch.c | 38 ++++++++++++++++++++++++++++++++++++++
>   1 file changed, 38 insertions(+)
> 
> diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c
> index d496ef5..ebf0aac 100644
> --- a/arch/arm64/kvm/hyp/switch.c
> +++ b/arch/arm64/kvm/hyp/switch.c
> @@ -373,6 +373,32 @@ static bool __hyp_text __hyp_switch_fpsimd(struct kvm_vcpu *vcpu)
>   	return true;
>   }
>   
> +static bool __hyp_text __pmu_switch_to_guest(struct kvm_cpu_context *host_ctxt)
> +{
> +	u32 host_only = host_ctxt->events_host_only;
> +	u32 guest_only = host_ctxt->events_guest_only;
> +
> +	if (host_only)
> +		write_sysreg(host_only, pmcntenclr_el0);
> +
> +	if (guest_only)
> +		write_sysreg(guest_only, pmcntenset_el0);
> +
> +	return (host_only || guest_only);
> +}
> +
> +static void __hyp_text __pmu_switch_to_host(struct kvm_cpu_context *host_ctxt)
> +{
> +	u32 host_only = host_ctxt->events_host_only;
> +	u32 guest_only = host_ctxt->events_guest_only;
> +
> +	if (guest_only)
> +		write_sysreg(guest_only, pmcntenclr_el0);
> +
> +	if (host_only)
> +		write_sysreg(host_only, pmcntenset_el0);

In the perf_event code, there is an ISB after enabling an event. I guess 
we don't need it when setting the guest events since I believe the eret 
to the guess give us the context synchronization. But don't we need one 
here when restoring host only events?

Thanks,
Andrew Murray Nov. 15, 2018, 3:57 p.m. UTC | #2
On Thu, Nov 15, 2018 at 02:00:39PM +0000, Julien Thierry wrote:
> Hi Andrew,
> 
> On 15/11/18 12:55, Andrew Murray wrote:
> > Enable/disable event counters as appropriate when entering and exiting
> > the guest to enable support for guest or host only event counting.
> > 
> > For both VHE and non-VHE we switch the counters between host/guest at
> > EL2. EL2 is filtered out by the PMU when we are using the :G modifier.
> > 
> > Signed-off-by: Andrew Murray <andrew.murray@arm.com>
> > ---
> >   arch/arm64/kvm/hyp/switch.c | 38 ++++++++++++++++++++++++++++++++++++++
> >   1 file changed, 38 insertions(+)
> > 
> > diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c
> > index d496ef5..ebf0aac 100644
> > --- a/arch/arm64/kvm/hyp/switch.c
> > +++ b/arch/arm64/kvm/hyp/switch.c
> > @@ -373,6 +373,32 @@ static bool __hyp_text __hyp_switch_fpsimd(struct kvm_vcpu *vcpu)
> >   	return true;
> >   }
> > +static bool __hyp_text __pmu_switch_to_guest(struct kvm_cpu_context *host_ctxt)
> > +{
> > +	u32 host_only = host_ctxt->events_host_only;
> > +	u32 guest_only = host_ctxt->events_guest_only;
> > +
> > +	if (host_only)
> > +		write_sysreg(host_only, pmcntenclr_el0);
> > +
> > +	if (guest_only)
> > +		write_sysreg(guest_only, pmcntenset_el0);
> > +
> > +	return (host_only || guest_only);
> > +}
> > +
> > +static void __hyp_text __pmu_switch_to_host(struct kvm_cpu_context *host_ctxt)
> > +{
> > +	u32 host_only = host_ctxt->events_host_only;
> > +	u32 guest_only = host_ctxt->events_guest_only;
> > +
> > +	if (guest_only)
> > +		write_sysreg(guest_only, pmcntenclr_el0);
> > +
> > +	if (host_only)
> > +		write_sysreg(host_only, pmcntenset_el0);
> 
> In the perf_event code, there is an ISB after enabling an event. I guess we
> don't need it when setting the guest events since I believe the eret to the
> guess give us the context synchronization. But don't we need one here when
> restoring host only events?

It's not really clear to me why the isb is present in the existing code,
this was only recently introduced when adding the chained events support.

Ideally for chained events you'd want to start the overflow counter first
(idx) followed by the low counter second (idx-1) as to not miss overflows
so an isb inbetween may be helpful. Though the isb is after both enables, this
sets a clear line of where event counting starts - but ideally this would be
symmetrical with an isb after the disable.

At present chained counters aren't supported in the guest but in any case
we turn them all on/off atomically rather than individually.

I guess we get a trivial gain in accuracy by adding ISB's at some performance
cost - I'm not sure I see the benefit - unless I'm missing something?

Thanks,

Andrew Murray

> 
> Thanks,
> 
> -- 
> Julien Thierry
Suzuki K Poulose Nov. 15, 2018, 5:40 p.m. UTC | #3
On 15/11/2018 15:57, Andrew Murray wrote:
> On Thu, Nov 15, 2018 at 02:00:39PM +0000, Julien Thierry wrote:
>> Hi Andrew,
>>
>> On 15/11/18 12:55, Andrew Murray wrote:
>>> Enable/disable event counters as appropriate when entering and exiting
>>> the guest to enable support for guest or host only event counting.
>>>
>>> For both VHE and non-VHE we switch the counters between host/guest at
>>> EL2. EL2 is filtered out by the PMU when we are using the :G modifier.
>>>
>>> Signed-off-by: Andrew Murray <andrew.murray@arm.com>
>>> ---
>>>    arch/arm64/kvm/hyp/switch.c | 38 ++++++++++++++++++++++++++++++++++++++
>>>    1 file changed, 38 insertions(+)
>>>
>>> diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c
>>> index d496ef5..ebf0aac 100644
>>> --- a/arch/arm64/kvm/hyp/switch.c
>>> +++ b/arch/arm64/kvm/hyp/switch.c
>>> @@ -373,6 +373,32 @@ static bool __hyp_text __hyp_switch_fpsimd(struct kvm_vcpu *vcpu)
>>>    	return true;
>>>    }
>>> +static bool __hyp_text __pmu_switch_to_guest(struct kvm_cpu_context *host_ctxt)
>>> +{
>>> +	u32 host_only = host_ctxt->events_host_only;
>>> +	u32 guest_only = host_ctxt->events_guest_only;
>>> +
>>> +	if (host_only)
>>> +		write_sysreg(host_only, pmcntenclr_el0);
>>> +
>>> +	if (guest_only)
>>> +		write_sysreg(guest_only, pmcntenset_el0);
>>> +
>>> +	return (host_only || guest_only);
>>> +}
>>> +
>>> +static void __hyp_text __pmu_switch_to_host(struct kvm_cpu_context *host_ctxt)
>>> +{
>>> +	u32 host_only = host_ctxt->events_host_only;
>>> +	u32 guest_only = host_ctxt->events_guest_only;
>>> +
>>> +	if (guest_only)
>>> +		write_sysreg(guest_only, pmcntenclr_el0);
>>> +
>>> +	if (host_only)
>>> +		write_sysreg(host_only, pmcntenset_el0);
>>
>> In the perf_event code, there is an ISB after enabling an event. I guess we
>> don't need it when setting the guest events since I believe the eret to the
>> guess give us the context synchronization. But don't we need one here when
>> restoring host only events?
> 
> It's not really clear to me why the isb is present in the existing code,
> this was only recently introduced when adding the chained events support.
> 
> Ideally for chained events you'd want to start the overflow counter first
> (idx) followed by the low counter second (idx-1) as to not miss overflows
> so an isb inbetween may be helpful. Though the isb is after both enables, this
> sets a clear line of where event counting starts - but ideally this would be
> symmetrical with an isb after the disable.

I think the isb() in the armv8_pmu_enable_event_counter() is
unnecessary, and might have been a left over from earlier versions
of the series. Please feel free to remove it.

> 
> At present chained counters aren't supported in the guest but in any case
> we turn them all on/off atomically rather than individually.
> 
> I guess we get a trivial gain in accuracy by adding ISB's at some performance
> cost - I'm not sure I see the benefit - unless I'm missing something?

But, I think Julien has a valid point here. When we modify the
pmcnten{set/clr} registers, the PMU could be enabled. (i.e, PMCR_E set).

So in order to synchronize the changes to the counters, we need an isb()
in the switch to host case to take immediate effect of the counter
changes.

Cheers
Suzuki
Andrew Murray Nov. 16, 2018, 12:12 p.m. UTC | #4
On Thu, Nov 15, 2018 at 05:40:24PM +0000, Suzuki K Poulose wrote:
> 
> 
> On 15/11/2018 15:57, Andrew Murray wrote:
> > On Thu, Nov 15, 2018 at 02:00:39PM +0000, Julien Thierry wrote:
> > > Hi Andrew,
> > > 
> > > On 15/11/18 12:55, Andrew Murray wrote:
> > > > Enable/disable event counters as appropriate when entering and exiting
> > > > the guest to enable support for guest or host only event counting.
> > > > 
> > > > For both VHE and non-VHE we switch the counters between host/guest at
> > > > EL2. EL2 is filtered out by the PMU when we are using the :G modifier.
> > > > 
> > > > Signed-off-by: Andrew Murray <andrew.murray@arm.com>
> > > > ---
> > > >    arch/arm64/kvm/hyp/switch.c | 38 ++++++++++++++++++++++++++++++++++++++
> > > >    1 file changed, 38 insertions(+)
> > > > 
> > > > diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c
> > > > index d496ef5..ebf0aac 100644
> > > > --- a/arch/arm64/kvm/hyp/switch.c
> > > > +++ b/arch/arm64/kvm/hyp/switch.c
> > > > @@ -373,6 +373,32 @@ static bool __hyp_text __hyp_switch_fpsimd(struct kvm_vcpu *vcpu)
> > > >    	return true;
> > > >    }
> > > > +static bool __hyp_text __pmu_switch_to_guest(struct kvm_cpu_context *host_ctxt)
> > > > +{
> > > > +	u32 host_only = host_ctxt->events_host_only;
> > > > +	u32 guest_only = host_ctxt->events_guest_only;
> > > > +
> > > > +	if (host_only)
> > > > +		write_sysreg(host_only, pmcntenclr_el0);
> > > > +
> > > > +	if (guest_only)
> > > > +		write_sysreg(guest_only, pmcntenset_el0);
> > > > +
> > > > +	return (host_only || guest_only);
> > > > +}
> > > > +
> > > > +static void __hyp_text __pmu_switch_to_host(struct kvm_cpu_context *host_ctxt)
> > > > +{
> > > > +	u32 host_only = host_ctxt->events_host_only;
> > > > +	u32 guest_only = host_ctxt->events_guest_only;
> > > > +
> > > > +	if (guest_only)
> > > > +		write_sysreg(guest_only, pmcntenclr_el0);
> > > > +
> > > > +	if (host_only)
> > > > +		write_sysreg(host_only, pmcntenset_el0);
> > > 
> > > In the perf_event code, there is an ISB after enabling an event. I guess we
> > > don't need it when setting the guest events since I believe the eret to the
> > > guess give us the context synchronization. But don't we need one here when
> > > restoring host only events?
> > 
> > It's not really clear to me why the isb is present in the existing code,
> > this was only recently introduced when adding the chained events support.
> > 
> > Ideally for chained events you'd want to start the overflow counter first
> > (idx) followed by the low counter second (idx-1) as to not miss overflows
> > so an isb inbetween may be helpful. Though the isb is after both enables, this
> > sets a clear line of where event counting starts - but ideally this would be
> > symmetrical with an isb after the disable.
> 
> I think the isb() in the armv8_pmu_enable_event_counter() is
> unnecessary, and might have been a left over from earlier versions
> of the series. Please feel free to remove it.

OK I'll do that.

> 
> > 
> > At present chained counters aren't supported in the guest but in any case
> > we turn them all on/off atomically rather than individually.
> > 
> > I guess we get a trivial gain in accuracy by adding ISB's at some performance
> > cost - I'm not sure I see the benefit - unless I'm missing something?
> 
> But, I think Julien has a valid point here. When we modify the
> pmcnten{set/clr} registers, the PMU could be enabled. (i.e, PMCR_E set).
> 
> So in order to synchronize the changes to the counters, we need an isb()
> in the switch to host case to take immediate effect of the counter
> changes.

For VHE we already do an isb in kvm_arm_vhe_guest_exit (next line of code
to kvm_arm_vhe_guest_exit).

For !VHE as I understand we will eret from EL2 (due to kvm_call_hyp call
completing) and thus also implicitly isb.

If that's correct we don't need to add any isb's right?

Thanks,

Andrew Murray

> 
> Cheers
> Suzuki
Suzuki K Poulose Nov. 16, 2018, 5:53 p.m. UTC | #5
On 16/11/2018 12:12, Andrew Murray wrote:
> On Thu, Nov 15, 2018 at 05:40:24PM +0000, Suzuki K Poulose wrote:
>>
>>
>> On 15/11/2018 15:57, Andrew Murray wrote:
>>> On Thu, Nov 15, 2018 at 02:00:39PM +0000, Julien Thierry wrote:
>>>> Hi Andrew,
>>>>
>>>> On 15/11/18 12:55, Andrew Murray wrote:
>>>>> Enable/disable event counters as appropriate when entering and exiting
>>>>> the guest to enable support for guest or host only event counting.
>>>>>
>>>>> For both VHE and non-VHE we switch the counters between host/guest at
>>>>> EL2. EL2 is filtered out by the PMU when we are using the :G modifier.
>>>>>
>>>>> Signed-off-by: Andrew Murray <andrew.murray@arm.com>
>>>>> ---
>>>>>     arch/arm64/kvm/hyp/switch.c | 38 ++++++++++++++++++++++++++++++++++++++
>>>>>     1 file changed, 38 insertions(+)
>>>>>
>>>>> diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c
>>>>> index d496ef5..ebf0aac 100644
>>>>> --- a/arch/arm64/kvm/hyp/switch.c
>>>>> +++ b/arch/arm64/kvm/hyp/switch.c
>>>>> @@ -373,6 +373,32 @@ static bool __hyp_text __hyp_switch_fpsimd(struct kvm_vcpu *vcpu)
>>>>>     	return true;
>>>>>     }
>>>>> +static bool __hyp_text __pmu_switch_to_guest(struct kvm_cpu_context *host_ctxt)
>>>>> +{
>>>>> +	u32 host_only = host_ctxt->events_host_only;
>>>>> +	u32 guest_only = host_ctxt->events_guest_only;
>>>>> +
>>>>> +	if (host_only)
>>>>> +		write_sysreg(host_only, pmcntenclr_el0);
>>>>> +
>>>>> +	if (guest_only)
>>>>> +		write_sysreg(guest_only, pmcntenset_el0);
>>>>> +
>>>>> +	return (host_only || guest_only);
>>>>> +}
>>>>> +
>>>>> +static void __hyp_text __pmu_switch_to_host(struct kvm_cpu_context *host_ctxt)
>>>>> +{
>>>>> +	u32 host_only = host_ctxt->events_host_only;
>>>>> +	u32 guest_only = host_ctxt->events_guest_only;
>>>>> +
>>>>> +	if (guest_only)
>>>>> +		write_sysreg(guest_only, pmcntenclr_el0);
>>>>> +
>>>>> +	if (host_only)
>>>>> +		write_sysreg(host_only, pmcntenset_el0);
>>>>
>>>> In the perf_event code, there is an ISB after enabling an event. I guess we
>>>> don't need it when setting the guest events since I believe the eret to the
>>>> guess give us the context synchronization. But don't we need one here when
>>>> restoring host only events?
>>>
>>> It's not really clear to me why the isb is present in the existing code,
>>> this was only recently introduced when adding the chained events support.
>>>
>>> Ideally for chained events you'd want to start the overflow counter first
>>> (idx) followed by the low counter second (idx-1) as to not miss overflows
>>> so an isb inbetween may be helpful. Though the isb is after both enables, this
>>> sets a clear line of where event counting starts - but ideally this would be
>>> symmetrical with an isb after the disable.
>>
>> I think the isb() in the armv8_pmu_enable_event_counter() is
>> unnecessary, and might have been a left over from earlier versions
>> of the series. Please feel free to remove it.
> 
> OK I'll do that.
> 
>>
>>>
>>> At present chained counters aren't supported in the guest but in any case
>>> we turn them all on/off atomically rather than individually.
>>>
>>> I guess we get a trivial gain in accuracy by adding ISB's at some performance
>>> cost - I'm not sure I see the benefit - unless I'm missing something?
>>
>> But, I think Julien has a valid point here. When we modify the
>> pmcnten{set/clr} registers, the PMU could be enabled. (i.e, PMCR_E set).
>>
>> So in order to synchronize the changes to the counters, we need an isb()
>> in the switch to host case to take immediate effect of the counter
>> changes.
> 
> For VHE we already do an isb in kvm_arm_vhe_guest_exit (next line of code
> to kvm_arm_vhe_guest_exit).
> 
> For !VHE as I understand we will eret from EL2 (due to kvm_call_hyp call
> completing) and thus also implicitly isb.
> 
> If that's correct we don't need to add any isb's right?

Yes, you're right. May be it is worth mentioning it where we switch to
host, so that we don't have to dig this again when we look at it later.

Cheers
Suzuki
diff mbox series

Patch

diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c
index d496ef5..ebf0aac 100644
--- a/arch/arm64/kvm/hyp/switch.c
+++ b/arch/arm64/kvm/hyp/switch.c
@@ -373,6 +373,32 @@  static bool __hyp_text __hyp_switch_fpsimd(struct kvm_vcpu *vcpu)
 	return true;
 }
 
+static bool __hyp_text __pmu_switch_to_guest(struct kvm_cpu_context *host_ctxt)
+{
+	u32 host_only = host_ctxt->events_host_only;
+	u32 guest_only = host_ctxt->events_guest_only;
+
+	if (host_only)
+		write_sysreg(host_only, pmcntenclr_el0);
+
+	if (guest_only)
+		write_sysreg(guest_only, pmcntenset_el0);
+
+	return (host_only || guest_only);
+}
+
+static void __hyp_text __pmu_switch_to_host(struct kvm_cpu_context *host_ctxt)
+{
+	u32 host_only = host_ctxt->events_host_only;
+	u32 guest_only = host_ctxt->events_guest_only;
+
+	if (guest_only)
+		write_sysreg(guest_only, pmcntenclr_el0);
+
+	if (host_only)
+		write_sysreg(host_only, pmcntenset_el0);
+}
+
 /*
  * Return true when we were able to fixup the guest exit and should return to
  * the guest, false when we should restore the host state and return to the
@@ -488,12 +514,15 @@  int kvm_vcpu_run_vhe(struct kvm_vcpu *vcpu)
 {
 	struct kvm_cpu_context *host_ctxt;
 	struct kvm_cpu_context *guest_ctxt;
+	bool pmu_switch_needed;
 	u64 exit_code;
 
 	host_ctxt = vcpu->arch.host_cpu_context;
 	host_ctxt->__hyp_running_vcpu = vcpu;
 	guest_ctxt = &vcpu->arch.ctxt;
 
+	pmu_switch_needed = __pmu_switch_to_guest(host_ctxt);
+
 	sysreg_save_host_state_vhe(host_ctxt);
 
 	__activate_traps(vcpu);
@@ -524,6 +553,9 @@  int kvm_vcpu_run_vhe(struct kvm_vcpu *vcpu)
 
 	__debug_switch_to_host(vcpu);
 
+	if (pmu_switch_needed)
+		__pmu_switch_to_host(host_ctxt);
+
 	return exit_code;
 }
 
@@ -532,6 +564,7 @@  int __hyp_text __kvm_vcpu_run_nvhe(struct kvm_vcpu *vcpu)
 {
 	struct kvm_cpu_context *host_ctxt;
 	struct kvm_cpu_context *guest_ctxt;
+	bool pmu_switch_needed;
 	u64 exit_code;
 
 	vcpu = kern_hyp_va(vcpu);
@@ -540,6 +573,8 @@  int __hyp_text __kvm_vcpu_run_nvhe(struct kvm_vcpu *vcpu)
 	host_ctxt->__hyp_running_vcpu = vcpu;
 	guest_ctxt = &vcpu->arch.ctxt;
 
+	pmu_switch_needed = __pmu_switch_to_guest(host_ctxt);
+
 	__sysreg_save_state_nvhe(host_ctxt);
 
 	__activate_traps(vcpu);
@@ -586,6 +621,9 @@  int __hyp_text __kvm_vcpu_run_nvhe(struct kvm_vcpu *vcpu)
 	 */
 	__debug_switch_to_host(vcpu);
 
+	if (pmu_switch_needed)
+		__pmu_switch_to_host(host_ctxt);
+
 	return exit_code;
 }