Message ID | 1550497684-26046-7-git-send-email-andrew.murray@arm.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | KVM: arm/arm64: add support for chained counters | expand |
On Mon, 18 Feb 2019 13:48:04 +0000 Andrew Murray <andrew.murray@arm.com> wrote: > Emulate chained PMU counters by creating a single 64 bit event counter > for a pair of chained KVM counters. > > Signed-off-by: Andrew Murray <andrew.murray@arm.com> > --- > include/kvm/arm_pmu.h | 1 + > virt/kvm/arm/pmu.c | 322 +++++++++++++++++++++++++++++++++++++++++--------- > 2 files changed, 270 insertions(+), 53 deletions(-) > > diff --git a/include/kvm/arm_pmu.h b/include/kvm/arm_pmu.h > index b73f31b..8e691ee 100644 > --- a/include/kvm/arm_pmu.h > +++ b/include/kvm/arm_pmu.h > @@ -29,6 +29,7 @@ struct kvm_pmc { > u8 idx; /* index into the pmu->pmc array */ > struct perf_event *perf_event; > u64 bitmask; > + u64 overflow_count; > }; > > struct kvm_pmu { > diff --git a/virt/kvm/arm/pmu.c b/virt/kvm/arm/pmu.c > index 524b27f..2852dc4 100644 > --- a/virt/kvm/arm/pmu.c > +++ b/virt/kvm/arm/pmu.c > @@ -24,9 +24,26 @@ > #include <kvm/arm_pmu.h> > #include <kvm/arm_vgic.h> > > +static void kvm_pmu_stop_release_perf_event_pair(struct kvm_vcpu *vcpu, > + u64 pair_low); > +static void kvm_pmu_stop_release_perf_event(struct kvm_vcpu *vcpu, > + u64 select_idx); > +static void kvm_pmu_sync_counter_enable_pair(struct kvm_vcpu *vcpu, u64 pair_low); > static void kvm_pmu_sync_counter_enable(struct kvm_vcpu *vcpu, u64 select_idx); > static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx); > -static void kvm_pmu_stop_counter(struct kvm_vcpu *vcpu, struct kvm_pmc *pmc); > + > +#define PERF_ATTR_CFG1_KVM_PMU_CHAINED 0x1 > + > +/** > + * kvm_pmu_counter_is_high_word - is select_idx high counter of 64bit event > + * @pmc: The PMU counter pointer > + * @select_idx: The counter index What is select_idx here? > + */ > +static inline bool kvm_pmu_counter_is_high_word(struct kvm_pmc *pmc) > +{ > + return ((pmc->perf_event->attr.config1 & PERF_ATTR_CFG1_KVM_PMU_CHAINED) > + && (pmc->idx % 2)); nit: This is usually expressed as 'idx & 1', but that's not a big deal. > +} > > /** > * kvm_pmu_get_counter_value - get PMU counter value > @@ -35,22 +52,70 @@ static void kvm_pmu_stop_counter(struct kvm_vcpu *vcpu, struct kvm_pmc *pmc); > */ > u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx) > { > - u64 counter, reg, enabled, running; > + u64 counter_idx, reg, enabled, running, incr; > struct kvm_pmu *pmu = &vcpu->arch.pmu; > struct kvm_pmc *pmc = &pmu->pmc[select_idx]; > > reg = (select_idx == ARMV8_PMU_CYCLE_IDX) > ? PMCCNTR_EL0 : PMEVCNTR0_EL0 + select_idx; > - counter = __vcpu_sys_reg(vcpu, reg); > + counter_idx = __vcpu_sys_reg(vcpu, reg); Sorry, but the _idx suffix is baffling. What you seem to have here is a counter value, not an index. 'counter' was the right name. > > /* The real counter value is equal to the value of counter register plus > * the value perf event counts. > */ > - if (pmc->perf_event) > - counter += perf_event_read_value(pmc->perf_event, &enabled, > + if (pmc->perf_event) { > + incr = perf_event_read_value(pmc->perf_event, &enabled, > &running); nit: please align the function arguments. > > - return counter & pmc->bitmask; > + if (kvm_pmu_counter_is_high_word(pmc)) { > + u64 counter_low, counter; > + > + reg = (select_idx == ARMV8_PMU_CYCLE_IDX) > + ? PMCCNTR_EL0 : PMEVCNTR0_EL0 + select_idx - 1; Why do you cater for the cycle counter here? It would feel wrong if we could reach this code with the cycle counter, given that it is 64bit and doesn't need any of this. > + counter_low = __vcpu_sys_reg(vcpu, reg); > + counter = lower_32_bits(counter_low) | (counter_idx << 32); > + counter_idx = upper_32_bits(counter + incr); > + } else { > + counter_idx += incr; > + } > + } > + > + return counter_idx & pmc->bitmask; > +} > + > +/** > + * kvm_pmu_counter_is_enabled - is a counter active > + * @vcpu: The vcpu pointer > + * @select_idx: The counter index > + */ > +static bool kvm_pmu_counter_is_enabled(struct kvm_vcpu *vcpu, u64 select_idx) > +{ > + u64 mask = kvm_pmu_valid_counter_mask(vcpu); > + > + return (__vcpu_sys_reg(vcpu, PMCR_EL0) & ARMV8_PMU_PMCR_E) && > + (__vcpu_sys_reg(vcpu, PMCNTENSET_EL0) & mask & BIT(select_idx)); > +} > + > +/** > + * kvnm_pmu_event_is_chained - is a pair of counters chained and enabled > + * @vcpu: The vcpu pointer > + * @select_idx: The low counter index > + */ > +static bool kvm_pmu_event_is_chained(struct kvm_vcpu *vcpu, u64 pair_low) > +{ > + u64 eventsel, reg; > + > + reg = (pair_low + 1 == ARMV8_PMU_CYCLE_IDX) > + ? PMCCFILTR_EL0 : PMEVTYPER0_EL0 + pair_low + 1; Again, what does chaining with the cycle counter mean? You shouldn't be able to get here if pair_low >= 30. > + eventsel = __vcpu_sys_reg(vcpu, reg) & ARMV8_PMU_EVTYPE_EVENT; > + if (!armv8pmu_evtype_is_chain(eventsel)) > + return false; OK, so this works with the cycle counter only because PMCCFILTR_EL0[15:0] are RES0. I'd expect something a bit more robust. > + > + if (kvm_pmu_counter_is_enabled(vcpu, pair_low) != > + kvm_pmu_counter_is_enabled(vcpu, pair_low + 1)) > + return false; > + > + return true; > } > > /** > @@ -61,29 +126,45 @@ u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx) > */ > void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu, u64 select_idx, u64 val) > { > - u64 reg; > - struct kvm_pmu *pmu = &vcpu->arch.pmu; > - struct kvm_pmc *pmc = &pmu->pmc[select_idx]; > + u64 reg, pair_low; > > reg = (select_idx == ARMV8_PMU_CYCLE_IDX) > ? PMCCNTR_EL0 : PMEVCNTR0_EL0 + select_idx; > __vcpu_sys_reg(vcpu, reg) += (s64)val - kvm_pmu_get_counter_value(vcpu, select_idx); > > - kvm_pmu_stop_counter(vcpu, pmc); > - kvm_pmu_sync_counter_enable(vcpu, select_idx); > + pair_low = (select_idx % 2) ? select_idx - 1 : select_idx; Let's get real: pair_low = select_idx & ~1UL; > + > + /* Recreate the perf event to reflect the updated sample_period */ > + if (kvm_pmu_event_is_chained(vcpu, pair_low)) { > + kvm_pmu_stop_release_perf_event_pair(vcpu, pair_low); > + kvm_pmu_sync_counter_enable_pair(vcpu, pair_low); > + } else { > + kvm_pmu_stop_release_perf_event(vcpu, select_idx); > + kvm_pmu_sync_counter_enable(vcpu, select_idx); > + } > } > > /** > * kvm_pmu_release_perf_event - remove the perf event > * @pmc: The PMU counter pointer Missing documentation update. > */ > -static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc) > +static void kvm_pmu_release_perf_event(struct kvm_vcpu *vcpu, > + struct kvm_pmc *pmc) > { > - if (pmc->perf_event) { > - perf_event_disable(pmc->perf_event); > + struct kvm_pmu *pmu = &vcpu->arch.pmu; The PMU structure is only a container_of away. See kvm_pmc_to_vcpu. > + struct kvm_pmc *pmc_alt; > + u64 pair_alt; > + > + pair_alt = (pmc->idx % 2) ? pmc->idx - 1 : pmc->idx + 1; pair_alt = pmc->idx ^ 1; > + pmc_alt = &pmu->pmc[pair_alt]; > + > + if (pmc->perf_event) > perf_event_release_kernel(pmc->perf_event); > - pmc->perf_event = NULL; > - } > + > + if (pmc->perf_event == pmc_alt->perf_event) > + pmc_alt->perf_event = NULL; Some people say that you shouldn't use pointers to memory that has already been freed. I don't really care, but this code strikes be a pretty fragile. I'd rather see something like: if (pmc->perf_event) { if (pmc->perf_event == pmc_alt->perf_event) pmc_alt->perf_event = NULL; perf_event_release_kernel(pmc->perf_event); pmc->perf_event = NULL; } which follows the current model, and doesn't mess with pointers post release. I also wonder why you don't rely on the primitives you've defined already (kvm_pmu_event_is_chained), and start reinventing it. Another thing that strikes me as odd is that all of a sudden, you start, dealing with this as full 64bit counter. Up to this point, you've been looking at it as a two separate counters. Here, you deal with both of them at once. I think this is a better approach, but consistency is key. > + > + pmc->perf_event = NULL; > } > > /** > @@ -91,22 +172,65 @@ static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc) > * @vcpu: The vcpu pointer > * @pmc: The PMU counter pointer > * > - * If this counter has been configured to monitor some event, release it here. > + * If this counter has been configured to monitor some event, stop it here. > */ > static void kvm_pmu_stop_counter(struct kvm_vcpu *vcpu, struct kvm_pmc *pmc) > { > u64 counter, reg; > > if (pmc->perf_event) { > + perf_event_disable(pmc->perf_event); > counter = kvm_pmu_get_counter_value(vcpu, pmc->idx); > reg = (pmc->idx == ARMV8_PMU_CYCLE_IDX) > ? PMCCNTR_EL0 : PMEVCNTR0_EL0 + pmc->idx; > __vcpu_sys_reg(vcpu, reg) = counter; > - kvm_pmu_release_perf_event(pmc); The ever changing semantics of these functions are becoming pretty tedious to follow. I suggest you define the semantic of your choosing early, and stick to it. This patch is only supposed to add 64bit counter support, and seems to contain all kind of infrastructure bits that should be better located in a separate patch. > } > } > > /** > + * kvm_pmu_stop_release_perf_event_pair - stop and release a pair of counters > + * @vcpu: The vcpu pointer > + * @pmc_low: The PMU counter pointer for lower word > + * @pmc_high: The PMU counter pointer for higher word > + * > + * As chained counters share the underlying perf event, we stop them > + * both first before discarding the underlying perf event > + */ > +static void kvm_pmu_stop_release_perf_event_pair(struct kvm_vcpu *vcpu, > + u64 idx_low) > +{ > + struct kvm_pmu *pmu = &vcpu->arch.pmu; > + struct kvm_pmc *pmc_low = &pmu->pmc[idx_low]; > + struct kvm_pmc *pmc_high = &pmu->pmc[idx_low + 1]; > + > + /* Stopping a counter involves adding the perf event value to the > + * vcpu sys register value prior to releasing the perf event. As > + * kvm_pmu_get_counter_value may depend on the low counter value we > + * must always stop the high counter first > + */ Comment style (I think Christoffer mentioned this already). > + kvm_pmu_stop_counter(vcpu, pmc_high); > + kvm_pmu_stop_counter(vcpu, pmc_low); > + > + kvm_pmu_release_perf_event(vcpu, pmc_high); > + kvm_pmu_release_perf_event(vcpu, pmc_low); > +} > + > +/** > + * kvm_pmu_stop_release_perf_event - stop and release a counter > + * @vcpu: The vcpu pointer > + * @select_idx: The counter index > + */ > +static void kvm_pmu_stop_release_perf_event(struct kvm_vcpu *vcpu, > + u64 select_idx) > +{ > + struct kvm_pmu *pmu = &vcpu->arch.pmu; > + struct kvm_pmc *pmc = &pmu->pmc[select_idx]; > + > + kvm_pmu_stop_counter(vcpu, pmc); > + kvm_pmu_release_perf_event(vcpu, pmc); > +} OK, I definitely dislike this kind of construct. Why don't we have a single 'stop_release' function that has deals with everything? You already have the infrastructure. Use it to introduce some abstraction. > + > +/** > * kvm_pmu_vcpu_reset - reset pmu state for cpu > * @vcpu: The vcpu pointer > * > @@ -117,7 +241,7 @@ void kvm_pmu_vcpu_reset(struct kvm_vcpu *vcpu) > struct kvm_pmu *pmu = &vcpu->arch.pmu; > > for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i++) { > - kvm_pmu_stop_counter(vcpu, &pmu->pmc[i]); > + kvm_pmu_stop_release_perf_event(vcpu, i); > pmu->pmc[i].idx = i; > pmu->pmc[i].bitmask = 0xffffffffUL; > } > @@ -134,7 +258,7 @@ void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu) > struct kvm_pmu *pmu = &vcpu->arch.pmu; > > for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i++) > - kvm_pmu_release_perf_event(&pmu->pmc[i]); > + kvm_pmu_release_perf_event(vcpu, &pmu->pmc[i]); > } > > u64 kvm_pmu_valid_counter_mask(struct kvm_vcpu *vcpu) > @@ -167,53 +291,115 @@ static void kvm_pmu_enable_counter(struct kvm_vcpu *vcpu, u64 select_idx) > } > > /** > + * kvm_pmu_sync_counter_enable - reenable a counter if it should be enabled Why the 'if it should be enabled'? Shouldn't it be 'just do it'? > + * @vcpu: The vcpu pointer > + * @select_idx: The counter index > + */ > +static void kvm_pmu_sync_counter_enable(struct kvm_vcpu *vcpu, > + u64 select_idx) > +{ > + kvm_pmu_enable_counter_mask(vcpu, BIT(select_idx)); > +} > + > +/** > + * kvm_pmu_sync_counter_enable_pair - reenable a pair if they should be enabled > + * @vcpu: The vcpu pointer > + * @pair_low: The low counter index > + */ > +static void kvm_pmu_sync_counter_enable_pair(struct kvm_vcpu *vcpu, u64 pair_low) > +{ > + kvm_pmu_enable_counter_mask(vcpu, BIT(pair_low) | BIT(pair_low + 1)); > +} > + > +/** > + * kvm_pmu_enable_counter_pair - enable counters pair at a time > + * @vcpu: The vcpu pointer > + * @val: counters to enable > + * @pair_low: The low counter index > + */ > +static void kvm_pmu_enable_counter_pair(struct kvm_vcpu *vcpu, u64 val, > + u64 pair_low) > +{ > + struct kvm_pmu *pmu = &vcpu->arch.pmu; > + struct kvm_pmc *pmc_low = &pmu->pmc[pair_low]; > + struct kvm_pmc *pmc_high = &pmu->pmc[pair_low + 1]; > + > + if (kvm_pmu_event_is_chained(vcpu, pair_low)) { > + if (pmc_low->perf_event != pmc_high->perf_event) > + kvm_pmu_stop_release_perf_event_pair(vcpu, pair_low); I find this double condition very odd. How can the second be false if the first is true? Or rather, why does the first condition matter at all? > + } > + > + if (val & BIT(pair_low)) > + kvm_pmu_enable_counter(vcpu, pair_low); > + > + if (val & BIT(pair_low + 1)) > + kvm_pmu_enable_counter(vcpu, pair_low + 1); > +} > + > +/** > * kvm_pmu_enable_counter_mask - enable selected PMU counters > * @vcpu: The vcpu pointer > - * @val: the value guest writes to PMCNTENSET register > + * @val: the value guest writes to PMCNTENSET register or a subset A subset? of what? from where? > * > * Call perf_event_enable to start counting the perf event > */ > void kvm_pmu_enable_counter_mask(struct kvm_vcpu *vcpu, u64 val) > { > int i; > + u64 mask = kvm_pmu_valid_counter_mask(vcpu); > + u64 set = __vcpu_sys_reg(vcpu, PMCNTENSET_EL0) & mask; > > if (!(__vcpu_sys_reg(vcpu, PMCR_EL0) & ARMV8_PMU_PMCR_E) || !val) > return; > > - for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i++) { > - if (!(val & BIT(i))) > - continue; > - > - kvm_pmu_enable_counter(vcpu, i); > - } > + for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i += 2) > + kvm_pmu_enable_counter_pair(vcpu, val & set, i); > } > > /** > - * kvm_pmu_sync_counter_enable - reenable a counter if it should be enabled > + * kvm_pmu_disable_counter - disable selected PMU counter > * @vcpu: The vcpu pointer > * @select_idx: The counter index > */ > -static void kvm_pmu_sync_counter_enable(struct kvm_vcpu *vcpu, > - u64 select_idx) > +static void kvm_pmu_disable_counter(struct kvm_vcpu *vcpu, u64 select_idx) > { > - u64 set = __vcpu_sys_reg(vcpu, PMCNTENSET_EL0); > + struct kvm_pmu *pmu = &vcpu->arch.pmu; > + struct kvm_pmc *pmc = &pmu->pmc[select_idx]; > > - if (set & BIT(select_idx)) > - kvm_pmu_enable_counter_mask(vcpu, BIT(select_idx)); > + if (pmc->perf_event) { > + perf_event_disable(pmc->perf_event); > + if (pmc->perf_event->state != PERF_EVENT_STATE_ACTIVE) > + kvm_debug("fail to enable perf event\n"); > + } > } > > /** > - * kvm_pmu_disable_counter - disable selected PMU counter > - * @vcpu: The vcpu pointer > - * @pmc: The counter to disable > + * kvm_pmu_disable_counter_pair - disable counters pair at a time > + * @val: counters to disable > + * @pair_low: The low counter index > */ > -static void kvm_pmu_disable_counter(struct kvm_vcpu *vcpu, u64 select_idx) > +static void kvm_pmu_disable_counter_pair(struct kvm_vcpu *vcpu, u64 val, > + u64 pair_low) > { > struct kvm_pmu *pmu = &vcpu->arch.pmu; > - struct kvm_pmc *pmc = &pmu->pmc[select_idx]; > + struct kvm_pmc *pmc_low = &pmu->pmc[pair_low]; > + struct kvm_pmc *pmc_high = &pmu->pmc[pair_low + 1]; > + > + if (!kvm_pmu_event_is_chained(vcpu, pair_low)) { > + if (pmc_low->perf_event == pmc_high->perf_event) { Again, I find this construct troubling. > + if (pmc_low->perf_event) { > + kvm_pmu_stop_release_perf_event_pair(vcpu, > + pair_low); > + kvm_pmu_sync_counter_enable_pair(vcpu, pair_low); > + } > + } > + } > > - if (pmc->perf_event) > - perf_event_disable(pmc->perf_event); > + if (val & BIT(pair_low)) > + kvm_pmu_disable_counter(vcpu, pair_low); > + > + if (val & BIT(pair_low + 1)) > + kvm_pmu_disable_counter(vcpu, pair_low + 1); > } > > /** > @@ -230,12 +416,8 @@ void kvm_pmu_disable_counter_mask(struct kvm_vcpu *vcpu, u64 val) > if (!val) > return; > > - for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i++) { > - if (!(val & BIT(i))) > - continue; > - > - kvm_pmu_disable_counter(vcpu, i); > - } > + for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i += 2) > + kvm_pmu_disable_counter_pair(vcpu, val, i); If disable_counter() was 64bit aware, we shouldn't need any of this. > } > > static u64 kvm_pmu_overflow_status(struct kvm_vcpu *vcpu) > @@ -346,6 +528,17 @@ static void kvm_pmu_perf_overflow(struct perf_event *perf_event, > > __vcpu_sys_reg(vcpu, PMOVSSET_EL0) |= BIT(idx); > > + if (kvm_pmu_event_is_chained(vcpu, idx)) { > + struct kvm_pmu *pmu = &vcpu->arch.pmu; > + struct kvm_pmc *pmc_high = &pmu->pmc[idx + 1]; > + > + if (!(--pmc_high->overflow_count)) { > + __vcpu_sys_reg(vcpu, PMOVSSET_EL0) |= BIT(idx + 1); > + pmc_high->overflow_count = U32_MAX + 1UL; > + } > + > + } > + > if (kvm_pmu_overflow_status(vcpu)) { > kvm_make_request(KVM_REQ_IRQ_PENDING, vcpu); > kvm_vcpu_kick(vcpu); > @@ -442,6 +635,10 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx) > select_idx != ARMV8_PMU_CYCLE_IDX) > return; > > + /* Handled by even event */ > + if (armv8pmu_evtype_is_chain(eventsel)) > + return; > + > memset(&attr, 0, sizeof(struct perf_event_attr)); > attr.type = PERF_TYPE_RAW; > attr.size = sizeof(attr); > @@ -454,6 +651,9 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx) > attr.config = (select_idx == ARMV8_PMU_CYCLE_IDX) ? > ARMV8_PMUV3_PERFCTR_CPU_CYCLES : eventsel; > > + if (kvm_pmu_event_is_chained(vcpu, select_idx)) > + attr.config1 |= PERF_ATTR_CFG1_KVM_PMU_CHAINED; > + > counter = kvm_pmu_get_counter_value(vcpu, select_idx); > /* The initial sample period (overflow count) of an event. */ > attr.sample_period = (-counter) & pmc->bitmask; > @@ -466,6 +666,14 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx) > return; > } > > + if (kvm_pmu_event_is_chained(vcpu, select_idx)) { > + struct kvm_pmc *pmc_next = &pmu->pmc[select_idx + 1]; > + > + pmc_next->perf_event = event; > + counter = kvm_pmu_get_counter_value(vcpu, select_idx + 1); > + pmc_next->overflow_count = (-counter) & pmc->bitmask; > + } > + > pmc->perf_event = event; > } > > @@ -482,17 +690,25 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx) > void kvm_pmu_set_counter_event_type(struct kvm_vcpu *vcpu, u64 data, > u64 select_idx) > { > - struct kvm_pmu *pmu = &vcpu->arch.pmu; > - struct kvm_pmc *pmc = &pmu->pmc[select_idx]; > - u64 reg, event_type = data & ARMV8_PMU_EVTYPE_MASK; > - > - kvm_pmu_stop_counter(vcpu, pmc); > + u64 eventsel, event_type, pair_low, reg; > > reg = (select_idx == ARMV8_PMU_CYCLE_IDX) > ? PMCCFILTR_EL0 : PMEVTYPER0_EL0 + select_idx; > > - __vcpu_sys_reg(vcpu, reg) = event_type; > - kvm_pmu_sync_counter_enable(vcpu, select_idx); > + eventsel = data & ARMV8_PMU_EVTYPE_EVENT; > + event_type = data & ARMV8_PMU_EVTYPE_MASK; > + pair_low = (select_idx % 2) ? select_idx - 1 : select_idx; > + > + if (kvm_pmu_event_is_chained(vcpu, pair_low) || > + armv8pmu_evtype_is_chain(eventsel)) { > + kvm_pmu_stop_release_perf_event_pair(vcpu, pair_low); > + __vcpu_sys_reg(vcpu, reg) = event_type; > + kvm_pmu_sync_counter_enable_pair(vcpu, pair_low); > + } else { > + kvm_pmu_stop_release_perf_event(vcpu, pair_low); > + __vcpu_sys_reg(vcpu, reg) = event_type; > + kvm_pmu_sync_counter_enable(vcpu, pair_low); > + } > } > > bool kvm_arm_support_pmu_v3(void) I've stopped early, as I think the way you handle pairs of register is not completely right. There is a distinct lack of abstraction which doesn't fill me with confidence, and the edge case of counter 30 requires some care. I think you should rework this patch to work with a more generic view of a counter, rather than trying to cram the single/dual counter in every single place. As it stands, I'm not planning to take this into 5.1. Hopefully you'll have the time to rework it for 5.2. Thanks, M.
Hi Marc, Thanks for the review... On Wed, Feb 20, 2019 at 01:36:59PM +0000, Marc Zyngier wrote: > On Mon, 18 Feb 2019 13:48:04 +0000 > Andrew Murray <andrew.murray@arm.com> wrote: > > > Emulate chained PMU counters by creating a single 64 bit event counter > > for a pair of chained KVM counters. > > > > Signed-off-by: Andrew Murray <andrew.murray@arm.com> > > --- > > include/kvm/arm_pmu.h | 1 + > > virt/kvm/arm/pmu.c | 322 +++++++++++++++++++++++++++++++++++++++++--------- > > 2 files changed, 270 insertions(+), 53 deletions(-) > > > > diff --git a/include/kvm/arm_pmu.h b/include/kvm/arm_pmu.h > > index b73f31b..8e691ee 100644 > > --- a/include/kvm/arm_pmu.h > > +++ b/include/kvm/arm_pmu.h > > @@ -29,6 +29,7 @@ struct kvm_pmc { > > u8 idx; /* index into the pmu->pmc array */ > > struct perf_event *perf_event; > > u64 bitmask; > > + u64 overflow_count; > > }; > > > > struct kvm_pmu { > > diff --git a/virt/kvm/arm/pmu.c b/virt/kvm/arm/pmu.c > > index 524b27f..2852dc4 100644 > > --- a/virt/kvm/arm/pmu.c > > +++ b/virt/kvm/arm/pmu.c > > @@ -24,9 +24,26 @@ > > #include <kvm/arm_pmu.h> > > #include <kvm/arm_vgic.h> > > > > +static void kvm_pmu_stop_release_perf_event_pair(struct kvm_vcpu *vcpu, > > + u64 pair_low); > > +static void kvm_pmu_stop_release_perf_event(struct kvm_vcpu *vcpu, > > + u64 select_idx); > > +static void kvm_pmu_sync_counter_enable_pair(struct kvm_vcpu *vcpu, u64 pair_low); > > static void kvm_pmu_sync_counter_enable(struct kvm_vcpu *vcpu, u64 select_idx); > > static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx); > > -static void kvm_pmu_stop_counter(struct kvm_vcpu *vcpu, struct kvm_pmc *pmc); > > + > > +#define PERF_ATTR_CFG1_KVM_PMU_CHAINED 0x1 > > + > > +/** > > + * kvm_pmu_counter_is_high_word - is select_idx high counter of 64bit event > > + * @pmc: The PMU counter pointer > > + * @select_idx: The counter index > > What is select_idx here? That should be removed. > > > + */ > > +static inline bool kvm_pmu_counter_is_high_word(struct kvm_pmc *pmc) > > +{ > > + return ((pmc->perf_event->attr.config1 & PERF_ATTR_CFG1_KVM_PMU_CHAINED) > > + && (pmc->idx % 2)); > > nit: This is usually expressed as 'idx & 1', but that's not a big deal. Noted, likewise for the others you point out later on. > > > +} > > > > /** > > * kvm_pmu_get_counter_value - get PMU counter value > > @@ -35,22 +52,70 @@ static void kvm_pmu_stop_counter(struct kvm_vcpu *vcpu, struct kvm_pmc *pmc); > > */ > > u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx) > > { > > - u64 counter, reg, enabled, running; > > + u64 counter_idx, reg, enabled, running, incr; > > struct kvm_pmu *pmu = &vcpu->arch.pmu; > > struct kvm_pmc *pmc = &pmu->pmc[select_idx]; > > > > reg = (select_idx == ARMV8_PMU_CYCLE_IDX) > > ? PMCCNTR_EL0 : PMEVCNTR0_EL0 + select_idx; > > - counter = __vcpu_sys_reg(vcpu, reg); > > + counter_idx = __vcpu_sys_reg(vcpu, reg); > > Sorry, but the _idx suffix is baffling. What you seem to have here is a > counter value, not an index. 'counter' was the right name. Gah - this was picked up in the previous review, I made the change, it's in my Git tree but somehow it didn't end up here - sorry about that. > > > > > /* The real counter value is equal to the value of counter register plus > > * the value perf event counts. > > */ > > - if (pmc->perf_event) > > - counter += perf_event_read_value(pmc->perf_event, &enabled, > > + if (pmc->perf_event) { > > + incr = perf_event_read_value(pmc->perf_event, &enabled, > > &running); > > nit: please align the function arguments. > > > > > - return counter & pmc->bitmask; > > + if (kvm_pmu_counter_is_high_word(pmc)) { > > + u64 counter_low, counter; > > + > > + reg = (select_idx == ARMV8_PMU_CYCLE_IDX) > > + ? PMCCNTR_EL0 : PMEVCNTR0_EL0 + select_idx - 1; > > Why do you cater for the cycle counter here? It would feel wrong if we > could reach this code with the cycle counter, given that it is 64bit > and doesn't need any of this. Yes this isn't needed. > > > + counter_low = __vcpu_sys_reg(vcpu, reg); > > + counter = lower_32_bits(counter_low) | (counter_idx << 32); > > + counter_idx = upper_32_bits(counter + incr); > > + } else { > > + counter_idx += incr; > > + } > > + } > > + > > + return counter_idx & pmc->bitmask; > > +} > > + > > +/** > > + * kvm_pmu_counter_is_enabled - is a counter active > > + * @vcpu: The vcpu pointer > > + * @select_idx: The counter index > > + */ > > +static bool kvm_pmu_counter_is_enabled(struct kvm_vcpu *vcpu, u64 select_idx) > > +{ > > + u64 mask = kvm_pmu_valid_counter_mask(vcpu); > > + > > + return (__vcpu_sys_reg(vcpu, PMCR_EL0) & ARMV8_PMU_PMCR_E) && > > + (__vcpu_sys_reg(vcpu, PMCNTENSET_EL0) & mask & BIT(select_idx)); > > +} > > + > > +/** > > + * kvnm_pmu_event_is_chained - is a pair of counters chained and enabled > > + * @vcpu: The vcpu pointer > > + * @select_idx: The low counter index > > + */ > > +static bool kvm_pmu_event_is_chained(struct kvm_vcpu *vcpu, u64 pair_low) > > +{ > > + u64 eventsel, reg; > > + > > + reg = (pair_low + 1 == ARMV8_PMU_CYCLE_IDX) > > + ? PMCCFILTR_EL0 : PMEVTYPER0_EL0 + pair_low + 1; > > Again, what does chaining with the cycle counter mean? You shouldn't be > able to get here if pair_low >= 30. > > > + eventsel = __vcpu_sys_reg(vcpu, reg) & ARMV8_PMU_EVTYPE_EVENT; > > + if (!armv8pmu_evtype_is_chain(eventsel)) > > + return false; > > OK, so this works with the cycle counter only because > PMCCFILTR_EL0[15:0] are RES0. I'd expect something a bit more robust. I can update this to bail out based on the value of pair_low for better clarity. > > > + > > + if (kvm_pmu_counter_is_enabled(vcpu, pair_low) != > > + kvm_pmu_counter_is_enabled(vcpu, pair_low + 1)) > > + return false; > > + > > + return true; > > } > > > > /** > > @@ -61,29 +126,45 @@ u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx) > > */ > > void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu, u64 select_idx, u64 val) > > { > > - u64 reg; > > - struct kvm_pmu *pmu = &vcpu->arch.pmu; > > - struct kvm_pmc *pmc = &pmu->pmc[select_idx]; > > + u64 reg, pair_low; > > > > reg = (select_idx == ARMV8_PMU_CYCLE_IDX) > > ? PMCCNTR_EL0 : PMEVCNTR0_EL0 + select_idx; > > __vcpu_sys_reg(vcpu, reg) += (s64)val - kvm_pmu_get_counter_value(vcpu, select_idx); > > > > - kvm_pmu_stop_counter(vcpu, pmc); > > - kvm_pmu_sync_counter_enable(vcpu, select_idx); > > + pair_low = (select_idx % 2) ? select_idx - 1 : select_idx; > > Let's get real: > > pair_low = select_idx & ~1UL; > > > + > > + /* Recreate the perf event to reflect the updated sample_period */ > > + if (kvm_pmu_event_is_chained(vcpu, pair_low)) { > > + kvm_pmu_stop_release_perf_event_pair(vcpu, pair_low); > > + kvm_pmu_sync_counter_enable_pair(vcpu, pair_low); > > + } else { > > + kvm_pmu_stop_release_perf_event(vcpu, select_idx); > > + kvm_pmu_sync_counter_enable(vcpu, select_idx); > > + } > > } > > > > /** > > * kvm_pmu_release_perf_event - remove the perf event > > * @pmc: The PMU counter pointer > > Missing documentation update. > > > */ > > -static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc) > > +static void kvm_pmu_release_perf_event(struct kvm_vcpu *vcpu, > > + struct kvm_pmc *pmc) > > { > > - if (pmc->perf_event) { > > - perf_event_disable(pmc->perf_event); > > + struct kvm_pmu *pmu = &vcpu->arch.pmu; > > The PMU structure is only a container_of away. See kvm_pmc_to_vcpu. > > > + struct kvm_pmc *pmc_alt; > > + u64 pair_alt; > > + > > + pair_alt = (pmc->idx % 2) ? pmc->idx - 1 : pmc->idx + 1; > > pair_alt = pmc->idx ^ 1; > > > + pmc_alt = &pmu->pmc[pair_alt]; > > + > > + if (pmc->perf_event) > > perf_event_release_kernel(pmc->perf_event); > > - pmc->perf_event = NULL; > > - } > > + > > + if (pmc->perf_event == pmc_alt->perf_event) > > + pmc_alt->perf_event = NULL; > > Some people say that you shouldn't use pointers to memory that has > already been freed. I don't really care, but this code strikes be a > pretty fragile. > > I'd rather see something like: > > if (pmc->perf_event) { > if (pmc->perf_event == pmc_alt->perf_event) > pmc_alt->perf_event = NULL; > > perf_event_release_kernel(pmc->perf_event); > pmc->perf_event = NULL; > } > > which follows the current model, and doesn't mess with pointers post > release. Thanks this makes sense. > I also wonder why you don't rely on the primitives you've > defined already (kvm_pmu_event_is_chained), and start reinventing it. Consider the scenario where you have two counters that form a chained counter. If the user initially enables just the low counter then despite the event types indicating its a chained counter, its actually a normal counter which we must back in perf with a 32 bit event. We reflect this in kvm_pmu_event_is_chained by considering if both counters are enabled. At this point low->perf_event will be something, high->perf_event will be NULL. If the user then additionally enables the high counter, then it becomes a chained counter backed by a 64 bit perf event. Despite it being a chained counter low->perf_event doesn't equal high->perf_event. There are other scenarios... For these reasons we can't reuse kvm_pmu_event_is_chained above. > > Another thing that strikes me as odd is that all of a sudden, you > start, dealing with this as full 64bit counter. Up to this point, > you've been looking at it as a two separate counters. Here, you deal > with both of them at once. I think this is a better approach, but > consistency is key. > > > + > > + pmc->perf_event = NULL; > > } > > > > /** > > @@ -91,22 +172,65 @@ static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc) > > * @vcpu: The vcpu pointer > > * @pmc: The PMU counter pointer > > * > > - * If this counter has been configured to monitor some event, release it here. > > + * If this counter has been configured to monitor some event, stop it here. > > */ > > static void kvm_pmu_stop_counter(struct kvm_vcpu *vcpu, struct kvm_pmc *pmc) > > { > > u64 counter, reg; > > > > if (pmc->perf_event) { > > + perf_event_disable(pmc->perf_event); > > counter = kvm_pmu_get_counter_value(vcpu, pmc->idx); > > reg = (pmc->idx == ARMV8_PMU_CYCLE_IDX) > > ? PMCCNTR_EL0 : PMEVCNTR0_EL0 + pmc->idx; > > __vcpu_sys_reg(vcpu, reg) = counter; > > - kvm_pmu_release_perf_event(pmc); > > The ever changing semantics of these functions are becoming pretty > tedious to follow. I suggest you define the semantic of your choosing > early, and stick to it. This patch is only supposed to add 64bit > counter support, and seems to contain all kind of infrastructure bits > that should be better located in a separate patch. This is relevant. We don't release the event when we stop a single counter because the underlying perf event is shared between both pmc's (if its chained). If we kill it here, then we can't read the counter value when we stop the pair'd pmc. This is why, if a counter is chained we release them as a pair in kvm_pmu_stop_release_perf_event_pair. > > > } > > } > > > > /** > > + * kvm_pmu_stop_release_perf_event_pair - stop and release a pair of counters > > + * @vcpu: The vcpu pointer > > + * @pmc_low: The PMU counter pointer for lower word > > + * @pmc_high: The PMU counter pointer for higher word > > + * > > + * As chained counters share the underlying perf event, we stop them > > + * both first before discarding the underlying perf event > > + */ > > +static void kvm_pmu_stop_release_perf_event_pair(struct kvm_vcpu *vcpu, > > + u64 idx_low) > > +{ > > + struct kvm_pmu *pmu = &vcpu->arch.pmu; > > + struct kvm_pmc *pmc_low = &pmu->pmc[idx_low]; > > + struct kvm_pmc *pmc_high = &pmu->pmc[idx_low + 1]; > > + > > + /* Stopping a counter involves adding the perf event value to the > > + * vcpu sys register value prior to releasing the perf event. As > > + * kvm_pmu_get_counter_value may depend on the low counter value we > > + * must always stop the high counter first > > + */ > > Comment style (I think Christoffer mentioned this already). I missed that. > > > + kvm_pmu_stop_counter(vcpu, pmc_high); > > + kvm_pmu_stop_counter(vcpu, pmc_low); > > + > > + kvm_pmu_release_perf_event(vcpu, pmc_high); > > + kvm_pmu_release_perf_event(vcpu, pmc_low); > > +} > > + > > +/** > > + * kvm_pmu_stop_release_perf_event - stop and release a counter > > + * @vcpu: The vcpu pointer > > + * @select_idx: The counter index > > + */ > > +static void kvm_pmu_stop_release_perf_event(struct kvm_vcpu *vcpu, > > + u64 select_idx) > > +{ > > + struct kvm_pmu *pmu = &vcpu->arch.pmu; > > + struct kvm_pmc *pmc = &pmu->pmc[select_idx]; > > + > > + kvm_pmu_stop_counter(vcpu, pmc); > > + kvm_pmu_release_perf_event(vcpu, pmc); > > +} > > OK, I definitely dislike this kind of construct. Why don't we have a > single 'stop_release' function that has deals with everything? You > already have the infrastructure. Use it to introduce some abstraction. It's possible to improve on this. Though I felt this was the simplest approach. Such a 'stop_release' function would have to determine if it's a chained counter or not and in many cases the caller of the kvm_pmu_stop_release_perf_event{,_pair} functions already knows - so it seemed to just add more code. > > > + > > +/** > > * kvm_pmu_vcpu_reset - reset pmu state for cpu > > * @vcpu: The vcpu pointer > > * > > @@ -117,7 +241,7 @@ void kvm_pmu_vcpu_reset(struct kvm_vcpu *vcpu) > > struct kvm_pmu *pmu = &vcpu->arch.pmu; > > > > for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i++) { > > - kvm_pmu_stop_counter(vcpu, &pmu->pmc[i]); > > + kvm_pmu_stop_release_perf_event(vcpu, i); > > pmu->pmc[i].idx = i; > > pmu->pmc[i].bitmask = 0xffffffffUL; > > } > > @@ -134,7 +258,7 @@ void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu) > > struct kvm_pmu *pmu = &vcpu->arch.pmu; > > > > for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i++) > > - kvm_pmu_release_perf_event(&pmu->pmc[i]); > > + kvm_pmu_release_perf_event(vcpu, &pmu->pmc[i]); > > } > > > > u64 kvm_pmu_valid_counter_mask(struct kvm_vcpu *vcpu) > > @@ -167,53 +291,115 @@ static void kvm_pmu_enable_counter(struct kvm_vcpu *vcpu, u64 select_idx) > > } > > > > /** > > + * kvm_pmu_sync_counter_enable - reenable a counter if it should be enabled > > Why the 'if it should be enabled'? Shouldn't it be 'just do it'? If the user changes the counter value, type or enable of a counter within a chained pair then we must destroy and recreate the underlying perf event. It's far easier to simply destroy the event upon these changes and then call kvm_pmu_sync_counter_enable to re-create the relevant events. Therefore we have the suitation where the PMCNTENSET bits are set but the perf backing events are not present, so we recreate the ones that should be enabled. > > > + * @vcpu: The vcpu pointer > > + * @select_idx: The counter index > > + */ > > +static void kvm_pmu_sync_counter_enable(struct kvm_vcpu *vcpu, > > + u64 select_idx) > > +{ > > + kvm_pmu_enable_counter_mask(vcpu, BIT(select_idx)); > > +} > > + > > +/** > > + * kvm_pmu_sync_counter_enable_pair - reenable a pair if they should be enabled > > + * @vcpu: The vcpu pointer > > + * @pair_low: The low counter index > > + */ > > +static void kvm_pmu_sync_counter_enable_pair(struct kvm_vcpu *vcpu, u64 pair_low) > > +{ > > + kvm_pmu_enable_counter_mask(vcpu, BIT(pair_low) | BIT(pair_low + 1)); > > +} > > + > > +/** > > + * kvm_pmu_enable_counter_pair - enable counters pair at a time > > + * @vcpu: The vcpu pointer > > + * @val: counters to enable > > + * @pair_low: The low counter index > > + */ > > +static void kvm_pmu_enable_counter_pair(struct kvm_vcpu *vcpu, u64 val, > > + u64 pair_low) > > +{ > > + struct kvm_pmu *pmu = &vcpu->arch.pmu; > > + struct kvm_pmc *pmc_low = &pmu->pmc[pair_low]; > > + struct kvm_pmc *pmc_high = &pmu->pmc[pair_low + 1]; > > + > > + if (kvm_pmu_event_is_chained(vcpu, pair_low)) { > > + if (pmc_low->perf_event != pmc_high->perf_event) > > + kvm_pmu_stop_release_perf_event_pair(vcpu, pair_low); > > I find this double condition very odd. How can the second be false if > the first is true? Or rather, why does the first condition matter at > all? I believe I've explained this above. I.e. if the event types indicate that two counters should make a pair, yet the perf events are different - then we need to destroy them and recreate them. This may happen if a user enables both counters from a pair whereas previously only one was enabled. > > > + } > > + > > + if (val & BIT(pair_low)) > > + kvm_pmu_enable_counter(vcpu, pair_low); > > + > > + if (val & BIT(pair_low + 1)) > > + kvm_pmu_enable_counter(vcpu, pair_low + 1); > > +} > > + > > +/** > > * kvm_pmu_enable_counter_mask - enable selected PMU counters > > * @vcpu: The vcpu pointer > > - * @val: the value guest writes to PMCNTENSET register > > + * @val: the value guest writes to PMCNTENSET register or a subset > > A subset? of what? from where? A subset (val) from PMCTENSET_EL0. Perhaps I can make this clearer. This comment was added to reflect the change of function name to append _mask. > > > * > > * Call perf_event_enable to start counting the perf event > > */ > > void kvm_pmu_enable_counter_mask(struct kvm_vcpu *vcpu, u64 val) > > { > > int i; > > + u64 mask = kvm_pmu_valid_counter_mask(vcpu); > > + u64 set = __vcpu_sys_reg(vcpu, PMCNTENSET_EL0) & mask; > > > > if (!(__vcpu_sys_reg(vcpu, PMCR_EL0) & ARMV8_PMU_PMCR_E) || !val) > > return; > > > > - for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i++) { > > - if (!(val & BIT(i))) > > - continue; > > - > > - kvm_pmu_enable_counter(vcpu, i); > > - } > > + for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i += 2) > > + kvm_pmu_enable_counter_pair(vcpu, val & set, i); > > } > > > > /** > > - * kvm_pmu_sync_counter_enable - reenable a counter if it should be enabled > > + * kvm_pmu_disable_counter - disable selected PMU counter > > * @vcpu: The vcpu pointer > > * @select_idx: The counter index > > */ > > -static void kvm_pmu_sync_counter_enable(struct kvm_vcpu *vcpu, > > - u64 select_idx) > > +static void kvm_pmu_disable_counter(struct kvm_vcpu *vcpu, u64 select_idx) > > { > > - u64 set = __vcpu_sys_reg(vcpu, PMCNTENSET_EL0); > > + struct kvm_pmu *pmu = &vcpu->arch.pmu; > > + struct kvm_pmc *pmc = &pmu->pmc[select_idx]; > > > > - if (set & BIT(select_idx)) > > - kvm_pmu_enable_counter_mask(vcpu, BIT(select_idx)); > > + if (pmc->perf_event) { > > + perf_event_disable(pmc->perf_event); > > + if (pmc->perf_event->state != PERF_EVENT_STATE_ACTIVE) > > + kvm_debug("fail to enable perf event\n"); > > + } > > } > > > > /** > > - * kvm_pmu_disable_counter - disable selected PMU counter > > - * @vcpu: The vcpu pointer > > - * @pmc: The counter to disable > > + * kvm_pmu_disable_counter_pair - disable counters pair at a time > > + * @val: counters to disable > > + * @pair_low: The low counter index > > */ > > -static void kvm_pmu_disable_counter(struct kvm_vcpu *vcpu, u64 select_idx) > > +static void kvm_pmu_disable_counter_pair(struct kvm_vcpu *vcpu, u64 val, > > + u64 pair_low) > > { > > struct kvm_pmu *pmu = &vcpu->arch.pmu; > > - struct kvm_pmc *pmc = &pmu->pmc[select_idx]; > > + struct kvm_pmc *pmc_low = &pmu->pmc[pair_low]; > > + struct kvm_pmc *pmc_high = &pmu->pmc[pair_low + 1]; > > + > > + if (!kvm_pmu_event_is_chained(vcpu, pair_low)) { > > + if (pmc_low->perf_event == pmc_high->perf_event) { > > Again, I find this construct troubling. The use-case here is when you have two counters in a pair enabled (second condition), and now you've disabled one of them (first condition). We need to destroy the shared perf event. > > > + if (pmc_low->perf_event) { > > + kvm_pmu_stop_release_perf_event_pair(vcpu, > > + pair_low); > > + kvm_pmu_sync_counter_enable_pair(vcpu, pair_low); > > + } > > + } > > + } > > > > - if (pmc->perf_event) > > - perf_event_disable(pmc->perf_event); > > + if (val & BIT(pair_low)) > > + kvm_pmu_disable_counter(vcpu, pair_low); > > + > > + if (val & BIT(pair_low + 1)) > > + kvm_pmu_disable_counter(vcpu, pair_low + 1); > > } > > > > /** > > @@ -230,12 +416,8 @@ void kvm_pmu_disable_counter_mask(struct kvm_vcpu *vcpu, u64 val) > > if (!val) > > return; > > > > - for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i++) { > > - if (!(val & BIT(i))) > > - continue; > > - > > - kvm_pmu_disable_counter(vcpu, i); > > - } > > + for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i += 2) > > + kvm_pmu_disable_counter_pair(vcpu, val, i); > > If disable_counter() was 64bit aware, we shouldn't need any of this. It's likely the OS will use a mix of chained and non-chained counters - and so we can't treat eveything as chained counters. If we were to combine the logic of kvm_pmu_disable_counter and kvm_pmu_disable_counter_pair we end up duplicating code - we can't disable a counter without considering the impact on its adjacent counter. Unless I'm not understanding what your point here? > > > } > > > > static u64 kvm_pmu_overflow_status(struct kvm_vcpu *vcpu) > > @@ -346,6 +528,17 @@ static void kvm_pmu_perf_overflow(struct perf_event *perf_event, > > > > __vcpu_sys_reg(vcpu, PMOVSSET_EL0) |= BIT(idx); > > > > + if (kvm_pmu_event_is_chained(vcpu, idx)) { > > + struct kvm_pmu *pmu = &vcpu->arch.pmu; > > + struct kvm_pmc *pmc_high = &pmu->pmc[idx + 1]; > > + > > + if (!(--pmc_high->overflow_count)) { > > + __vcpu_sys_reg(vcpu, PMOVSSET_EL0) |= BIT(idx + 1); > > + pmc_high->overflow_count = U32_MAX + 1UL; > > + } > > + > > + } > > + > > if (kvm_pmu_overflow_status(vcpu)) { > > kvm_make_request(KVM_REQ_IRQ_PENDING, vcpu); > > kvm_vcpu_kick(vcpu); > > @@ -442,6 +635,10 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx) > > select_idx != ARMV8_PMU_CYCLE_IDX) > > return; > > > > + /* Handled by even event */ > > + if (armv8pmu_evtype_is_chain(eventsel)) > > + return; > > + > > memset(&attr, 0, sizeof(struct perf_event_attr)); > > attr.type = PERF_TYPE_RAW; > > attr.size = sizeof(attr); > > @@ -454,6 +651,9 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx) > > attr.config = (select_idx == ARMV8_PMU_CYCLE_IDX) ? > > ARMV8_PMUV3_PERFCTR_CPU_CYCLES : eventsel; > > > > + if (kvm_pmu_event_is_chained(vcpu, select_idx)) > > + attr.config1 |= PERF_ATTR_CFG1_KVM_PMU_CHAINED; > > + > > counter = kvm_pmu_get_counter_value(vcpu, select_idx); > > /* The initial sample period (overflow count) of an event. */ > > attr.sample_period = (-counter) & pmc->bitmask; > > @@ -466,6 +666,14 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx) > > return; > > } > > > > + if (kvm_pmu_event_is_chained(vcpu, select_idx)) { > > + struct kvm_pmc *pmc_next = &pmu->pmc[select_idx + 1]; > > + > > + pmc_next->perf_event = event; > > + counter = kvm_pmu_get_counter_value(vcpu, select_idx + 1); > > + pmc_next->overflow_count = (-counter) & pmc->bitmask; > > + } > > + > > pmc->perf_event = event; > > } > > > > @@ -482,17 +690,25 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx) > > void kvm_pmu_set_counter_event_type(struct kvm_vcpu *vcpu, u64 data, > > u64 select_idx) > > { > > - struct kvm_pmu *pmu = &vcpu->arch.pmu; > > - struct kvm_pmc *pmc = &pmu->pmc[select_idx]; > > - u64 reg, event_type = data & ARMV8_PMU_EVTYPE_MASK; > > - > > - kvm_pmu_stop_counter(vcpu, pmc); > > + u64 eventsel, event_type, pair_low, reg; > > > > reg = (select_idx == ARMV8_PMU_CYCLE_IDX) > > ? PMCCFILTR_EL0 : PMEVTYPER0_EL0 + select_idx; > > > > - __vcpu_sys_reg(vcpu, reg) = event_type; > > - kvm_pmu_sync_counter_enable(vcpu, select_idx); > > + eventsel = data & ARMV8_PMU_EVTYPE_EVENT; > > + event_type = data & ARMV8_PMU_EVTYPE_MASK; > > + pair_low = (select_idx % 2) ? select_idx - 1 : select_idx; > > + > > + if (kvm_pmu_event_is_chained(vcpu, pair_low) || > > + armv8pmu_evtype_is_chain(eventsel)) { > > + kvm_pmu_stop_release_perf_event_pair(vcpu, pair_low); > > + __vcpu_sys_reg(vcpu, reg) = event_type; > > + kvm_pmu_sync_counter_enable_pair(vcpu, pair_low); > > + } else { > > + kvm_pmu_stop_release_perf_event(vcpu, pair_low); > > + __vcpu_sys_reg(vcpu, reg) = event_type; > > + kvm_pmu_sync_counter_enable(vcpu, pair_low); > > + } > > } > > > > bool kvm_arm_support_pmu_v3(void) > > I've stopped early, as I think the way you handle pairs of register is > not completely right. There is a distinct lack of abstraction which > doesn't fill me with confidence, and the edge case of counter 30 > requires some care. > > I think you should rework this patch to work with a more generic view > of a counter, rather than trying to cram the single/dual counter in > every single place. I'm struggling to find a better way to do this. It's made so much more complex due to the fact that we have one perf event backing two counters and that this event needs to be recreated whenever the user does anything with the PMU registers. This dependency is a challenge when the KVM guest can treat the counters as 32bit non-chained. The only way I can see to improve this is to emulate the behavior of chained counters by counting overflows, but then we don't take advantage of the host OS's 64 bit perf events (I'm not even sure how much of a benefit this is). With this approach we reduce the complexity to the kvm_pmu_perf_overflow handler which would be the only place where a dependency between two counters is considered. Everywhere else we treat the counters as non-chained 32 bit. This would so much simpler. Or is there another way? > > As it stands, I'm not planning to take this into 5.1. Hopefully you'll > have the time to rework it for 5.2. That's no problem, it's ready when it's ready. And this seems to be a complicated beast for what it is :|. Thanks, Andrew Murray > > Thanks, > > M. > -- > Without deviation from the norm, progress is not possible.
diff --git a/include/kvm/arm_pmu.h b/include/kvm/arm_pmu.h index b73f31b..8e691ee 100644 --- a/include/kvm/arm_pmu.h +++ b/include/kvm/arm_pmu.h @@ -29,6 +29,7 @@ struct kvm_pmc { u8 idx; /* index into the pmu->pmc array */ struct perf_event *perf_event; u64 bitmask; + u64 overflow_count; }; struct kvm_pmu { diff --git a/virt/kvm/arm/pmu.c b/virt/kvm/arm/pmu.c index 524b27f..2852dc4 100644 --- a/virt/kvm/arm/pmu.c +++ b/virt/kvm/arm/pmu.c @@ -24,9 +24,26 @@ #include <kvm/arm_pmu.h> #include <kvm/arm_vgic.h> +static void kvm_pmu_stop_release_perf_event_pair(struct kvm_vcpu *vcpu, + u64 pair_low); +static void kvm_pmu_stop_release_perf_event(struct kvm_vcpu *vcpu, + u64 select_idx); +static void kvm_pmu_sync_counter_enable_pair(struct kvm_vcpu *vcpu, u64 pair_low); static void kvm_pmu_sync_counter_enable(struct kvm_vcpu *vcpu, u64 select_idx); static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx); -static void kvm_pmu_stop_counter(struct kvm_vcpu *vcpu, struct kvm_pmc *pmc); + +#define PERF_ATTR_CFG1_KVM_PMU_CHAINED 0x1 + +/** + * kvm_pmu_counter_is_high_word - is select_idx high counter of 64bit event + * @pmc: The PMU counter pointer + * @select_idx: The counter index + */ +static inline bool kvm_pmu_counter_is_high_word(struct kvm_pmc *pmc) +{ + return ((pmc->perf_event->attr.config1 & PERF_ATTR_CFG1_KVM_PMU_CHAINED) + && (pmc->idx % 2)); +} /** * kvm_pmu_get_counter_value - get PMU counter value @@ -35,22 +52,70 @@ static void kvm_pmu_stop_counter(struct kvm_vcpu *vcpu, struct kvm_pmc *pmc); */ u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx) { - u64 counter, reg, enabled, running; + u64 counter_idx, reg, enabled, running, incr; struct kvm_pmu *pmu = &vcpu->arch.pmu; struct kvm_pmc *pmc = &pmu->pmc[select_idx]; reg = (select_idx == ARMV8_PMU_CYCLE_IDX) ? PMCCNTR_EL0 : PMEVCNTR0_EL0 + select_idx; - counter = __vcpu_sys_reg(vcpu, reg); + counter_idx = __vcpu_sys_reg(vcpu, reg); /* The real counter value is equal to the value of counter register plus * the value perf event counts. */ - if (pmc->perf_event) - counter += perf_event_read_value(pmc->perf_event, &enabled, + if (pmc->perf_event) { + incr = perf_event_read_value(pmc->perf_event, &enabled, &running); - return counter & pmc->bitmask; + if (kvm_pmu_counter_is_high_word(pmc)) { + u64 counter_low, counter; + + reg = (select_idx == ARMV8_PMU_CYCLE_IDX) + ? PMCCNTR_EL0 : PMEVCNTR0_EL0 + select_idx - 1; + counter_low = __vcpu_sys_reg(vcpu, reg); + counter = lower_32_bits(counter_low) | (counter_idx << 32); + counter_idx = upper_32_bits(counter + incr); + } else { + counter_idx += incr; + } + } + + return counter_idx & pmc->bitmask; +} + +/** + * kvm_pmu_counter_is_enabled - is a counter active + * @vcpu: The vcpu pointer + * @select_idx: The counter index + */ +static bool kvm_pmu_counter_is_enabled(struct kvm_vcpu *vcpu, u64 select_idx) +{ + u64 mask = kvm_pmu_valid_counter_mask(vcpu); + + return (__vcpu_sys_reg(vcpu, PMCR_EL0) & ARMV8_PMU_PMCR_E) && + (__vcpu_sys_reg(vcpu, PMCNTENSET_EL0) & mask & BIT(select_idx)); +} + +/** + * kvnm_pmu_event_is_chained - is a pair of counters chained and enabled + * @vcpu: The vcpu pointer + * @select_idx: The low counter index + */ +static bool kvm_pmu_event_is_chained(struct kvm_vcpu *vcpu, u64 pair_low) +{ + u64 eventsel, reg; + + reg = (pair_low + 1 == ARMV8_PMU_CYCLE_IDX) + ? PMCCFILTR_EL0 : PMEVTYPER0_EL0 + pair_low + 1; + eventsel = __vcpu_sys_reg(vcpu, reg) & ARMV8_PMU_EVTYPE_EVENT; + if (!armv8pmu_evtype_is_chain(eventsel)) + return false; + + if (kvm_pmu_counter_is_enabled(vcpu, pair_low) != + kvm_pmu_counter_is_enabled(vcpu, pair_low + 1)) + return false; + + return true; } /** @@ -61,29 +126,45 @@ u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx) */ void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu, u64 select_idx, u64 val) { - u64 reg; - struct kvm_pmu *pmu = &vcpu->arch.pmu; - struct kvm_pmc *pmc = &pmu->pmc[select_idx]; + u64 reg, pair_low; reg = (select_idx == ARMV8_PMU_CYCLE_IDX) ? PMCCNTR_EL0 : PMEVCNTR0_EL0 + select_idx; __vcpu_sys_reg(vcpu, reg) += (s64)val - kvm_pmu_get_counter_value(vcpu, select_idx); - kvm_pmu_stop_counter(vcpu, pmc); - kvm_pmu_sync_counter_enable(vcpu, select_idx); + pair_low = (select_idx % 2) ? select_idx - 1 : select_idx; + + /* Recreate the perf event to reflect the updated sample_period */ + if (kvm_pmu_event_is_chained(vcpu, pair_low)) { + kvm_pmu_stop_release_perf_event_pair(vcpu, pair_low); + kvm_pmu_sync_counter_enable_pair(vcpu, pair_low); + } else { + kvm_pmu_stop_release_perf_event(vcpu, select_idx); + kvm_pmu_sync_counter_enable(vcpu, select_idx); + } } /** * kvm_pmu_release_perf_event - remove the perf event * @pmc: The PMU counter pointer */ -static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc) +static void kvm_pmu_release_perf_event(struct kvm_vcpu *vcpu, + struct kvm_pmc *pmc) { - if (pmc->perf_event) { - perf_event_disable(pmc->perf_event); + struct kvm_pmu *pmu = &vcpu->arch.pmu; + struct kvm_pmc *pmc_alt; + u64 pair_alt; + + pair_alt = (pmc->idx % 2) ? pmc->idx - 1 : pmc->idx + 1; + pmc_alt = &pmu->pmc[pair_alt]; + + if (pmc->perf_event) perf_event_release_kernel(pmc->perf_event); - pmc->perf_event = NULL; - } + + if (pmc->perf_event == pmc_alt->perf_event) + pmc_alt->perf_event = NULL; + + pmc->perf_event = NULL; } /** @@ -91,22 +172,65 @@ static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc) * @vcpu: The vcpu pointer * @pmc: The PMU counter pointer * - * If this counter has been configured to monitor some event, release it here. + * If this counter has been configured to monitor some event, stop it here. */ static void kvm_pmu_stop_counter(struct kvm_vcpu *vcpu, struct kvm_pmc *pmc) { u64 counter, reg; if (pmc->perf_event) { + perf_event_disable(pmc->perf_event); counter = kvm_pmu_get_counter_value(vcpu, pmc->idx); reg = (pmc->idx == ARMV8_PMU_CYCLE_IDX) ? PMCCNTR_EL0 : PMEVCNTR0_EL0 + pmc->idx; __vcpu_sys_reg(vcpu, reg) = counter; - kvm_pmu_release_perf_event(pmc); } } /** + * kvm_pmu_stop_release_perf_event_pair - stop and release a pair of counters + * @vcpu: The vcpu pointer + * @pmc_low: The PMU counter pointer for lower word + * @pmc_high: The PMU counter pointer for higher word + * + * As chained counters share the underlying perf event, we stop them + * both first before discarding the underlying perf event + */ +static void kvm_pmu_stop_release_perf_event_pair(struct kvm_vcpu *vcpu, + u64 idx_low) +{ + struct kvm_pmu *pmu = &vcpu->arch.pmu; + struct kvm_pmc *pmc_low = &pmu->pmc[idx_low]; + struct kvm_pmc *pmc_high = &pmu->pmc[idx_low + 1]; + + /* Stopping a counter involves adding the perf event value to the + * vcpu sys register value prior to releasing the perf event. As + * kvm_pmu_get_counter_value may depend on the low counter value we + * must always stop the high counter first + */ + kvm_pmu_stop_counter(vcpu, pmc_high); + kvm_pmu_stop_counter(vcpu, pmc_low); + + kvm_pmu_release_perf_event(vcpu, pmc_high); + kvm_pmu_release_perf_event(vcpu, pmc_low); +} + +/** + * kvm_pmu_stop_release_perf_event - stop and release a counter + * @vcpu: The vcpu pointer + * @select_idx: The counter index + */ +static void kvm_pmu_stop_release_perf_event(struct kvm_vcpu *vcpu, + u64 select_idx) +{ + struct kvm_pmu *pmu = &vcpu->arch.pmu; + struct kvm_pmc *pmc = &pmu->pmc[select_idx]; + + kvm_pmu_stop_counter(vcpu, pmc); + kvm_pmu_release_perf_event(vcpu, pmc); +} + +/** * kvm_pmu_vcpu_reset - reset pmu state for cpu * @vcpu: The vcpu pointer * @@ -117,7 +241,7 @@ void kvm_pmu_vcpu_reset(struct kvm_vcpu *vcpu) struct kvm_pmu *pmu = &vcpu->arch.pmu; for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i++) { - kvm_pmu_stop_counter(vcpu, &pmu->pmc[i]); + kvm_pmu_stop_release_perf_event(vcpu, i); pmu->pmc[i].idx = i; pmu->pmc[i].bitmask = 0xffffffffUL; } @@ -134,7 +258,7 @@ void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu) struct kvm_pmu *pmu = &vcpu->arch.pmu; for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i++) - kvm_pmu_release_perf_event(&pmu->pmc[i]); + kvm_pmu_release_perf_event(vcpu, &pmu->pmc[i]); } u64 kvm_pmu_valid_counter_mask(struct kvm_vcpu *vcpu) @@ -167,53 +291,115 @@ static void kvm_pmu_enable_counter(struct kvm_vcpu *vcpu, u64 select_idx) } /** + * kvm_pmu_sync_counter_enable - reenable a counter if it should be enabled + * @vcpu: The vcpu pointer + * @select_idx: The counter index + */ +static void kvm_pmu_sync_counter_enable(struct kvm_vcpu *vcpu, + u64 select_idx) +{ + kvm_pmu_enable_counter_mask(vcpu, BIT(select_idx)); +} + +/** + * kvm_pmu_sync_counter_enable_pair - reenable a pair if they should be enabled + * @vcpu: The vcpu pointer + * @pair_low: The low counter index + */ +static void kvm_pmu_sync_counter_enable_pair(struct kvm_vcpu *vcpu, u64 pair_low) +{ + kvm_pmu_enable_counter_mask(vcpu, BIT(pair_low) | BIT(pair_low + 1)); +} + +/** + * kvm_pmu_enable_counter_pair - enable counters pair at a time + * @vcpu: The vcpu pointer + * @val: counters to enable + * @pair_low: The low counter index + */ +static void kvm_pmu_enable_counter_pair(struct kvm_vcpu *vcpu, u64 val, + u64 pair_low) +{ + struct kvm_pmu *pmu = &vcpu->arch.pmu; + struct kvm_pmc *pmc_low = &pmu->pmc[pair_low]; + struct kvm_pmc *pmc_high = &pmu->pmc[pair_low + 1]; + + if (kvm_pmu_event_is_chained(vcpu, pair_low)) { + if (pmc_low->perf_event != pmc_high->perf_event) + kvm_pmu_stop_release_perf_event_pair(vcpu, pair_low); + } + + if (val & BIT(pair_low)) + kvm_pmu_enable_counter(vcpu, pair_low); + + if (val & BIT(pair_low + 1)) + kvm_pmu_enable_counter(vcpu, pair_low + 1); +} + +/** * kvm_pmu_enable_counter_mask - enable selected PMU counters * @vcpu: The vcpu pointer - * @val: the value guest writes to PMCNTENSET register + * @val: the value guest writes to PMCNTENSET register or a subset * * Call perf_event_enable to start counting the perf event */ void kvm_pmu_enable_counter_mask(struct kvm_vcpu *vcpu, u64 val) { int i; + u64 mask = kvm_pmu_valid_counter_mask(vcpu); + u64 set = __vcpu_sys_reg(vcpu, PMCNTENSET_EL0) & mask; if (!(__vcpu_sys_reg(vcpu, PMCR_EL0) & ARMV8_PMU_PMCR_E) || !val) return; - for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i++) { - if (!(val & BIT(i))) - continue; - - kvm_pmu_enable_counter(vcpu, i); - } + for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i += 2) + kvm_pmu_enable_counter_pair(vcpu, val & set, i); } /** - * kvm_pmu_sync_counter_enable - reenable a counter if it should be enabled + * kvm_pmu_disable_counter - disable selected PMU counter * @vcpu: The vcpu pointer * @select_idx: The counter index */ -static void kvm_pmu_sync_counter_enable(struct kvm_vcpu *vcpu, - u64 select_idx) +static void kvm_pmu_disable_counter(struct kvm_vcpu *vcpu, u64 select_idx) { - u64 set = __vcpu_sys_reg(vcpu, PMCNTENSET_EL0); + struct kvm_pmu *pmu = &vcpu->arch.pmu; + struct kvm_pmc *pmc = &pmu->pmc[select_idx]; - if (set & BIT(select_idx)) - kvm_pmu_enable_counter_mask(vcpu, BIT(select_idx)); + if (pmc->perf_event) { + perf_event_disable(pmc->perf_event); + if (pmc->perf_event->state != PERF_EVENT_STATE_ACTIVE) + kvm_debug("fail to enable perf event\n"); + } } /** - * kvm_pmu_disable_counter - disable selected PMU counter - * @vcpu: The vcpu pointer - * @pmc: The counter to disable + * kvm_pmu_disable_counter_pair - disable counters pair at a time + * @val: counters to disable + * @pair_low: The low counter index */ -static void kvm_pmu_disable_counter(struct kvm_vcpu *vcpu, u64 select_idx) +static void kvm_pmu_disable_counter_pair(struct kvm_vcpu *vcpu, u64 val, + u64 pair_low) { struct kvm_pmu *pmu = &vcpu->arch.pmu; - struct kvm_pmc *pmc = &pmu->pmc[select_idx]; + struct kvm_pmc *pmc_low = &pmu->pmc[pair_low]; + struct kvm_pmc *pmc_high = &pmu->pmc[pair_low + 1]; + + if (!kvm_pmu_event_is_chained(vcpu, pair_low)) { + if (pmc_low->perf_event == pmc_high->perf_event) { + if (pmc_low->perf_event) { + kvm_pmu_stop_release_perf_event_pair(vcpu, + pair_low); + kvm_pmu_sync_counter_enable_pair(vcpu, pair_low); + } + } + } - if (pmc->perf_event) - perf_event_disable(pmc->perf_event); + if (val & BIT(pair_low)) + kvm_pmu_disable_counter(vcpu, pair_low); + + if (val & BIT(pair_low + 1)) + kvm_pmu_disable_counter(vcpu, pair_low + 1); } /** @@ -230,12 +416,8 @@ void kvm_pmu_disable_counter_mask(struct kvm_vcpu *vcpu, u64 val) if (!val) return; - for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i++) { - if (!(val & BIT(i))) - continue; - - kvm_pmu_disable_counter(vcpu, i); - } + for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i += 2) + kvm_pmu_disable_counter_pair(vcpu, val, i); } static u64 kvm_pmu_overflow_status(struct kvm_vcpu *vcpu) @@ -346,6 +528,17 @@ static void kvm_pmu_perf_overflow(struct perf_event *perf_event, __vcpu_sys_reg(vcpu, PMOVSSET_EL0) |= BIT(idx); + if (kvm_pmu_event_is_chained(vcpu, idx)) { + struct kvm_pmu *pmu = &vcpu->arch.pmu; + struct kvm_pmc *pmc_high = &pmu->pmc[idx + 1]; + + if (!(--pmc_high->overflow_count)) { + __vcpu_sys_reg(vcpu, PMOVSSET_EL0) |= BIT(idx + 1); + pmc_high->overflow_count = U32_MAX + 1UL; + } + + } + if (kvm_pmu_overflow_status(vcpu)) { kvm_make_request(KVM_REQ_IRQ_PENDING, vcpu); kvm_vcpu_kick(vcpu); @@ -442,6 +635,10 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx) select_idx != ARMV8_PMU_CYCLE_IDX) return; + /* Handled by even event */ + if (armv8pmu_evtype_is_chain(eventsel)) + return; + memset(&attr, 0, sizeof(struct perf_event_attr)); attr.type = PERF_TYPE_RAW; attr.size = sizeof(attr); @@ -454,6 +651,9 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx) attr.config = (select_idx == ARMV8_PMU_CYCLE_IDX) ? ARMV8_PMUV3_PERFCTR_CPU_CYCLES : eventsel; + if (kvm_pmu_event_is_chained(vcpu, select_idx)) + attr.config1 |= PERF_ATTR_CFG1_KVM_PMU_CHAINED; + counter = kvm_pmu_get_counter_value(vcpu, select_idx); /* The initial sample period (overflow count) of an event. */ attr.sample_period = (-counter) & pmc->bitmask; @@ -466,6 +666,14 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx) return; } + if (kvm_pmu_event_is_chained(vcpu, select_idx)) { + struct kvm_pmc *pmc_next = &pmu->pmc[select_idx + 1]; + + pmc_next->perf_event = event; + counter = kvm_pmu_get_counter_value(vcpu, select_idx + 1); + pmc_next->overflow_count = (-counter) & pmc->bitmask; + } + pmc->perf_event = event; } @@ -482,17 +690,25 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx) void kvm_pmu_set_counter_event_type(struct kvm_vcpu *vcpu, u64 data, u64 select_idx) { - struct kvm_pmu *pmu = &vcpu->arch.pmu; - struct kvm_pmc *pmc = &pmu->pmc[select_idx]; - u64 reg, event_type = data & ARMV8_PMU_EVTYPE_MASK; - - kvm_pmu_stop_counter(vcpu, pmc); + u64 eventsel, event_type, pair_low, reg; reg = (select_idx == ARMV8_PMU_CYCLE_IDX) ? PMCCFILTR_EL0 : PMEVTYPER0_EL0 + select_idx; - __vcpu_sys_reg(vcpu, reg) = event_type; - kvm_pmu_sync_counter_enable(vcpu, select_idx); + eventsel = data & ARMV8_PMU_EVTYPE_EVENT; + event_type = data & ARMV8_PMU_EVTYPE_MASK; + pair_low = (select_idx % 2) ? select_idx - 1 : select_idx; + + if (kvm_pmu_event_is_chained(vcpu, pair_low) || + armv8pmu_evtype_is_chain(eventsel)) { + kvm_pmu_stop_release_perf_event_pair(vcpu, pair_low); + __vcpu_sys_reg(vcpu, reg) = event_type; + kvm_pmu_sync_counter_enable_pair(vcpu, pair_low); + } else { + kvm_pmu_stop_release_perf_event(vcpu, pair_low); + __vcpu_sys_reg(vcpu, reg) = event_type; + kvm_pmu_sync_counter_enable(vcpu, pair_low); + } } bool kvm_arm_support_pmu_v3(void)
Emulate chained PMU counters by creating a single 64 bit event counter for a pair of chained KVM counters. Signed-off-by: Andrew Murray <andrew.murray@arm.com> --- include/kvm/arm_pmu.h | 1 + virt/kvm/arm/pmu.c | 322 +++++++++++++++++++++++++++++++++++++++++--------- 2 files changed, 270 insertions(+), 53 deletions(-)