Message ID | 20250302220112.17653-7-dongli.zhang@oracle.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | target/i386/kvm/pmu: PMU Enhancement, Bugfix and Cleanup | expand |
On 3/3/2025 6:00 AM, Dongli Zhang wrote: > AMD does not have what is commonly referred to as an architectural PMU. > Therefore, we need to rename the following variables to be applicable for > both Intel and AMD: > > - has_architectural_pmu_version > - num_architectural_pmu_gp_counters > - num_architectural_pmu_fixed_counters > > For Intel processors, the meaning of has_pmu_version remains unchanged. > > For AMD processors: > > has_pmu_version == 1 corresponds to versions before AMD PerfMonV2. > has_pmu_version == 2 corresponds to AMD PerfMonV2. > > Signed-off-by: Dongli Zhang <dongli.zhang@oracle.com> > --- > target/i386/kvm/kvm.c | 49 ++++++++++++++++++++++++------------------- > 1 file changed, 28 insertions(+), 21 deletions(-) > > diff --git a/target/i386/kvm/kvm.c b/target/i386/kvm/kvm.c > index 8f293ffd61..e895d22f94 100644 > --- a/target/i386/kvm/kvm.c > +++ b/target/i386/kvm/kvm.c > @@ -164,9 +164,16 @@ static bool has_msr_perf_capabs; > static bool has_msr_pkrs; > static bool has_msr_hwcr; > > -static uint32_t has_architectural_pmu_version; > -static uint32_t num_architectural_pmu_gp_counters; > -static uint32_t num_architectural_pmu_fixed_counters; > +/* > + * For Intel processors, the meaning is the architectural PMU version > + * number. > + * > + * For AMD processors: 1 corresponds to the prior versions, and 2 > + * corresponds to AMD PerfMonV2. > + */ > +static uint32_t has_pmu_version; > +static uint32_t num_pmu_gp_counters; > +static uint32_t num_pmu_fixed_counters; > > static int has_xsave2; > static int has_xcrs; > @@ -2072,24 +2079,24 @@ static void kvm_init_pmu_info(CPUX86State *env) > > cpu_x86_cpuid(env, 0x0a, 0, &eax, &unused, &unused, &edx); > > - has_architectural_pmu_version = eax & 0xff; > - if (has_architectural_pmu_version > 0) { > - num_architectural_pmu_gp_counters = (eax & 0xff00) >> 8; > + has_pmu_version = eax & 0xff; > + if (has_pmu_version > 0) { > + num_pmu_gp_counters = (eax & 0xff00) >> 8; > > /* > * Shouldn't be more than 32, since that's the number of bits > * available in EBX to tell us _which_ counters are available. > * Play it safe. > */ > - if (num_architectural_pmu_gp_counters > MAX_GP_COUNTERS) { > - num_architectural_pmu_gp_counters = MAX_GP_COUNTERS; > + if (num_pmu_gp_counters > MAX_GP_COUNTERS) { > + num_pmu_gp_counters = MAX_GP_COUNTERS; > } > > - if (has_architectural_pmu_version > 1) { > - num_architectural_pmu_fixed_counters = edx & 0x1f; > + if (has_pmu_version > 1) { > + num_pmu_fixed_counters = edx & 0x1f; > > - if (num_architectural_pmu_fixed_counters > MAX_FIXED_COUNTERS) { > - num_architectural_pmu_fixed_counters = MAX_FIXED_COUNTERS; > + if (num_pmu_fixed_counters > MAX_FIXED_COUNTERS) { > + num_pmu_fixed_counters = MAX_FIXED_COUNTERS; > } > } > } > @@ -4041,25 +4048,25 @@ static int kvm_put_msrs(X86CPU *cpu, int level) > kvm_msr_entry_add(cpu, MSR_KVM_POLL_CONTROL, env->poll_control_msr); > } > > - if (has_architectural_pmu_version > 0) { > - if (has_architectural_pmu_version > 1) { > + if (has_pmu_version > 0) { > + if (has_pmu_version > 1) { > /* Stop the counter. */ > kvm_msr_entry_add(cpu, MSR_CORE_PERF_FIXED_CTR_CTRL, 0); > kvm_msr_entry_add(cpu, MSR_CORE_PERF_GLOBAL_CTRL, 0); > } > > /* Set the counter values. */ > - for (i = 0; i < num_architectural_pmu_fixed_counters; i++) { > + for (i = 0; i < num_pmu_fixed_counters; i++) { > kvm_msr_entry_add(cpu, MSR_CORE_PERF_FIXED_CTR0 + i, > env->msr_fixed_counters[i]); > } > - for (i = 0; i < num_architectural_pmu_gp_counters; i++) { > + for (i = 0; i < num_pmu_gp_counters; i++) { > kvm_msr_entry_add(cpu, MSR_P6_PERFCTR0 + i, > env->msr_gp_counters[i]); > kvm_msr_entry_add(cpu, MSR_P6_EVNTSEL0 + i, > env->msr_gp_evtsel[i]); > } > - if (has_architectural_pmu_version > 1) { > + if (has_pmu_version > 1) { > kvm_msr_entry_add(cpu, MSR_CORE_PERF_GLOBAL_STATUS, > env->msr_global_status); > kvm_msr_entry_add(cpu, MSR_CORE_PERF_GLOBAL_OVF_CTRL, > @@ -4519,17 +4526,17 @@ static int kvm_get_msrs(X86CPU *cpu) > if (env->features[FEAT_KVM] & CPUID_KVM_POLL_CONTROL) { > kvm_msr_entry_add(cpu, MSR_KVM_POLL_CONTROL, 1); > } > - if (has_architectural_pmu_version > 0) { > - if (has_architectural_pmu_version > 1) { > + if (has_pmu_version > 0) { > + if (has_pmu_version > 1) { > kvm_msr_entry_add(cpu, MSR_CORE_PERF_FIXED_CTR_CTRL, 0); > kvm_msr_entry_add(cpu, MSR_CORE_PERF_GLOBAL_CTRL, 0); > kvm_msr_entry_add(cpu, MSR_CORE_PERF_GLOBAL_STATUS, 0); > kvm_msr_entry_add(cpu, MSR_CORE_PERF_GLOBAL_OVF_CTRL, 0); > } > - for (i = 0; i < num_architectural_pmu_fixed_counters; i++) { > + for (i = 0; i < num_pmu_fixed_counters; i++) { > kvm_msr_entry_add(cpu, MSR_CORE_PERF_FIXED_CTR0 + i, 0); > } > - for (i = 0; i < num_architectural_pmu_gp_counters; i++) { > + for (i = 0; i < num_pmu_gp_counters; i++) { > kvm_msr_entry_add(cpu, MSR_P6_PERFCTR0 + i, 0); > kvm_msr_entry_add(cpu, MSR_P6_EVNTSEL0 + i, 0); > } Reviewed-by: Dapeng Mi <dapeng1.mi@linux.intel.com>
> +/* > + * For Intel processors, the meaning is the architectural PMU version > + * number. > + * > + * For AMD processors: 1 corresponds to the prior versions, and 2 > + * corresponds to AMD PerfMonV2. > + */ > +static uint32_t has_pmu_version; The "has_" prefix sounds like a boolean type. So what about "pmu_version"? Others look good to me, Reviewed-by: Zhao Liu <zhao1.liu@intel.com>
On 3/7/25 1:19 AM, Zhao Liu wrote: >> +/* >> + * For Intel processors, the meaning is the architectural PMU version >> + * number. >> + * >> + * For AMD processors: 1 corresponds to the prior versions, and 2 >> + * corresponds to AMD PerfMonV2. >> + */ >> +static uint32_t has_pmu_version; > > The "has_" prefix sounds like a boolean type. So what about "pmu_version"? Sure. I will change to pmu_version. Thank you very much! Dongli Zhang > > Others look good to me, > > Reviewed-by: Zhao Liu <zhao1.liu@intel.com> > >
diff --git a/target/i386/kvm/kvm.c b/target/i386/kvm/kvm.c index 8f293ffd61..e895d22f94 100644 --- a/target/i386/kvm/kvm.c +++ b/target/i386/kvm/kvm.c @@ -164,9 +164,16 @@ static bool has_msr_perf_capabs; static bool has_msr_pkrs; static bool has_msr_hwcr; -static uint32_t has_architectural_pmu_version; -static uint32_t num_architectural_pmu_gp_counters; -static uint32_t num_architectural_pmu_fixed_counters; +/* + * For Intel processors, the meaning is the architectural PMU version + * number. + * + * For AMD processors: 1 corresponds to the prior versions, and 2 + * corresponds to AMD PerfMonV2. + */ +static uint32_t has_pmu_version; +static uint32_t num_pmu_gp_counters; +static uint32_t num_pmu_fixed_counters; static int has_xsave2; static int has_xcrs; @@ -2072,24 +2079,24 @@ static void kvm_init_pmu_info(CPUX86State *env) cpu_x86_cpuid(env, 0x0a, 0, &eax, &unused, &unused, &edx); - has_architectural_pmu_version = eax & 0xff; - if (has_architectural_pmu_version > 0) { - num_architectural_pmu_gp_counters = (eax & 0xff00) >> 8; + has_pmu_version = eax & 0xff; + if (has_pmu_version > 0) { + num_pmu_gp_counters = (eax & 0xff00) >> 8; /* * Shouldn't be more than 32, since that's the number of bits * available in EBX to tell us _which_ counters are available. * Play it safe. */ - if (num_architectural_pmu_gp_counters > MAX_GP_COUNTERS) { - num_architectural_pmu_gp_counters = MAX_GP_COUNTERS; + if (num_pmu_gp_counters > MAX_GP_COUNTERS) { + num_pmu_gp_counters = MAX_GP_COUNTERS; } - if (has_architectural_pmu_version > 1) { - num_architectural_pmu_fixed_counters = edx & 0x1f; + if (has_pmu_version > 1) { + num_pmu_fixed_counters = edx & 0x1f; - if (num_architectural_pmu_fixed_counters > MAX_FIXED_COUNTERS) { - num_architectural_pmu_fixed_counters = MAX_FIXED_COUNTERS; + if (num_pmu_fixed_counters > MAX_FIXED_COUNTERS) { + num_pmu_fixed_counters = MAX_FIXED_COUNTERS; } } } @@ -4041,25 +4048,25 @@ static int kvm_put_msrs(X86CPU *cpu, int level) kvm_msr_entry_add(cpu, MSR_KVM_POLL_CONTROL, env->poll_control_msr); } - if (has_architectural_pmu_version > 0) { - if (has_architectural_pmu_version > 1) { + if (has_pmu_version > 0) { + if (has_pmu_version > 1) { /* Stop the counter. */ kvm_msr_entry_add(cpu, MSR_CORE_PERF_FIXED_CTR_CTRL, 0); kvm_msr_entry_add(cpu, MSR_CORE_PERF_GLOBAL_CTRL, 0); } /* Set the counter values. */ - for (i = 0; i < num_architectural_pmu_fixed_counters; i++) { + for (i = 0; i < num_pmu_fixed_counters; i++) { kvm_msr_entry_add(cpu, MSR_CORE_PERF_FIXED_CTR0 + i, env->msr_fixed_counters[i]); } - for (i = 0; i < num_architectural_pmu_gp_counters; i++) { + for (i = 0; i < num_pmu_gp_counters; i++) { kvm_msr_entry_add(cpu, MSR_P6_PERFCTR0 + i, env->msr_gp_counters[i]); kvm_msr_entry_add(cpu, MSR_P6_EVNTSEL0 + i, env->msr_gp_evtsel[i]); } - if (has_architectural_pmu_version > 1) { + if (has_pmu_version > 1) { kvm_msr_entry_add(cpu, MSR_CORE_PERF_GLOBAL_STATUS, env->msr_global_status); kvm_msr_entry_add(cpu, MSR_CORE_PERF_GLOBAL_OVF_CTRL, @@ -4519,17 +4526,17 @@ static int kvm_get_msrs(X86CPU *cpu) if (env->features[FEAT_KVM] & CPUID_KVM_POLL_CONTROL) { kvm_msr_entry_add(cpu, MSR_KVM_POLL_CONTROL, 1); } - if (has_architectural_pmu_version > 0) { - if (has_architectural_pmu_version > 1) { + if (has_pmu_version > 0) { + if (has_pmu_version > 1) { kvm_msr_entry_add(cpu, MSR_CORE_PERF_FIXED_CTR_CTRL, 0); kvm_msr_entry_add(cpu, MSR_CORE_PERF_GLOBAL_CTRL, 0); kvm_msr_entry_add(cpu, MSR_CORE_PERF_GLOBAL_STATUS, 0); kvm_msr_entry_add(cpu, MSR_CORE_PERF_GLOBAL_OVF_CTRL, 0); } - for (i = 0; i < num_architectural_pmu_fixed_counters; i++) { + for (i = 0; i < num_pmu_fixed_counters; i++) { kvm_msr_entry_add(cpu, MSR_CORE_PERF_FIXED_CTR0 + i, 0); } - for (i = 0; i < num_architectural_pmu_gp_counters; i++) { + for (i = 0; i < num_pmu_gp_counters; i++) { kvm_msr_entry_add(cpu, MSR_P6_PERFCTR0 + i, 0); kvm_msr_entry_add(cpu, MSR_P6_EVNTSEL0 + i, 0); }
AMD does not have what is commonly referred to as an architectural PMU. Therefore, we need to rename the following variables to be applicable for both Intel and AMD: - has_architectural_pmu_version - num_architectural_pmu_gp_counters - num_architectural_pmu_fixed_counters For Intel processors, the meaning of has_pmu_version remains unchanged. For AMD processors: has_pmu_version == 1 corresponds to versions before AMD PerfMonV2. has_pmu_version == 2 corresponds to AMD PerfMonV2. Signed-off-by: Dongli Zhang <dongli.zhang@oracle.com> --- target/i386/kvm/kvm.c | 49 ++++++++++++++++++++++++------------------- 1 file changed, 28 insertions(+), 21 deletions(-)