Message ID | 1468220912-22828-5-git-send-email-sjitindarsingh@gmail.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On Mon, Jul 11, 2016 at 12:08 AM, Suraj Jitindar Singh <sjitindarsingh@gmail.com> wrote: > vcpu stats are used to collect information about a vcpu which can be viewed > in the debugfs. For example halt_attempted_poll and halt_successful_poll > are used to keep track of the number of times the vcpu attempts to and > successfully polls. These stats are currently not used on powerpc. > > Implement incrementation of the halt_attempted_poll and > halt_successful_poll vcpu stats for powerpc. Since these stats are summed > over all the vcpus for all running guests it doesn't matter which vcpu > they are attributed to, thus we choose the current runner vcpu of the > vcore. > > Also add new vcpu stats: halt_poll_time and halt_wait_time to be used to > accumulate the total time spend polling and waiting respectively, and > halt_successful_wait to accumulate the number of times the vcpu waits. > Given that halt_poll_time and halt_wait_time are expressed in nanoseconds > it is necessary to represent these as 64-bit quantities, otherwise they > would overflow after only about 4 seconds. > > Given that the total time spend either polling or waiting will be known and > the number of times that each was done, it will be possible to determine > the average poll and wait times. This will give the ability to tune the kvm > module parameters based on the calculated average wait and poll times. > > --- > Change Log: > > V1 -> V2: > - Nothing > > Signed-off-by: Suraj Jitindar Singh <sjitindarsingh@gmail.com> > --- > arch/powerpc/include/asm/kvm_host.h | 3 +++ > arch/powerpc/kvm/book3s.c | 3 +++ > arch/powerpc/kvm/book3s_hv.c | 14 +++++++++++++- > 3 files changed, 19 insertions(+), 1 deletion(-) > > diff --git a/arch/powerpc/include/asm/kvm_host.h b/arch/powerpc/include/asm/kvm_host.h > index 610f393..66a7198 100644 > --- a/arch/powerpc/include/asm/kvm_host.h > +++ b/arch/powerpc/include/asm/kvm_host.h > @@ -114,8 +114,11 @@ struct kvm_vcpu_stat { > u32 emulated_inst_exits; > u32 dec_exits; > u32 ext_intr_exits; > + u64 halt_poll_time; > + u64 halt_wait_time; > u32 halt_successful_poll; > u32 halt_attempted_poll; > + u32 halt_successful_wait; > u32 halt_poll_invalid; > u32 halt_wakeup; > u32 dbell_exits; > diff --git a/arch/powerpc/kvm/book3s.c b/arch/powerpc/kvm/book3s.c > index ed9132b..6217bea 100644 > --- a/arch/powerpc/kvm/book3s.c > +++ b/arch/powerpc/kvm/book3s.c > @@ -53,8 +53,11 @@ struct kvm_stats_debugfs_item debugfs_entries[] = { > { "dec", VCPU_STAT(dec_exits) }, > { "ext_intr", VCPU_STAT(ext_intr_exits) }, > { "queue_intr", VCPU_STAT(queue_intr) }, > + { "halt_poll_time_ns", VCPU_STAT_U64(halt_poll_time) }, > + { "halt_wait_time_ns", VCPU_STAT_U64(halt_wait_time) }, > { "halt_successful_poll", VCPU_STAT(halt_successful_poll), }, > { "halt_attempted_poll", VCPU_STAT(halt_attempted_poll), }, > + { "halt_successful_wait", VCPU_STAT(halt_successful_wait) }, > { "halt_poll_invalid", VCPU_STAT(halt_poll_invalid) }, > { "halt_wakeup", VCPU_STAT(halt_wakeup) }, > { "pf_storage", VCPU_STAT(pf_storage) }, > diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c > index 0d8ce14..a0dae63 100644 > --- a/arch/powerpc/kvm/book3s_hv.c > +++ b/arch/powerpc/kvm/book3s_hv.c > @@ -2688,6 +2688,7 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) > cur = start = ktime_get(); > if (vc->halt_poll_ns) { > ktime_t stop = ktime_add_ns(start, vc->halt_poll_ns); > + ++vc->runner->stat.halt_attempted_poll; > > vc->vcore_state = VCORE_POLLING; > spin_unlock(&vc->lock); > @@ -2703,8 +2704,10 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) > spin_lock(&vc->lock); > vc->vcore_state = VCORE_INACTIVE; > > - if (!do_sleep) > + if (!do_sleep) { > + ++vc->runner->stat.halt_successful_poll; > goto out; > + } > } > > prepare_to_swait(&vc->wq, &wait, TASK_INTERRUPTIBLE); > @@ -2712,6 +2715,9 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) > if (kvmppc_vcore_check_block(vc)) { > finish_swait(&vc->wq, &wait); > do_sleep = 0; > + /* If we polled, count this as a successful poll */ > + if (vc->halt_poll_ns) > + ++vc->runner->stat.halt_successful_poll; > goto out; > } > > @@ -2723,12 +2729,18 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) > spin_lock(&vc->lock); > vc->vcore_state = VCORE_INACTIVE; > trace_kvmppc_vcore_blocked(vc, 1); > + ++vc->runner->stat.halt_successful_wait; > > cur = ktime_get(); > > out: > block_ns = ktime_to_ns(cur) - ktime_to_ns(start); > > + if (do_sleep) > + vc->runner->stat.halt_wait_time += block_ns; It's possible to poll and wait in one halt, conflating this stat with polling time. Is it useful to split out a third stat, halt_poll_fail_ns which counts how long we polled which ended up sleeping? Then halt_wait_time only counts the time the VCPU spent on the wait queue. The sum of all 3 is still the total time spent halted. > + else if (vc->halt_poll_ns) > + vc->runner->stat.halt_poll_time += block_ns; > + > if (halt_poll_max_ns) { > if (block_ns <= vc->halt_poll_ns) > ; > -- > 2.5.5 > > -- > To unsubscribe from this list: send the line "unsubscribe kvm" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On 12/07/16 02:49, David Matlack wrote: > On Mon, Jul 11, 2016 at 12:08 AM, Suraj Jitindar Singh > <sjitindarsingh@gmail.com> wrote: >> vcpu stats are used to collect information about a vcpu which can be viewed >> in the debugfs. For example halt_attempted_poll and halt_successful_poll >> are used to keep track of the number of times the vcpu attempts to and >> successfully polls. These stats are currently not used on powerpc. >> >> Implement incrementation of the halt_attempted_poll and >> halt_successful_poll vcpu stats for powerpc. Since these stats are summed >> over all the vcpus for all running guests it doesn't matter which vcpu >> they are attributed to, thus we choose the current runner vcpu of the >> vcore. >> >> Also add new vcpu stats: halt_poll_time and halt_wait_time to be used to >> accumulate the total time spend polling and waiting respectively, and >> halt_successful_wait to accumulate the number of times the vcpu waits. >> Given that halt_poll_time and halt_wait_time are expressed in nanoseconds >> it is necessary to represent these as 64-bit quantities, otherwise they >> would overflow after only about 4 seconds. >> >> Given that the total time spend either polling or waiting will be known and >> the number of times that each was done, it will be possible to determine >> the average poll and wait times. This will give the ability to tune the kvm >> module parameters based on the calculated average wait and poll times. >> >> --- >> Change Log: >> >> V1 -> V2: >> - Nothing >> >> Signed-off-by: Suraj Jitindar Singh <sjitindarsingh@gmail.com> >> --- >> arch/powerpc/include/asm/kvm_host.h | 3 +++ >> arch/powerpc/kvm/book3s.c | 3 +++ >> arch/powerpc/kvm/book3s_hv.c | 14 +++++++++++++- >> 3 files changed, 19 insertions(+), 1 deletion(-) >> >> diff --git a/arch/powerpc/include/asm/kvm_host.h b/arch/powerpc/include/asm/kvm_host.h >> index 610f393..66a7198 100644 >> --- a/arch/powerpc/include/asm/kvm_host.h >> +++ b/arch/powerpc/include/asm/kvm_host.h >> @@ -114,8 +114,11 @@ struct kvm_vcpu_stat { >> u32 emulated_inst_exits; >> u32 dec_exits; >> u32 ext_intr_exits; >> + u64 halt_poll_time; >> + u64 halt_wait_time; >> u32 halt_successful_poll; >> u32 halt_attempted_poll; >> + u32 halt_successful_wait; >> u32 halt_poll_invalid; >> u32 halt_wakeup; >> u32 dbell_exits; >> diff --git a/arch/powerpc/kvm/book3s.c b/arch/powerpc/kvm/book3s.c >> index ed9132b..6217bea 100644 >> --- a/arch/powerpc/kvm/book3s.c >> +++ b/arch/powerpc/kvm/book3s.c >> @@ -53,8 +53,11 @@ struct kvm_stats_debugfs_item debugfs_entries[] = { >> { "dec", VCPU_STAT(dec_exits) }, >> { "ext_intr", VCPU_STAT(ext_intr_exits) }, >> { "queue_intr", VCPU_STAT(queue_intr) }, >> + { "halt_poll_time_ns", VCPU_STAT_U64(halt_poll_time) }, >> + { "halt_wait_time_ns", VCPU_STAT_U64(halt_wait_time) }, >> { "halt_successful_poll", VCPU_STAT(halt_successful_poll), }, >> { "halt_attempted_poll", VCPU_STAT(halt_attempted_poll), }, >> + { "halt_successful_wait", VCPU_STAT(halt_successful_wait) }, >> { "halt_poll_invalid", VCPU_STAT(halt_poll_invalid) }, >> { "halt_wakeup", VCPU_STAT(halt_wakeup) }, >> { "pf_storage", VCPU_STAT(pf_storage) }, >> diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c >> index 0d8ce14..a0dae63 100644 >> --- a/arch/powerpc/kvm/book3s_hv.c >> +++ b/arch/powerpc/kvm/book3s_hv.c >> @@ -2688,6 +2688,7 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) >> cur = start = ktime_get(); >> if (vc->halt_poll_ns) { >> ktime_t stop = ktime_add_ns(start, vc->halt_poll_ns); >> + ++vc->runner->stat.halt_attempted_poll; >> >> vc->vcore_state = VCORE_POLLING; >> spin_unlock(&vc->lock); >> @@ -2703,8 +2704,10 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) >> spin_lock(&vc->lock); >> vc->vcore_state = VCORE_INACTIVE; >> >> - if (!do_sleep) >> + if (!do_sleep) { >> + ++vc->runner->stat.halt_successful_poll; >> goto out; >> + } >> } >> >> prepare_to_swait(&vc->wq, &wait, TASK_INTERRUPTIBLE); >> @@ -2712,6 +2715,9 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) >> if (kvmppc_vcore_check_block(vc)) { >> finish_swait(&vc->wq, &wait); >> do_sleep = 0; >> + /* If we polled, count this as a successful poll */ >> + if (vc->halt_poll_ns) >> + ++vc->runner->stat.halt_successful_poll; >> goto out; >> } >> >> @@ -2723,12 +2729,18 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) >> spin_lock(&vc->lock); >> vc->vcore_state = VCORE_INACTIVE; >> trace_kvmppc_vcore_blocked(vc, 1); >> + ++vc->runner->stat.halt_successful_wait; >> >> cur = ktime_get(); >> >> out: >> block_ns = ktime_to_ns(cur) - ktime_to_ns(start); >> >> + if (do_sleep) >> + vc->runner->stat.halt_wait_time += block_ns; > It's possible to poll and wait in one halt, conflating this stat with > polling time. Is it useful to split out a third stat, > halt_poll_fail_ns which counts how long we polled which ended up > sleeping? Then halt_wait_time only counts the time the VCPU spent on > the wait queue. The sum of all 3 is still the total time spent halted. > I see what you're saying. I would say that in the event that you do wait then the most useful number is going to be the total block time (the sum of the wait and poll time) as this is the minimum value you would have to set the halt_poll_max_ns module parameter in order to ensure you poll for long enough (in most circumstances) to avoid waiting, which is the main use case I envision for this statistic. That being said this is definitely a source of ambiguity and splitting this into two statistics would make the distinction clearer without any loss of data, you could simply sum the two stats to get the same number. Either way I don't think it really makes much of a difference, but in the interest of clarity I think I'll split the statistic. >> + else if (vc->halt_poll_ns) >> + vc->runner->stat.halt_poll_time += block_ns; >> + >> if (halt_poll_max_ns) { >> if (block_ns <= vc->halt_poll_ns) >> ; >> -- >> 2.5.5 >> >> -- >> To unsubscribe from this list: send the line "unsubscribe kvm" in >> the body of a message to majordomo@vger.kernel.org >> More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On 12/07/16 16:17, Suraj Jitindar Singh wrote: > On 12/07/16 02:49, David Matlack wrote: >> On Mon, Jul 11, 2016 at 12:08 AM, Suraj Jitindar Singh >> <sjitindarsingh@gmail.com> wrote: >>> vcpu stats are used to collect information about a vcpu which can be viewed >>> in the debugfs. For example halt_attempted_poll and halt_successful_poll >>> are used to keep track of the number of times the vcpu attempts to and >>> successfully polls. These stats are currently not used on powerpc. >>> >>> Implement incrementation of the halt_attempted_poll and >>> halt_successful_poll vcpu stats for powerpc. Since these stats are summed >>> over all the vcpus for all running guests it doesn't matter which vcpu >>> they are attributed to, thus we choose the current runner vcpu of the >>> vcore. >>> >>> Also add new vcpu stats: halt_poll_time and halt_wait_time to be used to >>> accumulate the total time spend polling and waiting respectively, and >>> halt_successful_wait to accumulate the number of times the vcpu waits. >>> Given that halt_poll_time and halt_wait_time are expressed in nanoseconds >>> it is necessary to represent these as 64-bit quantities, otherwise they >>> would overflow after only about 4 seconds. >>> >>> Given that the total time spend either polling or waiting will be known and >>> the number of times that each was done, it will be possible to determine >>> the average poll and wait times. This will give the ability to tune the kvm >>> module parameters based on the calculated average wait and poll times. >>> >>> --- >>> Change Log: >>> >>> V1 -> V2: >>> - Nothing >>> >>> Signed-off-by: Suraj Jitindar Singh <sjitindarsingh@gmail.com> >>> --- >>> arch/powerpc/include/asm/kvm_host.h | 3 +++ >>> arch/powerpc/kvm/book3s.c | 3 +++ >>> arch/powerpc/kvm/book3s_hv.c | 14 +++++++++++++- >>> 3 files changed, 19 insertions(+), 1 deletion(-) >>> >>> diff --git a/arch/powerpc/include/asm/kvm_host.h b/arch/powerpc/include/asm/kvm_host.h >>> index 610f393..66a7198 100644 >>> --- a/arch/powerpc/include/asm/kvm_host.h >>> +++ b/arch/powerpc/include/asm/kvm_host.h >>> @@ -114,8 +114,11 @@ struct kvm_vcpu_stat { >>> u32 emulated_inst_exits; >>> u32 dec_exits; >>> u32 ext_intr_exits; >>> + u64 halt_poll_time; >>> + u64 halt_wait_time; >>> u32 halt_successful_poll; >>> u32 halt_attempted_poll; >>> + u32 halt_successful_wait; >>> u32 halt_poll_invalid; >>> u32 halt_wakeup; >>> u32 dbell_exits; >>> diff --git a/arch/powerpc/kvm/book3s.c b/arch/powerpc/kvm/book3s.c >>> index ed9132b..6217bea 100644 >>> --- a/arch/powerpc/kvm/book3s.c >>> +++ b/arch/powerpc/kvm/book3s.c >>> @@ -53,8 +53,11 @@ struct kvm_stats_debugfs_item debugfs_entries[] = { >>> { "dec", VCPU_STAT(dec_exits) }, >>> { "ext_intr", VCPU_STAT(ext_intr_exits) }, >>> { "queue_intr", VCPU_STAT(queue_intr) }, >>> + { "halt_poll_time_ns", VCPU_STAT_U64(halt_poll_time) }, >>> + { "halt_wait_time_ns", VCPU_STAT_U64(halt_wait_time) }, >>> { "halt_successful_poll", VCPU_STAT(halt_successful_poll), }, >>> { "halt_attempted_poll", VCPU_STAT(halt_attempted_poll), }, >>> + { "halt_successful_wait", VCPU_STAT(halt_successful_wait) }, >>> { "halt_poll_invalid", VCPU_STAT(halt_poll_invalid) }, >>> { "halt_wakeup", VCPU_STAT(halt_wakeup) }, >>> { "pf_storage", VCPU_STAT(pf_storage) }, >>> diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c >>> index 0d8ce14..a0dae63 100644 >>> --- a/arch/powerpc/kvm/book3s_hv.c >>> +++ b/arch/powerpc/kvm/book3s_hv.c >>> @@ -2688,6 +2688,7 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) >>> cur = start = ktime_get(); >>> if (vc->halt_poll_ns) { >>> ktime_t stop = ktime_add_ns(start, vc->halt_poll_ns); >>> + ++vc->runner->stat.halt_attempted_poll; >>> >>> vc->vcore_state = VCORE_POLLING; >>> spin_unlock(&vc->lock); >>> @@ -2703,8 +2704,10 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) >>> spin_lock(&vc->lock); >>> vc->vcore_state = VCORE_INACTIVE; >>> >>> - if (!do_sleep) >>> + if (!do_sleep) { >>> + ++vc->runner->stat.halt_successful_poll; >>> goto out; >>> + } >>> } >>> >>> prepare_to_swait(&vc->wq, &wait, TASK_INTERRUPTIBLE); >>> @@ -2712,6 +2715,9 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) >>> if (kvmppc_vcore_check_block(vc)) { >>> finish_swait(&vc->wq, &wait); >>> do_sleep = 0; >>> + /* If we polled, count this as a successful poll */ >>> + if (vc->halt_poll_ns) >>> + ++vc->runner->stat.halt_successful_poll; >>> goto out; >>> } >>> >>> @@ -2723,12 +2729,18 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) >>> spin_lock(&vc->lock); >>> vc->vcore_state = VCORE_INACTIVE; >>> trace_kvmppc_vcore_blocked(vc, 1); >>> + ++vc->runner->stat.halt_successful_wait; >>> >>> cur = ktime_get(); >>> >>> out: >>> block_ns = ktime_to_ns(cur) - ktime_to_ns(start); >>> >>> + if (do_sleep) >>> + vc->runner->stat.halt_wait_time += block_ns; >> It's possible to poll and wait in one halt, conflating this stat with >> polling time. Is it useful to split out a third stat, >> halt_poll_fail_ns which counts how long we polled which ended up >> sleeping? Then halt_wait_time only counts the time the VCPU spent on >> the wait queue. The sum of all 3 is still the total time spent halted. >> > I see what you're saying. I would say that in the event that you do wait > then the most useful number is going to be the total block time (the sum > of the wait and poll time) as this is the minimum value you would have to > set the halt_poll_max_ns module parameter in order to ensure you poll > for long enough (in most circumstances) to avoid waiting, which is the main > use case I envision for this statistic. That being said this is definitely > a source of ambiguity and splitting this into two statistics would make the > distinction clearer without any loss of data, you could simply sum the two > stats to get the same number. > > Either way I don't think it really makes much of a difference, but in the > interest of clarity I think I'll split the statistic. On further though, I really think that splitting this statistic is an unnecessary source of ambiguity. In reality the interesting piece of information is going to be the average time that you blocked on either an unsuccessful poll or a successful poll. So instead of splitting the statistic I'm going to rename them as: halt_poll_time -> halt_block_time_successful_poll halt_wait_time -> halt_block_time_waited > >>> + else if (vc->halt_poll_ns) >>> + vc->runner->stat.halt_poll_time += block_ns; >>> + >>> if (halt_poll_max_ns) { >>> if (block_ns <= vc->halt_poll_ns) >>> ; >>> -- >>> 2.5.5 >>> >>> -- >>> To unsubscribe from this list: send the line "unsubscribe kvm" in >>> the body of a message to majordomo@vger.kernel.org >>> More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On Tue, Jul 12, 2016 at 11:07 PM, Suraj Jitindar Singh <sjitindarsingh@gmail.com> wrote: > On 12/07/16 16:17, Suraj Jitindar Singh wrote: >> On 12/07/16 02:49, David Matlack wrote: [snip] >>> It's possible to poll and wait in one halt, conflating this stat with >>> polling time. Is it useful to split out a third stat, >>> halt_poll_fail_ns which counts how long we polled which ended up >>> sleeping? Then halt_wait_time only counts the time the VCPU spent on >>> the wait queue. The sum of all 3 is still the total time spent halted. >>> >> I see what you're saying. I would say that in the event that you do wait >> then the most useful number is going to be the total block time (the sum >> of the wait and poll time) as this is the minimum value you would have to >> set the halt_poll_max_ns module parameter in order to ensure you poll >> for long enough (in most circumstances) to avoid waiting, which is the main >> use case I envision for this statistic. That being said this is definitely >> a source of ambiguity and splitting this into two statistics would make the >> distinction clearer without any loss of data, you could simply sum the two >> stats to get the same number. >> >> Either way I don't think it really makes much of a difference, but in the >> interest of clarity I think I'll split the statistic. > > On further though, I really think that splitting this statistic is an > unnecessary source of ambiguity. In reality the interesting piece of > information is going to be the average time that you blocked on > either an unsuccessful poll or a successful poll. > > So instead of splitting the statistic I'm going to rename them as: > halt_poll_time -> halt_block_time_successful_poll > halt_wait_time -> halt_block_time_waited The downside of having only these 2 stats is there is no way to see the total time spent halt-polling. Halt-polling shows up as host kernel CPU usage on the VCPU thread, despite it really being idle cycles that could be reclaimed. It's useful to have the total amount of time spent halt-polling (halt_poll_fail + halt_poll_success) to feed into provisioning/monitoring systems that look at CPU usage. FWIW, I have a very similar patch internally. It adds 2 stats, halt_poll_success_ns and halt_poll_fail_ns, to the halt-polling code in virt/kvm/kvm_main.c. So if you agree splitting the stats makes sense, it would be helpful to us if we can adopt the same naming convention. -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On 14/07/16 03:20, David Matlack wrote: > On Tue, Jul 12, 2016 at 11:07 PM, Suraj Jitindar Singh > <sjitindarsingh@gmail.com> wrote: >> On 12/07/16 16:17, Suraj Jitindar Singh wrote: >>> On 12/07/16 02:49, David Matlack wrote: > [snip] >>>> It's possible to poll and wait in one halt, conflating this stat with >>>> polling time. Is it useful to split out a third stat, >>>> halt_poll_fail_ns which counts how long we polled which ended up >>>> sleeping? Then halt_wait_time only counts the time the VCPU spent on >>>> the wait queue. The sum of all 3 is still the total time spent halted. >>>> >>> I see what you're saying. I would say that in the event that you do wait >>> then the most useful number is going to be the total block time (the sum >>> of the wait and poll time) as this is the minimum value you would have to >>> set the halt_poll_max_ns module parameter in order to ensure you poll >>> for long enough (in most circumstances) to avoid waiting, which is the main >>> use case I envision for this statistic. That being said this is definitely >>> a source of ambiguity and splitting this into two statistics would make the >>> distinction clearer without any loss of data, you could simply sum the two >>> stats to get the same number. >>> >>> Either way I don't think it really makes much of a difference, but in the >>> interest of clarity I think I'll split the statistic. >> On further though, I really think that splitting this statistic is an >> unnecessary source of ambiguity. In reality the interesting piece of >> information is going to be the average time that you blocked on >> either an unsuccessful poll or a successful poll. >> >> So instead of splitting the statistic I'm going to rename them as: >> halt_poll_time -> halt_block_time_successful_poll >> halt_wait_time -> halt_block_time_waited > The downside of having only these 2 stats is there is no way to see > the total time spent halt-polling. Halt-polling shows up as host > kernel CPU usage on the VCPU thread, despite it really being idle > cycles that could be reclaimed. It's useful to have the total amount > of time spent halt-polling (halt_poll_fail + halt_poll_success) to > feed into provisioning/monitoring systems that look at CPU usage. > > FWIW, I have a very similar patch internally. It adds 2 stats, > halt_poll_success_ns and halt_poll_fail_ns, to the halt-polling code > in virt/kvm/kvm_main.c. So if you agree splitting the stats makes > sense, it would be helpful to us if we can adopt the same naming > convention. Ok, I didn't realise that was a use case. Makes sense, I'll split it and adopt those names. Thanks -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
diff --git a/arch/powerpc/include/asm/kvm_host.h b/arch/powerpc/include/asm/kvm_host.h index 610f393..66a7198 100644 --- a/arch/powerpc/include/asm/kvm_host.h +++ b/arch/powerpc/include/asm/kvm_host.h @@ -114,8 +114,11 @@ struct kvm_vcpu_stat { u32 emulated_inst_exits; u32 dec_exits; u32 ext_intr_exits; + u64 halt_poll_time; + u64 halt_wait_time; u32 halt_successful_poll; u32 halt_attempted_poll; + u32 halt_successful_wait; u32 halt_poll_invalid; u32 halt_wakeup; u32 dbell_exits; diff --git a/arch/powerpc/kvm/book3s.c b/arch/powerpc/kvm/book3s.c index ed9132b..6217bea 100644 --- a/arch/powerpc/kvm/book3s.c +++ b/arch/powerpc/kvm/book3s.c @@ -53,8 +53,11 @@ struct kvm_stats_debugfs_item debugfs_entries[] = { { "dec", VCPU_STAT(dec_exits) }, { "ext_intr", VCPU_STAT(ext_intr_exits) }, { "queue_intr", VCPU_STAT(queue_intr) }, + { "halt_poll_time_ns", VCPU_STAT_U64(halt_poll_time) }, + { "halt_wait_time_ns", VCPU_STAT_U64(halt_wait_time) }, { "halt_successful_poll", VCPU_STAT(halt_successful_poll), }, { "halt_attempted_poll", VCPU_STAT(halt_attempted_poll), }, + { "halt_successful_wait", VCPU_STAT(halt_successful_wait) }, { "halt_poll_invalid", VCPU_STAT(halt_poll_invalid) }, { "halt_wakeup", VCPU_STAT(halt_wakeup) }, { "pf_storage", VCPU_STAT(pf_storage) }, diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c index 0d8ce14..a0dae63 100644 --- a/arch/powerpc/kvm/book3s_hv.c +++ b/arch/powerpc/kvm/book3s_hv.c @@ -2688,6 +2688,7 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) cur = start = ktime_get(); if (vc->halt_poll_ns) { ktime_t stop = ktime_add_ns(start, vc->halt_poll_ns); + ++vc->runner->stat.halt_attempted_poll; vc->vcore_state = VCORE_POLLING; spin_unlock(&vc->lock); @@ -2703,8 +2704,10 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) spin_lock(&vc->lock); vc->vcore_state = VCORE_INACTIVE; - if (!do_sleep) + if (!do_sleep) { + ++vc->runner->stat.halt_successful_poll; goto out; + } } prepare_to_swait(&vc->wq, &wait, TASK_INTERRUPTIBLE); @@ -2712,6 +2715,9 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) if (kvmppc_vcore_check_block(vc)) { finish_swait(&vc->wq, &wait); do_sleep = 0; + /* If we polled, count this as a successful poll */ + if (vc->halt_poll_ns) + ++vc->runner->stat.halt_successful_poll; goto out; } @@ -2723,12 +2729,18 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc) spin_lock(&vc->lock); vc->vcore_state = VCORE_INACTIVE; trace_kvmppc_vcore_blocked(vc, 1); + ++vc->runner->stat.halt_successful_wait; cur = ktime_get(); out: block_ns = ktime_to_ns(cur) - ktime_to_ns(start); + if (do_sleep) + vc->runner->stat.halt_wait_time += block_ns; + else if (vc->halt_poll_ns) + vc->runner->stat.halt_poll_time += block_ns; + if (halt_poll_max_ns) { if (block_ns <= vc->halt_poll_ns) ;
vcpu stats are used to collect information about a vcpu which can be viewed in the debugfs. For example halt_attempted_poll and halt_successful_poll are used to keep track of the number of times the vcpu attempts to and successfully polls. These stats are currently not used on powerpc. Implement incrementation of the halt_attempted_poll and halt_successful_poll vcpu stats for powerpc. Since these stats are summed over all the vcpus for all running guests it doesn't matter which vcpu they are attributed to, thus we choose the current runner vcpu of the vcore. Also add new vcpu stats: halt_poll_time and halt_wait_time to be used to accumulate the total time spend polling and waiting respectively, and halt_successful_wait to accumulate the number of times the vcpu waits. Given that halt_poll_time and halt_wait_time are expressed in nanoseconds it is necessary to represent these as 64-bit quantities, otherwise they would overflow after only about 4 seconds. Given that the total time spend either polling or waiting will be known and the number of times that each was done, it will be possible to determine the average poll and wait times. This will give the ability to tune the kvm module parameters based on the calculated average wait and poll times. --- Change Log: V1 -> V2: - Nothing Signed-off-by: Suraj Jitindar Singh <sjitindarsingh@gmail.com> --- arch/powerpc/include/asm/kvm_host.h | 3 +++ arch/powerpc/kvm/book3s.c | 3 +++ arch/powerpc/kvm/book3s_hv.c | 14 +++++++++++++- 3 files changed, 19 insertions(+), 1 deletion(-)