Message ID | 20210730220455.26054-5-peterx@redhat.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | KVM: X86: Some light optimizations on rmap logic | expand |
On 31/07/21 00:04, Peter Xu wrote: > Use this file to dump rmap statistic information. The statistic is done by > calculating the rmap count and the result is log-2-based. > > An example output of this looks like (idle 6GB guest, right after boot linux): > > Rmap_Count: 0 1 2-3 4-7 8-15 16-31 32-63 64-127 128-255 256-511 512-1023 > Level=4K: 3086676 53045 12330 1272 502 121 76 2 0 0 0 > Level=2M: 5947 231 0 0 0 0 0 0 0 0 0 > Level=1G: 32 0 0 0 0 0 0 0 0 0 0 > > Signed-off-by: Peter Xu <peterx@redhat.com> > --- > arch/x86/kvm/x86.c | 113 +++++++++++++++++++++++++++++++++++++++++++++ > 1 file changed, 113 insertions(+) This should be in debugfs.c, meaning that the kvm_mmu_slot_lpages() must be in a header. I think mmu.h should do, let me take a look and I can post myself a v4 of these debugfs parts. Paolo > diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c > index e44d8f7781b6..0877340dc6ff 100644 > --- a/arch/x86/kvm/x86.c > +++ b/arch/x86/kvm/x86.c > @@ -30,6 +30,7 @@ > #include "hyperv.h" > #include "lapic.h" > #include "xen.h" > +#include "mmu/mmu_internal.h" > > #include <linux/clocksource.h> > #include <linux/interrupt.h> > @@ -59,6 +60,7 @@ > #include <linux/mem_encrypt.h> > #include <linux/entry-kvm.h> > #include <linux/suspend.h> > +#include <linux/debugfs.h> > > #include <trace/events/kvm.h> > > @@ -11193,6 +11195,117 @@ int kvm_arch_post_init_vm(struct kvm *kvm) > return kvm_mmu_post_init_vm(kvm); > } > > +/* > + * This covers statistics <1024 (11=log(1024)+1), which should be enough to > + * cover RMAP_RECYCLE_THRESHOLD. > + */ > +#define RMAP_LOG_SIZE 11 > + > +static const char *kvm_lpage_str[KVM_NR_PAGE_SIZES] = { "4K", "2M", "1G" }; > + > +static int kvm_mmu_rmaps_stat_show(struct seq_file *m, void *v) > +{ > + struct kvm_rmap_head *rmap; > + struct kvm *kvm = m->private; > + struct kvm_memory_slot *slot; > + struct kvm_memslots *slots; > + unsigned int lpage_size, index; > + /* Still small enough to be on the stack */ > + unsigned int *log[KVM_NR_PAGE_SIZES], *cur; > + int i, j, k, l, ret; > + > + memset(log, 0, sizeof(log)); > + > + ret = -ENOMEM; > + for (i = 0; i < KVM_NR_PAGE_SIZES; i++) { > + log[i] = kzalloc(RMAP_LOG_SIZE * sizeof(unsigned int), GFP_KERNEL); > + if (!log[i]) > + goto out; > + } > + > + mutex_lock(&kvm->slots_lock); > + write_lock(&kvm->mmu_lock); > + > + for (i = 0; i < KVM_ADDRESS_SPACE_NUM; i++) { > + slots = __kvm_memslots(kvm, i); > + for (j = 0; j < slots->used_slots; j++) { > + slot = &slots->memslots[j]; > + for (k = 0; k < KVM_NR_PAGE_SIZES; k++) { > + rmap = slot->arch.rmap[k]; > + lpage_size = kvm_mmu_slot_lpages(slot, k + 1); > + cur = log[k]; > + for (l = 0; l < lpage_size; l++) { > + index = ffs(pte_list_count(&rmap[l])); > + if (WARN_ON_ONCE(index >= RMAP_LOG_SIZE)) > + index = RMAP_LOG_SIZE - 1; > + cur[index]++; > + } > + } > + } > + } > + > + write_unlock(&kvm->mmu_lock); > + mutex_unlock(&kvm->slots_lock); > + > + /* index=0 counts no rmap; index=1 counts 1 rmap */ > + seq_printf(m, "Rmap_Count:\t0\t1\t"); > + for (i = 2; i < RMAP_LOG_SIZE; i++) { > + j = 1 << (i - 1); > + k = (1 << i) - 1; > + seq_printf(m, "%d-%d\t", j, k); > + } > + seq_printf(m, "\n"); > + > + for (i = 0; i < KVM_NR_PAGE_SIZES; i++) { > + seq_printf(m, "Level=%s:\t", kvm_lpage_str[i]); > + cur = log[i]; > + for (j = 0; j < RMAP_LOG_SIZE; j++) > + seq_printf(m, "%d\t", cur[j]); > + seq_printf(m, "\n"); > + } > + > + ret = 0; > +out: > + for (i = 0; i < KVM_NR_PAGE_SIZES; i++) > + if (log[i]) > + kfree(log[i]); > + > + return ret; > +} > + > +static int kvm_mmu_rmaps_stat_open(struct inode *inode, struct file *file) > +{ > + struct kvm *kvm = inode->i_private; > + > + if (!kvm_get_kvm_safe(kvm)) > + return -ENOENT; > + > + return single_open(file, kvm_mmu_rmaps_stat_show, kvm); > +} > + > +static int kvm_mmu_rmaps_stat_release(struct inode *inode, struct file *file) > +{ > + struct kvm *kvm = inode->i_private; > + > + kvm_put_kvm(kvm); > + > + return single_release(inode, file); > +} > + > +static const struct file_operations mmu_rmaps_stat_fops = { > + .open = kvm_mmu_rmaps_stat_open, > + .read = seq_read, > + .llseek = seq_lseek, > + .release = kvm_mmu_rmaps_stat_release, > +}; > + > +int kvm_arch_create_vm_debugfs(struct kvm *kvm) > +{ > + debugfs_create_file("mmu_rmaps_stat", 0644, kvm->debugfs_dentry, kvm, > + &mmu_rmaps_stat_fops); > + return 0; > +} > + > static void kvm_unload_vcpu_mmu(struct kvm_vcpu *vcpu) > { > vcpu_load(vcpu); >
On Mon, Aug 02, 2021 at 05:25:12PM +0200, Paolo Bonzini wrote: > On 31/07/21 00:04, Peter Xu wrote: > > Use this file to dump rmap statistic information. The statistic is done by > > calculating the rmap count and the result is log-2-based. > > > > An example output of this looks like (idle 6GB guest, right after boot linux): > > > > Rmap_Count: 0 1 2-3 4-7 8-15 16-31 32-63 64-127 128-255 256-511 512-1023 > > Level=4K: 3086676 53045 12330 1272 502 121 76 2 0 0 0 > > Level=2M: 5947 231 0 0 0 0 0 0 0 0 0 > > Level=1G: 32 0 0 0 0 0 0 0 0 0 0 > > > > Signed-off-by: Peter Xu <peterx@redhat.com> > > --- > > arch/x86/kvm/x86.c | 113 +++++++++++++++++++++++++++++++++++++++++++++ > > 1 file changed, 113 insertions(+) > > This should be in debugfs.c, meaning that the kvm_mmu_slot_lpages() must be > in a header. I think mmu.h should do, let me take a look and I can post > myself a v4 of these debugfs parts. Thanks, Paolo!
On Mon, Aug 02, 2021, Paolo Bonzini wrote: > On 31/07/21 00:04, Peter Xu wrote: > > Use this file to dump rmap statistic information. The statistic is done by > > calculating the rmap count and the result is log-2-based. > > > > An example output of this looks like (idle 6GB guest, right after boot linux): > > > > Rmap_Count: 0 1 2-3 4-7 8-15 16-31 32-63 64-127 128-255 256-511 512-1023 > > Level=4K: 3086676 53045 12330 1272 502 121 76 2 0 0 0 > > Level=2M: 5947 231 0 0 0 0 0 0 0 0 0 > > Level=1G: 32 0 0 0 0 0 0 0 0 0 0 > > > > Signed-off-by: Peter Xu <peterx@redhat.com> > > --- > > arch/x86/kvm/x86.c | 113 +++++++++++++++++++++++++++++++++++++++++++++ > > 1 file changed, 113 insertions(+) > > This should be in debugfs.c, meaning that the kvm_mmu_slot_lpages() must be > in a header. I think mmu.h should do, let me take a look and I can post > myself a v4 of these debugfs parts. When you do post v4, don't forget to include both mmu.h and mmu/mmu_internal.h. :-) kvm/queue is still broken... arch/x86/kvm/debugfs.c: In function ‘kvm_mmu_rmaps_stat_show’: arch/x86/kvm/debugfs.c:115:18: error: implicit declaration of function ‘kvm_mmu_slot_lpages’; 115 | lpage_size = kvm_mmu_slot_lpages(slot, k + 1); | ^~~~~~~~~~~~~~~~~~~ | kvm_mmu_gfn_allow_lpage
diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c index e44d8f7781b6..0877340dc6ff 100644 --- a/arch/x86/kvm/x86.c +++ b/arch/x86/kvm/x86.c @@ -30,6 +30,7 @@ #include "hyperv.h" #include "lapic.h" #include "xen.h" +#include "mmu/mmu_internal.h" #include <linux/clocksource.h> #include <linux/interrupt.h> @@ -59,6 +60,7 @@ #include <linux/mem_encrypt.h> #include <linux/entry-kvm.h> #include <linux/suspend.h> +#include <linux/debugfs.h> #include <trace/events/kvm.h> @@ -11193,6 +11195,117 @@ int kvm_arch_post_init_vm(struct kvm *kvm) return kvm_mmu_post_init_vm(kvm); } +/* + * This covers statistics <1024 (11=log(1024)+1), which should be enough to + * cover RMAP_RECYCLE_THRESHOLD. + */ +#define RMAP_LOG_SIZE 11 + +static const char *kvm_lpage_str[KVM_NR_PAGE_SIZES] = { "4K", "2M", "1G" }; + +static int kvm_mmu_rmaps_stat_show(struct seq_file *m, void *v) +{ + struct kvm_rmap_head *rmap; + struct kvm *kvm = m->private; + struct kvm_memory_slot *slot; + struct kvm_memslots *slots; + unsigned int lpage_size, index; + /* Still small enough to be on the stack */ + unsigned int *log[KVM_NR_PAGE_SIZES], *cur; + int i, j, k, l, ret; + + memset(log, 0, sizeof(log)); + + ret = -ENOMEM; + for (i = 0; i < KVM_NR_PAGE_SIZES; i++) { + log[i] = kzalloc(RMAP_LOG_SIZE * sizeof(unsigned int), GFP_KERNEL); + if (!log[i]) + goto out; + } + + mutex_lock(&kvm->slots_lock); + write_lock(&kvm->mmu_lock); + + for (i = 0; i < KVM_ADDRESS_SPACE_NUM; i++) { + slots = __kvm_memslots(kvm, i); + for (j = 0; j < slots->used_slots; j++) { + slot = &slots->memslots[j]; + for (k = 0; k < KVM_NR_PAGE_SIZES; k++) { + rmap = slot->arch.rmap[k]; + lpage_size = kvm_mmu_slot_lpages(slot, k + 1); + cur = log[k]; + for (l = 0; l < lpage_size; l++) { + index = ffs(pte_list_count(&rmap[l])); + if (WARN_ON_ONCE(index >= RMAP_LOG_SIZE)) + index = RMAP_LOG_SIZE - 1; + cur[index]++; + } + } + } + } + + write_unlock(&kvm->mmu_lock); + mutex_unlock(&kvm->slots_lock); + + /* index=0 counts no rmap; index=1 counts 1 rmap */ + seq_printf(m, "Rmap_Count:\t0\t1\t"); + for (i = 2; i < RMAP_LOG_SIZE; i++) { + j = 1 << (i - 1); + k = (1 << i) - 1; + seq_printf(m, "%d-%d\t", j, k); + } + seq_printf(m, "\n"); + + for (i = 0; i < KVM_NR_PAGE_SIZES; i++) { + seq_printf(m, "Level=%s:\t", kvm_lpage_str[i]); + cur = log[i]; + for (j = 0; j < RMAP_LOG_SIZE; j++) + seq_printf(m, "%d\t", cur[j]); + seq_printf(m, "\n"); + } + + ret = 0; +out: + for (i = 0; i < KVM_NR_PAGE_SIZES; i++) + if (log[i]) + kfree(log[i]); + + return ret; +} + +static int kvm_mmu_rmaps_stat_open(struct inode *inode, struct file *file) +{ + struct kvm *kvm = inode->i_private; + + if (!kvm_get_kvm_safe(kvm)) + return -ENOENT; + + return single_open(file, kvm_mmu_rmaps_stat_show, kvm); +} + +static int kvm_mmu_rmaps_stat_release(struct inode *inode, struct file *file) +{ + struct kvm *kvm = inode->i_private; + + kvm_put_kvm(kvm); + + return single_release(inode, file); +} + +static const struct file_operations mmu_rmaps_stat_fops = { + .open = kvm_mmu_rmaps_stat_open, + .read = seq_read, + .llseek = seq_lseek, + .release = kvm_mmu_rmaps_stat_release, +}; + +int kvm_arch_create_vm_debugfs(struct kvm *kvm) +{ + debugfs_create_file("mmu_rmaps_stat", 0644, kvm->debugfs_dentry, kvm, + &mmu_rmaps_stat_fops); + return 0; +} + static void kvm_unload_vcpu_mmu(struct kvm_vcpu *vcpu) { vcpu_load(vcpu);
Use this file to dump rmap statistic information. The statistic is done by calculating the rmap count and the result is log-2-based. An example output of this looks like (idle 6GB guest, right after boot linux): Rmap_Count: 0 1 2-3 4-7 8-15 16-31 32-63 64-127 128-255 256-511 512-1023 Level=4K: 3086676 53045 12330 1272 502 121 76 2 0 0 0 Level=2M: 5947 231 0 0 0 0 0 0 0 0 0 Level=1G: 32 0 0 0 0 0 0 0 0 0 0 Signed-off-by: Peter Xu <peterx@redhat.com> --- arch/x86/kvm/x86.c | 113 +++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 113 insertions(+)