Message ID | 20170724200303.12197-17-brijesh.singh@amd.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On Mon, Jul 24, 2017 at 03:02:53PM -0500, Brijesh Singh wrote: > The command is used for encrypting the guest memory region using the VM > encryption key (VEK) created during LAUNCH_START. Yap, this is one good commit message! :-) > Signed-off-by: Brijesh Singh <brijesh.singh@amd.com> > --- > arch/x86/kvm/svm.c | 165 +++++++++++++++++++++++++++++++++++++++++++++++++++++ > 1 file changed, 165 insertions(+) > > diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c > index 3e325578..91b070f 100644 > --- a/arch/x86/kvm/svm.c > +++ b/arch/x86/kvm/svm.c > @@ -39,6 +39,8 @@ > #include <linux/frame.h> > #include <linux/psp-sev.h> > #include <linux/file.h> > +#include <linux/pagemap.h> > +#include <linux/swap.h> > > #include <asm/apic.h> > #include <asm/perf_event.h> > @@ -331,6 +333,7 @@ static int sev_asid_new(void); > static void sev_asid_free(int asid); > static void sev_deactivate_handle(struct kvm *kvm, int *error); > static void sev_decommission_handle(struct kvm *kvm, int *error); > +#define __sme_page_pa(x) __sme_set(page_to_pfn(x) << PAGE_SHIFT) > > static bool svm_sev_enabled(void) > { > @@ -5796,6 +5799,164 @@ static int sev_launch_start(struct kvm *kvm, struct kvm_sev_cmd *argp) > return ret; > } > > +static struct page **sev_pin_memory(unsigned long uaddr, unsigned long ulen, > + unsigned long *n, int write) > +{ > + unsigned long npages, pinned, size; > + struct page **pages; > + int first, last; > + > + /* Get number of pages */ > + first = (uaddr & PAGE_MASK) >> PAGE_SHIFT; > + last = ((uaddr + ulen - 1) & PAGE_MASK) >> PAGE_SHIFT; > + npages = (last - first + 1); > + > + /* Avoid using vmalloc for smaller buffer */ > + size = npages * sizeof(struct page *); > + if (size > PAGE_SIZE) > + pages = vmalloc(size); > + else > + pages = kmalloc(size, GFP_KERNEL); > + > + if (!pages) > + return NULL; > + > + /* pin the user virtual address */ > + pinned = get_user_pages_fast(uaddr, npages, write ? FOLL_WRITE : 0, > + pages); Let it stick out. > + if (pinned != npages) { > + pr_err("failed to pin %ld pages (got %ld)\n", npages, pinned); > + goto err; > + } > + > + *n = npages; > + return pages; > +err: > + if (pinned > 0) > + release_pages(pages, pinned, 0); <---- newline here. > + kvfree(pages); > + > + return NULL; > +} > + > +static void sev_unpin_memory(struct page **pages, unsigned long npages) > +{ > + release_pages(pages, npages, 0); > + kvfree(pages); > +} > + > +static void sev_clflush_pages(struct page *pages[], unsigned long npages) > +{ > + uint8_t *page_virtual; > + unsigned long i; > + > + if (npages == 0 || pages == NULL) > + return; > + > + for (i = 0; i < npages; i++) { > + page_virtual = kmap_atomic(pages[i]); > + clflush_cache_range(page_virtual, PAGE_SIZE); > + kunmap_atomic(page_virtual); > + } > +} > + > +static int get_num_contig_pages(int idx, struct page **inpages, > + unsigned long npages) > +{ > + int i = idx + 1, pages = 1; > + unsigned long paddr, next_paddr; > + > + /* find the number of contiguous pages starting from idx */ > + paddr = __sme_page_pa(inpages[idx]); > + while (i < npages) { > + next_paddr = __sme_page_pa(inpages[i++]); > + if ((paddr + PAGE_SIZE) == next_paddr) { > + pages++; > + paddr = next_paddr; > + continue; > + } > + break; > + } > + > + return pages; > +} > + > +static int sev_launch_update_data(struct kvm *kvm, struct kvm_sev_cmd *argp) > +{ > + unsigned long vaddr, vaddr_end, next_vaddr, npages, size; > + struct kvm_sev_launch_update_data params; > + struct sev_data_launch_update_data *data; > + struct page **inpages; > + int i, ret, pages; > + > + if (!sev_guest(kvm)) > + return -ENOTTY; > + > + if (copy_from_user(¶ms, (void *)argp->data, > + sizeof(struct kvm_sev_launch_update_data))) > + return -EFAULT; > + > + data = kzalloc(sizeof(*data), GFP_KERNEL); > + if (!data) > + return -ENOMEM; Same issues as before. > + > + vaddr = params.address; > + size = params.length; > + vaddr_end = vaddr + size; > + > + /* lock the user memory */ > + inpages = sev_pin_memory(vaddr, size, &npages, 1); This way user basically controls how many pages to pin and you need to limit that on the upper end. > + if (!inpages) { > + ret = -ENOMEM; > + goto e_free; > + } > + > + /* > + * invalidate the cache to ensure that DRAM has recent content before recent content? > + * calling the SEV commands. > + */ > + sev_clflush_pages(inpages, npages); > + > + for (i = 0; vaddr < vaddr_end; vaddr = next_vaddr, i += pages) { > + int offset, len; > + > + /* > + * since user buffer may not be page aligned, calculate the > + * offset within the page. > + */ > + offset = vaddr & (PAGE_SIZE - 1); > + > + /* > + * calculate the number of pages that can be encrypted in one go > + */ > + pages = get_num_contig_pages(i, inpages, npages); > + > + len = min_t(size_t, ((pages * PAGE_SIZE) - offset), size); > + > + data->handle = sev_get_handle(kvm); > + data->length = len; > + data->address = __sme_page_pa(inpages[i]) + offset; > + ret = sev_issue_cmd(kvm, SEV_CMD_LAUNCH_UPDATE_DATA, data, > + &argp->error); Yah, let it stick out.
On 09/13/2017 12:55 PM, Borislav Petkov wrote: ... >> + >> + /* pin the user virtual address */ >> + pinned = get_user_pages_fast(uaddr, npages, write ? FOLL_WRITE : 0, >> + pages); > > Let it stick out. Will do. ... >> + vaddr = params.address; >> + size = params.length; >> + vaddr_end = vaddr + size; >> + >> + /* lock the user memory */ >> + inpages = sev_pin_memory(vaddr, size, &npages, 1); > > This way user basically controls how many pages to pin and you need to > limit that on the upper end. > Actually I don't know what should be sane upper bound in this case -- typically we encrypt the guest BIOS using LAUNCH_UPDATE_DATA command. I have heard that some user may want to create a pre-encrypted image (which may contains guest BIOS + kernel + initrd) -- this can be huge. For SEV guest, we have been needing to pin the memory hence how about if we limit the number of pages to pin with rlimit ? The rlimit check can also include the guest RAM pinning. >> + if (!inpages) { >> + ret = -ENOMEM; >> + goto e_free; >> + } >> + >> + /* >> + * invalidate the cache to ensure that DRAM has recent content before > > recent content? Cache access from the PSP are coherent with x86 but not other way around -- I will update the comments to reflect the true meaning. ... > > Yah, let it stick out. > Okay.
On Wed, Sep 13, 2017 at 02:45:37PM -0500, Brijesh Singh wrote: > Actually I don't know what should be sane upper bound in this case -- > typically we encrypt the guest BIOS using LAUNCH_UPDATE_DATA command. > I have heard that some user may want to create a pre-encrypted image > (which may contains guest BIOS + kernel + initrd) -- this can be huge. > > For SEV guest, we have been needing to pin the memory hence how about if > we limit the number of pages to pin with rlimit ? The rlimit check can > also include the guest RAM pinning. rlimit sounds like a sensible thing to do. It would be interesting to hear what the general policy is wrt guest sizes that KVM folk do ...
diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c index 3e325578..91b070f 100644 --- a/arch/x86/kvm/svm.c +++ b/arch/x86/kvm/svm.c @@ -39,6 +39,8 @@ #include <linux/frame.h> #include <linux/psp-sev.h> #include <linux/file.h> +#include <linux/pagemap.h> +#include <linux/swap.h> #include <asm/apic.h> #include <asm/perf_event.h> @@ -331,6 +333,7 @@ static int sev_asid_new(void); static void sev_asid_free(int asid); static void sev_deactivate_handle(struct kvm *kvm, int *error); static void sev_decommission_handle(struct kvm *kvm, int *error); +#define __sme_page_pa(x) __sme_set(page_to_pfn(x) << PAGE_SHIFT) static bool svm_sev_enabled(void) { @@ -5796,6 +5799,164 @@ static int sev_launch_start(struct kvm *kvm, struct kvm_sev_cmd *argp) return ret; } +static struct page **sev_pin_memory(unsigned long uaddr, unsigned long ulen, + unsigned long *n, int write) +{ + unsigned long npages, pinned, size; + struct page **pages; + int first, last; + + /* Get number of pages */ + first = (uaddr & PAGE_MASK) >> PAGE_SHIFT; + last = ((uaddr + ulen - 1) & PAGE_MASK) >> PAGE_SHIFT; + npages = (last - first + 1); + + /* Avoid using vmalloc for smaller buffer */ + size = npages * sizeof(struct page *); + if (size > PAGE_SIZE) + pages = vmalloc(size); + else + pages = kmalloc(size, GFP_KERNEL); + + if (!pages) + return NULL; + + /* pin the user virtual address */ + pinned = get_user_pages_fast(uaddr, npages, write ? FOLL_WRITE : 0, + pages); + if (pinned != npages) { + pr_err("failed to pin %ld pages (got %ld)\n", npages, pinned); + goto err; + } + + *n = npages; + return pages; +err: + if (pinned > 0) + release_pages(pages, pinned, 0); + kvfree(pages); + + return NULL; +} + +static void sev_unpin_memory(struct page **pages, unsigned long npages) +{ + release_pages(pages, npages, 0); + kvfree(pages); +} + +static void sev_clflush_pages(struct page *pages[], unsigned long npages) +{ + uint8_t *page_virtual; + unsigned long i; + + if (npages == 0 || pages == NULL) + return; + + for (i = 0; i < npages; i++) { + page_virtual = kmap_atomic(pages[i]); + clflush_cache_range(page_virtual, PAGE_SIZE); + kunmap_atomic(page_virtual); + } +} + +static int get_num_contig_pages(int idx, struct page **inpages, + unsigned long npages) +{ + int i = idx + 1, pages = 1; + unsigned long paddr, next_paddr; + + /* find the number of contiguous pages starting from idx */ + paddr = __sme_page_pa(inpages[idx]); + while (i < npages) { + next_paddr = __sme_page_pa(inpages[i++]); + if ((paddr + PAGE_SIZE) == next_paddr) { + pages++; + paddr = next_paddr; + continue; + } + break; + } + + return pages; +} + +static int sev_launch_update_data(struct kvm *kvm, struct kvm_sev_cmd *argp) +{ + unsigned long vaddr, vaddr_end, next_vaddr, npages, size; + struct kvm_sev_launch_update_data params; + struct sev_data_launch_update_data *data; + struct page **inpages; + int i, ret, pages; + + if (!sev_guest(kvm)) + return -ENOTTY; + + if (copy_from_user(¶ms, (void *)argp->data, + sizeof(struct kvm_sev_launch_update_data))) + return -EFAULT; + + data = kzalloc(sizeof(*data), GFP_KERNEL); + if (!data) + return -ENOMEM; + + vaddr = params.address; + size = params.length; + vaddr_end = vaddr + size; + + /* lock the user memory */ + inpages = sev_pin_memory(vaddr, size, &npages, 1); + if (!inpages) { + ret = -ENOMEM; + goto e_free; + } + + /* + * invalidate the cache to ensure that DRAM has recent content before + * calling the SEV commands. + */ + sev_clflush_pages(inpages, npages); + + for (i = 0; vaddr < vaddr_end; vaddr = next_vaddr, i += pages) { + int offset, len; + + /* + * since user buffer may not be page aligned, calculate the + * offset within the page. + */ + offset = vaddr & (PAGE_SIZE - 1); + + /* + * calculate the number of pages that can be encrypted in one go + */ + pages = get_num_contig_pages(i, inpages, npages); + + len = min_t(size_t, ((pages * PAGE_SIZE) - offset), size); + + data->handle = sev_get_handle(kvm); + data->length = len; + data->address = __sme_page_pa(inpages[i]) + offset; + ret = sev_issue_cmd(kvm, SEV_CMD_LAUNCH_UPDATE_DATA, data, + &argp->error); + if (ret) + goto e_unpin; + + size -= len; + next_vaddr = vaddr + len; + } +e_unpin: + /* content of memory is updated, mark pages dirty */ + for (i = 0; i < npages; i++) { + set_page_dirty_lock(inpages[i]); + mark_page_accessed(inpages[i]); + } + /* unlock the user pages */ + sev_unpin_memory(inpages, npages); +e_free: + kfree(data); + return ret; +} + static int svm_memory_encryption_op(struct kvm *kvm, void __user *argp) { struct kvm_sev_cmd sev_cmd; @@ -5815,6 +5976,10 @@ static int svm_memory_encryption_op(struct kvm *kvm, void __user *argp) r = sev_launch_start(kvm, &sev_cmd); break; } + case KVM_SEV_LAUNCH_UPDATE_DATA: { + r = sev_launch_update_data(kvm, &sev_cmd); + break; + } default: break; }
The command is used for encrypting the guest memory region using the VM encryption key (VEK) created during LAUNCH_START. Signed-off-by: Brijesh Singh <brijesh.singh@amd.com> --- arch/x86/kvm/svm.c | 165 +++++++++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 165 insertions(+)