diff mbox

[RFC,Part2,v3,16/26] KVM: SVM: Add support for SEV LAUNCH_UPDATE_DATA command

Message ID 20170724200303.12197-17-brijesh.singh@amd.com (mailing list archive)
State New, archived
Headers show

Commit Message

Brijesh Singh July 24, 2017, 8:02 p.m. UTC
The command is used for encrypting the guest memory region using the VM
encryption key (VEK) created during LAUNCH_START.

Signed-off-by: Brijesh Singh <brijesh.singh@amd.com>
---
 arch/x86/kvm/svm.c | 165 +++++++++++++++++++++++++++++++++++++++++++++++++++++
 1 file changed, 165 insertions(+)

Comments

Borislav Petkov Sept. 13, 2017, 5:55 p.m. UTC | #1
On Mon, Jul 24, 2017 at 03:02:53PM -0500, Brijesh Singh wrote:
> The command is used for encrypting the guest memory region using the VM
> encryption key (VEK) created during LAUNCH_START.

Yap, this is one good commit message!

:-)

> Signed-off-by: Brijesh Singh <brijesh.singh@amd.com>
> ---
>  arch/x86/kvm/svm.c | 165 +++++++++++++++++++++++++++++++++++++++++++++++++++++
>  1 file changed, 165 insertions(+)
> 
> diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c
> index 3e325578..91b070f 100644
> --- a/arch/x86/kvm/svm.c
> +++ b/arch/x86/kvm/svm.c
> @@ -39,6 +39,8 @@
>  #include <linux/frame.h>
>  #include <linux/psp-sev.h>
>  #include <linux/file.h>
> +#include <linux/pagemap.h>
> +#include <linux/swap.h>
>  
>  #include <asm/apic.h>
>  #include <asm/perf_event.h>
> @@ -331,6 +333,7 @@ static int sev_asid_new(void);
>  static void sev_asid_free(int asid);
>  static void sev_deactivate_handle(struct kvm *kvm, int *error);
>  static void sev_decommission_handle(struct kvm *kvm, int *error);
> +#define __sme_page_pa(x) __sme_set(page_to_pfn(x) << PAGE_SHIFT)
>  
>  static bool svm_sev_enabled(void)
>  {
> @@ -5796,6 +5799,164 @@ static int sev_launch_start(struct kvm *kvm, struct kvm_sev_cmd *argp)
>  	return ret;
>  }
>  
> +static struct page **sev_pin_memory(unsigned long uaddr, unsigned long ulen,
> +				    unsigned long *n, int write)
> +{
> +	unsigned long npages, pinned, size;
> +	struct page **pages;
> +	int first, last;
> +
> +	/* Get number of pages */
> +	first = (uaddr & PAGE_MASK) >> PAGE_SHIFT;
> +	last = ((uaddr + ulen - 1) & PAGE_MASK) >> PAGE_SHIFT;
> +	npages = (last - first + 1);
> +
> +	/* Avoid using vmalloc for smaller buffer */
> +	size = npages * sizeof(struct page *);
> +	if (size > PAGE_SIZE)
> +		pages = vmalloc(size);
> +	else
> +		pages = kmalloc(size, GFP_KERNEL);
> +
> +	if (!pages)
> +		return NULL;
> +
> +	/* pin the user virtual address */
> +	pinned = get_user_pages_fast(uaddr, npages, write ? FOLL_WRITE : 0,
> +					pages);

Let it stick out.

> +	if (pinned != npages) {
> +		pr_err("failed to pin %ld pages (got %ld)\n", npages, pinned);
> +		goto err;
> +	}
> +
> +	*n = npages;
> +	return pages;
> +err:
> +	if (pinned > 0)
> +		release_pages(pages, pinned, 0);

<---- newline here.

> +	kvfree(pages);
> +
> +	return NULL;
> +}
> +
> +static void sev_unpin_memory(struct page **pages, unsigned long npages)
> +{
> +	release_pages(pages, npages, 0);
> +	kvfree(pages);
> +}
> +
> +static void sev_clflush_pages(struct page *pages[], unsigned long npages)
> +{
> +	uint8_t *page_virtual;
> +	unsigned long i;
> +
> +	if (npages == 0 || pages == NULL)
> +		return;
> +
> +	for (i = 0; i < npages; i++) {
> +		page_virtual = kmap_atomic(pages[i]);
> +		clflush_cache_range(page_virtual, PAGE_SIZE);
> +		kunmap_atomic(page_virtual);
> +	}
> +}
> +
> +static int get_num_contig_pages(int idx, struct page **inpages,
> +				unsigned long npages)
> +{
> +	int i = idx + 1, pages = 1;
> +	unsigned long paddr, next_paddr;
> +
> +	/* find the number of contiguous pages starting from idx */
> +	paddr = __sme_page_pa(inpages[idx]);
> +	while (i < npages) {
> +		next_paddr = __sme_page_pa(inpages[i++]);
> +		if ((paddr + PAGE_SIZE) == next_paddr) {
> +			pages++;
> +			paddr = next_paddr;
> +			continue;
> +		}
> +		break;
> +	}
> +
> +	return pages;
> +}
> +
> +static int sev_launch_update_data(struct kvm *kvm, struct kvm_sev_cmd *argp)
> +{
> +	unsigned long vaddr, vaddr_end, next_vaddr, npages, size;
> +	struct kvm_sev_launch_update_data params;
> +	struct sev_data_launch_update_data *data;
> +	struct page **inpages;
> +	int i, ret, pages;
> +
> +	if (!sev_guest(kvm))
> +		return -ENOTTY;
> +
> +	if (copy_from_user(&params, (void *)argp->data,
> +			sizeof(struct kvm_sev_launch_update_data)))
> +		return -EFAULT;
> +
> +	data = kzalloc(sizeof(*data), GFP_KERNEL);
> +	if (!data)
> +		return -ENOMEM;

Same issues as before.

> +
> +	vaddr = params.address;
> +	size = params.length;
> +	vaddr_end = vaddr + size;
> +
> +	/* lock the user memory */
> +	inpages = sev_pin_memory(vaddr, size, &npages, 1);

This way user basically controls how many pages to pin and you need to
limit that on the upper end.

> +	if (!inpages) {
> +		ret = -ENOMEM;
> +		goto e_free;
> +	}
> +
> +	/*
> +	 * invalidate the cache to ensure that DRAM has recent content before

recent content?

> +	 * calling the SEV commands.
> +	 */
> +	sev_clflush_pages(inpages, npages);
> +
> +	for (i = 0; vaddr < vaddr_end; vaddr = next_vaddr, i += pages) {
> +		int offset, len;
> +
> +		/*
> +		 * since user buffer may not be page aligned, calculate the
> +		 * offset within the page.
> +		 */
> +		offset = vaddr & (PAGE_SIZE - 1);
> +
> +		/*
> +		 * calculate the number of pages that can be encrypted in one go
> +		 */
> +		pages = get_num_contig_pages(i, inpages, npages);
> +
> +		len = min_t(size_t, ((pages * PAGE_SIZE) - offset), size);
> +
> +		data->handle = sev_get_handle(kvm);
> +		data->length = len;
> +		data->address = __sme_page_pa(inpages[i]) + offset;
> +		ret = sev_issue_cmd(kvm, SEV_CMD_LAUNCH_UPDATE_DATA, data,
> +				&argp->error);

Yah, let it stick out.
Brijesh Singh Sept. 13, 2017, 7:45 p.m. UTC | #2
On 09/13/2017 12:55 PM, Borislav Petkov wrote:
...

>> +
>> +	/* pin the user virtual address */
>> +	pinned = get_user_pages_fast(uaddr, npages, write ? FOLL_WRITE : 0,
>> +					pages);
> 
> Let it stick out.


Will do.

...


>> +	vaddr = params.address;
>> +	size = params.length;
>> +	vaddr_end = vaddr + size;
>> +
>> +	/* lock the user memory */
>> +	inpages = sev_pin_memory(vaddr, size, &npages, 1);
> 
> This way user basically controls how many pages to pin and you need to
> limit that on the upper end.
> 

Actually I don't know what should be sane upper bound in this case --
typically we encrypt the guest BIOS using LAUNCH_UPDATE_DATA command.
I have heard that some user may want to create a pre-encrypted image
(which may contains guest BIOS + kernel + initrd) -- this can be huge.

For SEV guest, we have been needing to pin the memory hence how about if
we limit the number of pages to pin with rlimit ? The rlimit check can
also include the guest RAM pinning.


>> +	if (!inpages) {
>> +		ret = -ENOMEM;
>> +		goto e_free;
>> +	}
>> +
>> +	/*
>> +	 * invalidate the cache to ensure that DRAM has recent content before
> 
> recent content?


Cache access from the PSP are coherent with x86 but not other way around --
I will update the comments to reflect the true meaning.

...

> 
> Yah, let it stick out.
> 

Okay.
Borislav Petkov Sept. 13, 2017, 9:07 p.m. UTC | #3
On Wed, Sep 13, 2017 at 02:45:37PM -0500, Brijesh Singh wrote:
> Actually I don't know what should be sane upper bound in this case --
> typically we encrypt the guest BIOS using LAUNCH_UPDATE_DATA command.
> I have heard that some user may want to create a pre-encrypted image
> (which may contains guest BIOS + kernel + initrd) -- this can be huge.
> 
> For SEV guest, we have been needing to pin the memory hence how about if
> we limit the number of pages to pin with rlimit ? The rlimit check can
> also include the guest RAM pinning.

rlimit sounds like a sensible thing to do. It would be interesting to
hear what the general policy is wrt guest sizes that KVM folk do ...
diff mbox

Patch

diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c
index 3e325578..91b070f 100644
--- a/arch/x86/kvm/svm.c
+++ b/arch/x86/kvm/svm.c
@@ -39,6 +39,8 @@ 
 #include <linux/frame.h>
 #include <linux/psp-sev.h>
 #include <linux/file.h>
+#include <linux/pagemap.h>
+#include <linux/swap.h>
 
 #include <asm/apic.h>
 #include <asm/perf_event.h>
@@ -331,6 +333,7 @@  static int sev_asid_new(void);
 static void sev_asid_free(int asid);
 static void sev_deactivate_handle(struct kvm *kvm, int *error);
 static void sev_decommission_handle(struct kvm *kvm, int *error);
+#define __sme_page_pa(x) __sme_set(page_to_pfn(x) << PAGE_SHIFT)
 
 static bool svm_sev_enabled(void)
 {
@@ -5796,6 +5799,164 @@  static int sev_launch_start(struct kvm *kvm, struct kvm_sev_cmd *argp)
 	return ret;
 }
 
+static struct page **sev_pin_memory(unsigned long uaddr, unsigned long ulen,
+				    unsigned long *n, int write)
+{
+	unsigned long npages, pinned, size;
+	struct page **pages;
+	int first, last;
+
+	/* Get number of pages */
+	first = (uaddr & PAGE_MASK) >> PAGE_SHIFT;
+	last = ((uaddr + ulen - 1) & PAGE_MASK) >> PAGE_SHIFT;
+	npages = (last - first + 1);
+
+	/* Avoid using vmalloc for smaller buffer */
+	size = npages * sizeof(struct page *);
+	if (size > PAGE_SIZE)
+		pages = vmalloc(size);
+	else
+		pages = kmalloc(size, GFP_KERNEL);
+
+	if (!pages)
+		return NULL;
+
+	/* pin the user virtual address */
+	pinned = get_user_pages_fast(uaddr, npages, write ? FOLL_WRITE : 0,
+					pages);
+	if (pinned != npages) {
+		pr_err("failed to pin %ld pages (got %ld)\n", npages, pinned);
+		goto err;
+	}
+
+	*n = npages;
+	return pages;
+err:
+	if (pinned > 0)
+		release_pages(pages, pinned, 0);
+	kvfree(pages);
+
+	return NULL;
+}
+
+static void sev_unpin_memory(struct page **pages, unsigned long npages)
+{
+	release_pages(pages, npages, 0);
+	kvfree(pages);
+}
+
+static void sev_clflush_pages(struct page *pages[], unsigned long npages)
+{
+	uint8_t *page_virtual;
+	unsigned long i;
+
+	if (npages == 0 || pages == NULL)
+		return;
+
+	for (i = 0; i < npages; i++) {
+		page_virtual = kmap_atomic(pages[i]);
+		clflush_cache_range(page_virtual, PAGE_SIZE);
+		kunmap_atomic(page_virtual);
+	}
+}
+
+static int get_num_contig_pages(int idx, struct page **inpages,
+				unsigned long npages)
+{
+	int i = idx + 1, pages = 1;
+	unsigned long paddr, next_paddr;
+
+	/* find the number of contiguous pages starting from idx */
+	paddr = __sme_page_pa(inpages[idx]);
+	while (i < npages) {
+		next_paddr = __sme_page_pa(inpages[i++]);
+		if ((paddr + PAGE_SIZE) == next_paddr) {
+			pages++;
+			paddr = next_paddr;
+			continue;
+		}
+		break;
+	}
+
+	return pages;
+}
+
+static int sev_launch_update_data(struct kvm *kvm, struct kvm_sev_cmd *argp)
+{
+	unsigned long vaddr, vaddr_end, next_vaddr, npages, size;
+	struct kvm_sev_launch_update_data params;
+	struct sev_data_launch_update_data *data;
+	struct page **inpages;
+	int i, ret, pages;
+
+	if (!sev_guest(kvm))
+		return -ENOTTY;
+
+	if (copy_from_user(&params, (void *)argp->data,
+			sizeof(struct kvm_sev_launch_update_data)))
+		return -EFAULT;
+
+	data = kzalloc(sizeof(*data), GFP_KERNEL);
+	if (!data)
+		return -ENOMEM;
+
+	vaddr = params.address;
+	size = params.length;
+	vaddr_end = vaddr + size;
+
+	/* lock the user memory */
+	inpages = sev_pin_memory(vaddr, size, &npages, 1);
+	if (!inpages) {
+		ret = -ENOMEM;
+		goto e_free;
+	}
+
+	/*
+	 * invalidate the cache to ensure that DRAM has recent content before
+	 * calling the SEV commands.
+	 */
+	sev_clflush_pages(inpages, npages);
+
+	for (i = 0; vaddr < vaddr_end; vaddr = next_vaddr, i += pages) {
+		int offset, len;
+
+		/*
+		 * since user buffer may not be page aligned, calculate the
+		 * offset within the page.
+		 */
+		offset = vaddr & (PAGE_SIZE - 1);
+
+		/*
+		 * calculate the number of pages that can be encrypted in one go
+		 */
+		pages = get_num_contig_pages(i, inpages, npages);
+
+		len = min_t(size_t, ((pages * PAGE_SIZE) - offset), size);
+
+		data->handle = sev_get_handle(kvm);
+		data->length = len;
+		data->address = __sme_page_pa(inpages[i]) + offset;
+		ret = sev_issue_cmd(kvm, SEV_CMD_LAUNCH_UPDATE_DATA, data,
+				&argp->error);
+		if (ret)
+			goto e_unpin;
+
+		size -= len;
+		next_vaddr = vaddr + len;
+	}
+e_unpin:
+	/* content of memory is updated, mark pages dirty */
+	for (i = 0; i < npages; i++) {
+		set_page_dirty_lock(inpages[i]);
+		mark_page_accessed(inpages[i]);
+	}
+	/* unlock the user pages */
+	sev_unpin_memory(inpages, npages);
+e_free:
+	kfree(data);
+	return ret;
+}
+
 static int svm_memory_encryption_op(struct kvm *kvm, void __user *argp)
 {
 	struct kvm_sev_cmd sev_cmd;
@@ -5815,6 +5976,10 @@  static int svm_memory_encryption_op(struct kvm *kvm, void __user *argp)
 		r = sev_launch_start(kvm, &sev_cmd);
 		break;
 	}
+	case KVM_SEV_LAUNCH_UPDATE_DATA: {
+		r = sev_launch_update_data(kvm, &sev_cmd);
+		break;
+	}
 	default:
 		break;
 	}