diff mbox series

[v4,16/18] nitro_enclaves: Add sample for ioctl interface usage

Message ID 20200622200329.52996-17-andraprs@amazon.com (mailing list archive)
State New, archived
Headers show
Series Add support for Nitro Enclaves | expand

Commit Message

Paraschiv, Andra-Irina June 22, 2020, 8:03 p.m. UTC
Signed-off-by: Alexandru Vasile <lexnv@amazon.com>
Signed-off-by: Andra Paraschiv <andraprs@amazon.com>
---
Changelog

v3 -> v4

* Update usage details to match the updates in v4.
* Update NE ioctl interface usage.

v2 -> v3

* Remove the include directory to use the uapi from the kernel.
* Remove the GPL additional wording as SPDX-License-Identifier is
  already in place.

v1 -> v2

* New in v2.
---
 samples/nitro_enclaves/.gitignore        |   2 +
 samples/nitro_enclaves/Makefile          |  16 +
 samples/nitro_enclaves/ne_ioctl_sample.c | 520 +++++++++++++++++++++++
 3 files changed, 538 insertions(+)
 create mode 100644 samples/nitro_enclaves/.gitignore
 create mode 100644 samples/nitro_enclaves/Makefile
 create mode 100644 samples/nitro_enclaves/ne_ioctl_sample.c

Comments

Alexander Graf July 6, 2020, 11:39 a.m. UTC | #1
On 22.06.20 22:03, Andra Paraschiv wrote:
> Signed-off-by: Alexandru Vasile <lexnv@amazon.com>
> Signed-off-by: Andra Paraschiv <andraprs@amazon.com>
> ---
> Changelog
> 
> v3 -> v4
> 
> * Update usage details to match the updates in v4.
> * Update NE ioctl interface usage.
> 
> v2 -> v3
> 
> * Remove the include directory to use the uapi from the kernel.
> * Remove the GPL additional wording as SPDX-License-Identifier is
>    already in place.
> 
> v1 -> v2
> 
> * New in v2.
> ---
>   samples/nitro_enclaves/.gitignore        |   2 +
>   samples/nitro_enclaves/Makefile          |  16 +
>   samples/nitro_enclaves/ne_ioctl_sample.c | 520 +++++++++++++++++++++++
>   3 files changed, 538 insertions(+)
>   create mode 100644 samples/nitro_enclaves/.gitignore
>   create mode 100644 samples/nitro_enclaves/Makefile
>   create mode 100644 samples/nitro_enclaves/ne_ioctl_sample.c
> 
> diff --git a/samples/nitro_enclaves/.gitignore b/samples/nitro_enclaves/.gitignore
> new file mode 100644
> index 000000000000..827934129c90
> --- /dev/null
> +++ b/samples/nitro_enclaves/.gitignore
> @@ -0,0 +1,2 @@
> +# SPDX-License-Identifier: GPL-2.0
> +ne_ioctl_sample
> diff --git a/samples/nitro_enclaves/Makefile b/samples/nitro_enclaves/Makefile
> new file mode 100644
> index 000000000000..a3ec78fefb52
> --- /dev/null
> +++ b/samples/nitro_enclaves/Makefile
> @@ -0,0 +1,16 @@
> +# SPDX-License-Identifier: GPL-2.0
> +#
> +# Copyright 2020 Amazon.com, Inc. or its affiliates. All Rights Reserved.
> +
> +# Enclave lifetime management support for Nitro Enclaves (NE) - ioctl sample
> +# usage.
> +
> +.PHONY: all clean
> +
> +CFLAGS += -Wall
> +
> +all:
> +	$(CC) $(CFLAGS) -o ne_ioctl_sample ne_ioctl_sample.c -lpthread
> +
> +clean:
> +	rm -f ne_ioctl_sample
> diff --git a/samples/nitro_enclaves/ne_ioctl_sample.c b/samples/nitro_enclaves/ne_ioctl_sample.c
> new file mode 100644
> index 000000000000..572143d55d77
> --- /dev/null
> +++ b/samples/nitro_enclaves/ne_ioctl_sample.c
> @@ -0,0 +1,520 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Copyright 2020 Amazon.com, Inc. or its affiliates. All Rights Reserved.
> + */
> +
> +/**
> + * Sample flow of using the ioctl interface provided by the Nitro Enclaves (NE)
> + * kernel driver.
> + *
> + * Usage
> + * -----
> + *
> + * Load the nitro_enclaves module, setting also the enclave CPU pool. The
> + * enclave CPUs need to be full cores from the same NUMA node. CPU 0 and its
> + * siblings have to remain available for the primary / parent VM, so they
> + * cannot be included in the enclave CPU pool.
> + *
> + * See the cpu list section from the kernel documentation.
> + * https://www.kernel.org/doc/html/latest/admin-guide/kernel-parameters.html
> + *
> + *	insmod drivers/virt/nitro_enclaves/nitro_enclaves.ko
> + *	lsmod
> + *
> + *	The CPU pool can be set at runtime, after the kernel module is loaded.
> + *
> + *	echo <cpu-list> > /sys/module/nitro_enclaves/parameters/ne_cpus
> + *
> + *	NUMA and CPU siblings information can be found using
> + *
> + *	lscpu
> + *	/proc/cpuinfo
> + *
> + * Check the online / offline CPU list. The CPUs from the pool should be
> + * offlined.
> + *
> + *	lscpu
> + *
> + * Check dmesg for any warnings / errors through the NE driver lifetime / usage.
> + * The NE logs contain the "nitro_enclaves" or "pci 0000:00:02.0" pattern.
> + *
> + *	dmesg
> + *
> + * Setup hugetlbfs huge pages. The memory needs to be from the same NUMA node as
> + * the enclave CPUs.
> + * https://www.kernel.org/doc/Documentation/vm/hugetlbpage.txt
> + *
> + *	echo <nr_hugepages> > /proc/sys/vm/nr_hugepages
> + *
> + *	or set the number of 2 MiB / 1 GiB hugepages using
> + *
> + *	/sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
> + *	/sys/kernel/mm/hugepages/hugepages-1048576kB/nr_hugepages
> + *
> + *	In this example 256 hugepages of 2 MiB are used.
> + *
> + * Build and run the NE sample.
> + *
> + *	make -C samples/nitro_enclaves clean
> + *	make -C samples/nitro_enclaves
> + *	./samples/nitro_enclaves/ne_ioctl_sample <path_to_enclave_image>
> + *
> + * Unload the nitro_enclaves module.
> + *
> + *	rmmod nitro_enclaves
> + *	lsmod
> + */
> +
> +#include <stdio.h>
> +#include <stdlib.h>
> +#include <errno.h>
> +#include <fcntl.h>
> +#include <limits.h>
> +#include <poll.h>
> +#include <pthread.h>
> +#include <string.h>
> +#include <sys/ioctl.h>
> +#include <sys/eventfd.h>
> +#include <sys/mman.h>
> +#include <sys/socket.h>
> +#include <sys/types.h>
> +#include <unistd.h>
> +
> +#include <linux/nitro_enclaves.h>
> +#include <linux/vm_sockets.h>
> +
> +/* Nitro Enclaves (NE) misc device that provides the ioctl interface. */
> +#define NE_DEV_NAME "/dev/nitro_enclaves"
> +#define NE_EXPECTED_API_VERSION (1)
> +
> +/* Timeout in seconds / milliseconds for each poll event. */
> +#define NE_POLL_WAIT_TIME (60)
> +#define NE_POLL_WAIT_TIME_MS (NE_POLL_WAIT_TIME * 1000)
> +
> +/* Amount of time in seconds for the process to keep the enclave alive. */
> +#define NE_SLEEP_TIME (300)
> +
> +/* Enclave vCPUs metadata. */
> +#define NE_DEFAULT_NR_VCPUS (2)
> +
> +/* Enclave memory metadata */
> +
> +/* Min memory size - 2 MiB */
> +#define NE_MIN_MEM_REGION_SIZE (2 * 1024 * 1024)
> +
> +/* 256 memory regions of 2 MiB */
> +#define NE_DEFAULT_NR_MEM_REGIONS (256)
> +
> +/* Vsock addressing for enclave image loading heartbeat. */
> +#define NE_IMAGE_LOAD_VSOCK_CID (3)
> +#define NE_IMAGE_LOAD_VSOCK_PORT (9000)
> +#define NE_IMAGE_LOAD_HEARTBEAT_VALUE (0xb7)
> +
> +struct ne_mem_region {
> +	void *mem_addr;
> +	size_t mem_size;
> +};
> +
> +struct ne_vcpu {
> +	int vcpu_fd;
> +	unsigned int vcpu_id;
> +};
> +
> +/* Thread function for polling the enclave fd. */
> +void *ne_poll_enclave_fd(void *data)
> +{
> +	int enclave_fd = *(int *)data;
> +	struct pollfd fds[1] = {};
> +	int i = 0;
> +	int rc = 0;
> +
> +	printf("Running from poll thread, enclave fd %d\n", enclave_fd);
> +
> +	fds[0].fd = enclave_fd;
> +	fds[0].events = POLLIN | POLLERR | POLLHUP;
> +
> +	/* Keep on polling until the current process is terminated. */
> +	while (1) {
> +		printf("[iter %d] Polling ...\n", i);
> +
> +		rc = poll(fds, 1, NE_POLL_WAIT_TIME_MS);
> +		if (rc < 0) {
> +			printf("Error in poll [%m]\n");
> +
> +			return NULL;
> +		}
> +
> +		i++;
> +
> +		if (!rc) {
> +			printf("Poll: %d seconds elapsed\n",
> +			       i * NE_POLL_WAIT_TIME);
> +
> +			continue;
> +		}
> +
> +		printf("Poll received value %d\n", fds[0].revents);
> +	}
> +
> +	return NULL;
> +}
> +
> +/* Allocate memory region that will be used for the enclave. */
> +static int ne_alloc_mem_region(struct ne_mem_region *ne_mem_region)
> +{
> +	if (!ne_mem_region)
> +		return -EINVAL;
> +
> +	if (!ne_mem_region->mem_size)
> +		return -EINVAL;
> +
> +	ne_mem_region->mem_addr = mmap(NULL, ne_mem_region->mem_size,
> +				       PROT_READ | PROT_WRITE,
> +				       MAP_PRIVATE | MAP_ANONYMOUS |
> +				       MAP_HUGETLB, -1, 0);
> +	if (ne_mem_region->mem_addr == MAP_FAILED) {
> +		printf("Error in mmap memory [%m]\n");
> +
> +		return -1;
> +	}
> +
> +	return 0;
> +}
> +
> +/* Place enclave image in enclave memory. */
> +static int ne_load_enclave_image(int enclave_fd,
> +	struct ne_mem_region ne_mem_regions[], char enclave_image_path[])
> +{
> +	struct ne_image_load_info image_load_info = {};
> +	int rc = 0;
> +
> +	if (enclave_fd < 0)
> +		return -EINVAL;
> +
> +	image_load_info.flags = NE_EIF_IMAGE;
> +
> +	rc = ioctl(enclave_fd, NE_GET_IMAGE_LOAD_INFO, &image_load_info);
> +	if (rc < 0) {
> +		printf("Error in get image load info [rc=%d]\n", rc);
> +
> +		return rc;
> +	}
> +
> +	printf("Enclave image offset in enclave memory is %lld\n",
> +	       image_load_info.memory_offset);
> +
> +	/*
> +	 * TODO: Copy enclave image in enclave memory starting from the given
> +	 * offset.
> +	 */

Just open and read into the buffer at the given offset? :)

> +
> +	return 0;
> +}
> +
> +/* Wait for a hearbeat from the enclave to check it has booted. */
> +static int ne_check_enclave_booted(void)
> +{
> +	struct sockaddr_vm client_vsock_addr = {};
> +	socklen_t client_vsock_len = sizeof(client_vsock_addr);
> +	struct pollfd fds[1] = {};
> +	int rc = 0;
> +	unsigned char recv_buf = 0;
> +	struct sockaddr_vm server_vsock_addr = {
> +		.svm_family = AF_VSOCK,
> +		.svm_cid = NE_IMAGE_LOAD_VSOCK_CID,
> +		.svm_port = NE_IMAGE_LOAD_VSOCK_PORT,
> +	};
> +	int server_vsock_fd = 0;
> +
> +	server_vsock_fd = socket(AF_VSOCK, SOCK_STREAM, 0);
> +	if (server_vsock_fd < 0) {
> +		rc = server_vsock_fd;
> +
> +		printf("Error in socket [rc=%d]\n", rc);
> +
> +		return rc;
> +	}
> +
> +	rc = bind(server_vsock_fd, (struct sockaddr *)&server_vsock_addr,
> +		  sizeof(server_vsock_addr));
> +	if (rc < 0) {
> +		printf("Error in bind [rc=%d]\n", rc);
> +
> +		goto out;
> +	}
> +
> +	rc = listen(server_vsock_fd, 1);
> +	if (rc < 0) {
> +		printf("Error in listen [rc=%d]\n", rc);
> +
> +		goto out;
> +	}
> +
> +	fds[0].fd = server_vsock_fd;
> +	fds[0].events = POLLIN;
> +
> +	rc = poll(fds, 1, NE_POLL_WAIT_TIME_MS);
> +	if (rc < 0) {
> +		printf("Error in poll [%m]\n");
> +
> +		goto out;
> +	}
> +
> +	if (!rc) {
> +		printf("Poll timeout, %d seconds elapsed\n", NE_POLL_WAIT_TIME);
> +
> +		rc = -ETIMEDOUT;
> +
> +		goto out;
> +	}
> +
> +	if ((fds[0].revents & POLLIN) == 0) {
> +		printf("Poll received value %d\n", fds[0].revents);
> +
> +		rc = -EINVAL;
> +
> +		goto out;
> +	}
> +
> +	rc = accept(server_vsock_fd, (struct sockaddr *)&client_vsock_addr,
> +		    &client_vsock_len);
> +	if (rc < 0) {
> +		printf("Error in accept [rc=%d]\n", rc);
> +
> +		goto out;
> +	}
> +
> +	/*
> +	 * Read the heartbeat value that the init process in the enclave sends
> +	 * after vsock connect.
> +	 */
> +	rc = read(server_vsock_fd, &recv_buf, sizeof(recv_buf));
> +	if (rc < 0) {
> +		printf("Error in read [rc=%d]\n", rc);
> +
> +		goto out;
> +	}
> +
> +	if (rc != sizeof(recv_buf) ||
> +	    recv_buf != NE_IMAGE_LOAD_HEARTBEAT_VALUE) {
> +		printf("Read %d instead of %d\n", recv_buf,
> +		       NE_IMAGE_LOAD_HEARTBEAT_VALUE);
> +
> +		goto out;
> +	}
> +
> +	close(server_vsock_fd);
> +
> +	return 0;
> +
> +out:
> +	close(server_vsock_fd);
> +
> +	return rc;
> +}
> +
> +/* Set memory region for the given enclave. */
> +static int ne_set_mem_region(int enclave_fd, struct ne_mem_region ne_mem_region)
> +{
> +	struct ne_user_memory_region mem_region = {};
> +	int rc = 0;
> +
> +	if (enclave_fd < 0)
> +		return -EINVAL;
> +
> +	mem_region.memory_size = ne_mem_region.mem_size;
> +	mem_region.userspace_addr = (__u64)ne_mem_region.mem_addr;
> +
> +	rc = ioctl(enclave_fd, NE_SET_USER_MEMORY_REGION, &mem_region);
> +	if (rc < 0) {
> +		printf("Error in set user memory region [rc=%d]\n", rc);
> +
> +		return rc;
> +	}
> +
> +	return 0;
> +}
> +
> +/* Unmap all the memory regions that were set aside for the  enclave. */
> +static void ne_free_mem_regions(struct ne_mem_region ne_mem_regions[])
> +{
> +	unsigned int i = 0;
> +
> +	for (i = 0; i < NE_DEFAULT_NR_MEM_REGIONS; i++)
> +		munmap(ne_mem_regions[i].mem_addr, ne_mem_regions[i].mem_size);
> +}
> +
> +/* Create enclave vCPU. */
> +static int ne_create_vcpu(int enclave_fd, struct ne_vcpu *ne_vcpu)
> +{
> +	if (enclave_fd < 0)
> +		return -EINVAL;
> +
> +	if (!ne_vcpu)
> +		return -EINVAL;
> +
> +	ne_vcpu->vcpu_fd = ioctl(enclave_fd, NE_CREATE_VCPU, &ne_vcpu->vcpu_id);
> +	if (ne_vcpu->vcpu_fd < 0) {
> +		printf("Error in create vcpu [rc=%d]\n", ne_vcpu->vcpu_fd);
> +
> +		return ne_vcpu->vcpu_fd;
> +	}
> +
> +	return 0;
> +}
> +
> +/* Release enclave vCPU fd(s). */
> +static void ne_release_vcpus(struct ne_vcpu ne_vcpus[])
> +{
> +	unsigned int i = 0;
> +
> +	for (i = 0; i < NE_DEFAULT_NR_VCPUS; i++)
> +		if (ne_vcpus[i].vcpu_fd > 0)
> +			close(ne_vcpus[i].vcpu_fd);
> +}
> +
> +int main(int argc, char *argv[])
> +{
> +	int enclave_fd = 0;
> +	char enclave_image_path[PATH_MAX] = {};
> +	struct ne_enclave_start_info enclave_start_info = {};
> +	unsigned int i = 0;
> +	int ne_api_version = 0;
> +	int ne_dev_fd = 0;
> +	struct ne_mem_region ne_mem_regions[NE_DEFAULT_NR_MEM_REGIONS] = {};
> +	struct ne_vcpu ne_vcpus[NE_DEFAULT_NR_VCPUS] = {};
> +	int rc = 0;
> +	unsigned long slot_uid = 0;
> +	pthread_t thread_id = 0;
> +
> +	if (argc != 2) {
> +		printf("Usage: %s <path_to_enclave_image>\n", argv[0]);
> +
> +		exit(EXIT_FAILURE);
> +	}
> +
> +	strncpy(enclave_image_path, argv[1], sizeof(enclave_image_path) - 1);

Why can you not just pass argv[1] as path?

> +
> +	ne_dev_fd = open(NE_DEV_NAME, O_RDWR | O_CLOEXEC);
> +	if (ne_dev_fd < 0) {
> +		printf("Error in open NE device [rc=%d]\n", ne_dev_fd);
> +
> +		exit(EXIT_FAILURE);
> +	}
> +
> +	ne_api_version = ioctl(ne_dev_fd, NE_GET_API_VERSION);
> +	if (ne_api_version != NE_EXPECTED_API_VERSION) {
> +		printf("Expected API version %d, provided API version %d\n",
> +		       NE_EXPECTED_API_VERSION, ne_api_version);
> +
> +		close(ne_dev_fd);
> +
> +		exit(EXIT_FAILURE);
> +	}
> +
> +	printf("Creating enclave slot ...\n");
> +
> +	enclave_fd = ioctl(ne_dev_fd, NE_CREATE_VM, &slot_uid);
> +
> +	close(ne_dev_fd);
> +
> +	if (enclave_fd < 0) {
> +		printf("Error in create enclave slot [rc=%d]\n", enclave_fd);
> +
> +		exit(EXIT_FAILURE);
> +	}
> +
> +	printf("Enclave fd %d\n", enclave_fd);
> +
> +	rc = pthread_create(&thread_id, NULL, ne_poll_enclave_fd,
> +			    (void *)&enclave_fd);
> +	if (rc < 0) {
> +		printf("Error in thread create [rc=%d]\n", rc);
> +
> +		close(enclave_fd);
> +
> +		exit(EXIT_FAILURE);
> +	}
> +
> +	for (i = 0; i < NE_DEFAULT_NR_MEM_REGIONS; i++) {
> +		ne_mem_regions[i].mem_size = NE_MIN_MEM_REGION_SIZE;
> +		rc = ne_alloc_mem_region(&ne_mem_regions[i]);
> +		if (rc < 0) {
> +			printf("Error in alloc mem region, iter %d [rc=%d]\n",
> +			       i, rc);
> +
> +			goto release_enclave_fd;
> +		}
> +	}
> +
> +	rc = ne_load_enclave_image(enclave_fd, ne_mem_regions,
> +				   enclave_image_path);
> +	if (rc < 0) {
> +		printf("Error in load enclave image [rc=%d]\n", rc);
> +
> +		goto release_enclave_fd;
> +	}
> +
> +	for (i = 0; i < NE_DEFAULT_NR_MEM_REGIONS; i++) {
> +		rc = ne_set_mem_region(enclave_fd, ne_mem_regions[i]);
> +		if (rc < 0) {
> +			printf("Error in set mem region, iter %d [rc=%d]\n",
> +			       i, rc);
> +
> +			goto release_enclave_fd;
> +		}
> +	}
> +
> +	printf("Enclave memory regions were added\n");
> +
> +	for (i = 0; i < NE_DEFAULT_NR_VCPUS; i++) {
> +		/*
> +		 * The vCPU is chosen from the enclave vCPU pool, if the value
> +		 * of the vcpu_id is 0.
> +		 */
> +		ne_vcpus[i].vcpu_id = 0;
> +		rc = ne_create_vcpu(enclave_fd, &ne_vcpus[i]);
> +		if (rc < 0) {
> +			printf("Error in create vcpu, iter %d [rc=%d]\n",
> +			       i, rc);
> +
> +			goto release_enclave_vcpu_fds;
> +		}
> +	}
> +
> +	printf("Enclave vCPUs were created\n");
> +
> +	rc = ioctl(enclave_fd, NE_START_ENCLAVE, &enclave_start_info);
> +	if (rc < 0) {
> +		printf("Error in start enclave [rc=%d]\n", rc);
> +
> +		goto release_enclave_vcpu_fds;
> +	}
> +
> +	printf("Enclave started, CID %llu\n", enclave_start_info.enclave_cid);
> +
> +	/*
> +	 * TODO: Check for enclave hearbeat after it has started to see if it
> +	 * has booted.
> +	 */

So you wrote the function to check for the heartbeat, but don't call it? 
Why?


Alex

> +
> +	printf("Entering sleep for %d seconds ...\n", NE_SLEEP_TIME);
> +
> +	sleep(NE_SLEEP_TIME);
> +
> +	ne_release_vcpus(ne_vcpus);
> +
> +	close(enclave_fd);
> +
> +	ne_free_mem_regions(ne_mem_regions);
> +
> +	exit(EXIT_SUCCESS);
> +
> +release_enclave_vcpu_fds:
> +	ne_release_vcpus(ne_vcpus);
> +release_enclave_fd:
> +	close(enclave_fd);
> +	ne_free_mem_regions(ne_mem_regions);
> +
> +	exit(EXIT_FAILURE);
> +}
> 



Amazon Development Center Germany GmbH
Krausenstr. 38
10117 Berlin
Geschaeftsfuehrung: Christian Schlaeger, Jonathan Weiss
Eingetragen am Amtsgericht Charlottenburg unter HRB 149173 B
Sitz: Berlin
Ust-ID: DE 289 237 879
Paraschiv, Andra-Irina July 7, 2020, 7:03 p.m. UTC | #2
On 06/07/2020 14:39, Alexander Graf wrote:
>
>
> On 22.06.20 22:03, Andra Paraschiv wrote:
>> Signed-off-by: Alexandru Vasile <lexnv@amazon.com>
>> Signed-off-by: Andra Paraschiv <andraprs@amazon.com>
>> ---
>> Changelog
>>
>> v3 -> v4
>>
>> * Update usage details to match the updates in v4.
>> * Update NE ioctl interface usage.
>>
>> v2 -> v3
>>
>> * Remove the include directory to use the uapi from the kernel.
>> * Remove the GPL additional wording as SPDX-License-Identifier is
>>    already in place.
>>
>> v1 -> v2
>>
>> * New in v2.
>> ---
>>   samples/nitro_enclaves/.gitignore        |   2 +
>>   samples/nitro_enclaves/Makefile          |  16 +
>>   samples/nitro_enclaves/ne_ioctl_sample.c | 520 +++++++++++++++++++++++
>>   3 files changed, 538 insertions(+)
>>   create mode 100644 samples/nitro_enclaves/.gitignore
>>   create mode 100644 samples/nitro_enclaves/Makefile
>>   create mode 100644 samples/nitro_enclaves/ne_ioctl_sample.c
>>
>> diff --git a/samples/nitro_enclaves/.gitignore 
>> b/samples/nitro_enclaves/.gitignore
>> new file mode 100644
>> index 000000000000..827934129c90
>> --- /dev/null
>> +++ b/samples/nitro_enclaves/.gitignore
>> @@ -0,0 +1,2 @@
>> +# SPDX-License-Identifier: GPL-2.0
>> +ne_ioctl_sample
>> diff --git a/samples/nitro_enclaves/Makefile 
>> b/samples/nitro_enclaves/Makefile
>> new file mode 100644
>> index 000000000000..a3ec78fefb52
>> --- /dev/null
>> +++ b/samples/nitro_enclaves/Makefile
>> @@ -0,0 +1,16 @@
>> +# SPDX-License-Identifier: GPL-2.0
>> +#
>> +# Copyright 2020 Amazon.com, Inc. or its affiliates. All Rights 
>> Reserved.
>> +
>> +# Enclave lifetime management support for Nitro Enclaves (NE) - 
>> ioctl sample
>> +# usage.
>> +
>> +.PHONY: all clean
>> +
>> +CFLAGS += -Wall
>> +
>> +all:
>> +    $(CC) $(CFLAGS) -o ne_ioctl_sample ne_ioctl_sample.c -lpthread
>> +
>> +clean:
>> +    rm -f ne_ioctl_sample
>> diff --git a/samples/nitro_enclaves/ne_ioctl_sample.c 
>> b/samples/nitro_enclaves/ne_ioctl_sample.c
>> new file mode 100644
>> index 000000000000..572143d55d77
>> --- /dev/null
>> +++ b/samples/nitro_enclaves/ne_ioctl_sample.c
>> @@ -0,0 +1,520 @@
>> +// SPDX-License-Identifier: GPL-2.0
>> +/*
>> + * Copyright 2020 Amazon.com, Inc. or its affiliates. All Rights 
>> Reserved.
>> + */
>> +
>> +/**
>> + * Sample flow of using the ioctl interface provided by the Nitro 
>> Enclaves (NE)
>> + * kernel driver.
>> + *
>> + * Usage
>> + * -----
>> + *
>> + * Load the nitro_enclaves module, setting also the enclave CPU 
>> pool. The
>> + * enclave CPUs need to be full cores from the same NUMA node. CPU 0 
>> and its
>> + * siblings have to remain available for the primary / parent VM, so 
>> they
>> + * cannot be included in the enclave CPU pool.
>> + *
>> + * See the cpu list section from the kernel documentation.
>> + * 
>> https://www.kernel.org/doc/html/latest/admin-guide/kernel-parameters.html
>> + *
>> + *    insmod drivers/virt/nitro_enclaves/nitro_enclaves.ko
>> + *    lsmod
>> + *
>> + *    The CPU pool can be set at runtime, after the kernel module is 
>> loaded.
>> + *
>> + *    echo <cpu-list> > /sys/module/nitro_enclaves/parameters/ne_cpus
>> + *
>> + *    NUMA and CPU siblings information can be found using
>> + *
>> + *    lscpu
>> + *    /proc/cpuinfo
>> + *
>> + * Check the online / offline CPU list. The CPUs from the pool 
>> should be
>> + * offlined.
>> + *
>> + *    lscpu
>> + *
>> + * Check dmesg for any warnings / errors through the NE driver 
>> lifetime / usage.
>> + * The NE logs contain the "nitro_enclaves" or "pci 0000:00:02.0" 
>> pattern.
>> + *
>> + *    dmesg
>> + *
>> + * Setup hugetlbfs huge pages. The memory needs to be from the same 
>> NUMA node as
>> + * the enclave CPUs.
>> + * https://www.kernel.org/doc/Documentation/vm/hugetlbpage.txt
>> + *
>> + *    echo <nr_hugepages> > /proc/sys/vm/nr_hugepages
>> + *
>> + *    or set the number of 2 MiB / 1 GiB hugepages using
>> + *
>> + *    /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
>> + *    /sys/kernel/mm/hugepages/hugepages-1048576kB/nr_hugepages
>> + *
>> + *    In this example 256 hugepages of 2 MiB are used.
>> + *
>> + * Build and run the NE sample.
>> + *
>> + *    make -C samples/nitro_enclaves clean
>> + *    make -C samples/nitro_enclaves
>> + *    ./samples/nitro_enclaves/ne_ioctl_sample <path_to_enclave_image>
>> + *
>> + * Unload the nitro_enclaves module.
>> + *
>> + *    rmmod nitro_enclaves
>> + *    lsmod
>> + */
>> +
>> +#include <stdio.h>
>> +#include <stdlib.h>
>> +#include <errno.h>
>> +#include <fcntl.h>
>> +#include <limits.h>
>> +#include <poll.h>
>> +#include <pthread.h>
>> +#include <string.h>
>> +#include <sys/ioctl.h>
>> +#include <sys/eventfd.h>
>> +#include <sys/mman.h>
>> +#include <sys/socket.h>
>> +#include <sys/types.h>
>> +#include <unistd.h>
>> +
>> +#include <linux/nitro_enclaves.h>
>> +#include <linux/vm_sockets.h>
>> +
>> +/* Nitro Enclaves (NE) misc device that provides the ioctl 
>> interface. */
>> +#define NE_DEV_NAME "/dev/nitro_enclaves"
>> +#define NE_EXPECTED_API_VERSION (1)
>> +
>> +/* Timeout in seconds / milliseconds for each poll event. */
>> +#define NE_POLL_WAIT_TIME (60)
>> +#define NE_POLL_WAIT_TIME_MS (NE_POLL_WAIT_TIME * 1000)
>> +
>> +/* Amount of time in seconds for the process to keep the enclave 
>> alive. */
>> +#define NE_SLEEP_TIME (300)
>> +
>> +/* Enclave vCPUs metadata. */
>> +#define NE_DEFAULT_NR_VCPUS (2)
>> +
>> +/* Enclave memory metadata */
>> +
>> +/* Min memory size - 2 MiB */
>> +#define NE_MIN_MEM_REGION_SIZE (2 * 1024 * 1024)
>> +
>> +/* 256 memory regions of 2 MiB */
>> +#define NE_DEFAULT_NR_MEM_REGIONS (256)
>> +
>> +/* Vsock addressing for enclave image loading heartbeat. */
>> +#define NE_IMAGE_LOAD_VSOCK_CID (3)
>> +#define NE_IMAGE_LOAD_VSOCK_PORT (9000)
>> +#define NE_IMAGE_LOAD_HEARTBEAT_VALUE (0xb7)
>> +
>> +struct ne_mem_region {
>> +    void *mem_addr;
>> +    size_t mem_size;
>> +};
>> +
>> +struct ne_vcpu {
>> +    int vcpu_fd;
>> +    unsigned int vcpu_id;
>> +};
>> +
>> +/* Thread function for polling the enclave fd. */
>> +void *ne_poll_enclave_fd(void *data)
>> +{
>> +    int enclave_fd = *(int *)data;
>> +    struct pollfd fds[1] = {};
>> +    int i = 0;
>> +    int rc = 0;
>> +
>> +    printf("Running from poll thread, enclave fd %d\n", enclave_fd);
>> +
>> +    fds[0].fd = enclave_fd;
>> +    fds[0].events = POLLIN | POLLERR | POLLHUP;
>> +
>> +    /* Keep on polling until the current process is terminated. */
>> +    while (1) {
>> +        printf("[iter %d] Polling ...\n", i);
>> +
>> +        rc = poll(fds, 1, NE_POLL_WAIT_TIME_MS);
>> +        if (rc < 0) {
>> +            printf("Error in poll [%m]\n");
>> +
>> +            return NULL;
>> +        }
>> +
>> +        i++;
>> +
>> +        if (!rc) {
>> +            printf("Poll: %d seconds elapsed\n",
>> +                   i * NE_POLL_WAIT_TIME);
>> +
>> +            continue;
>> +        }
>> +
>> +        printf("Poll received value %d\n", fds[0].revents);
>> +    }
>> +
>> +    return NULL;
>> +}
>> +
>> +/* Allocate memory region that will be used for the enclave. */
>> +static int ne_alloc_mem_region(struct ne_mem_region *ne_mem_region)
>> +{
>> +    if (!ne_mem_region)
>> +        return -EINVAL;
>> +
>> +    if (!ne_mem_region->mem_size)
>> +        return -EINVAL;
>> +
>> +    ne_mem_region->mem_addr = mmap(NULL, ne_mem_region->mem_size,
>> +                       PROT_READ | PROT_WRITE,
>> +                       MAP_PRIVATE | MAP_ANONYMOUS |
>> +                       MAP_HUGETLB, -1, 0);
>> +    if (ne_mem_region->mem_addr == MAP_FAILED) {
>> +        printf("Error in mmap memory [%m]\n");
>> +
>> +        return -1;
>> +    }
>> +
>> +    return 0;
>> +}
>> +
>> +/* Place enclave image in enclave memory. */
>> +static int ne_load_enclave_image(int enclave_fd,
>> +    struct ne_mem_region ne_mem_regions[], char enclave_image_path[])
>> +{
>> +    struct ne_image_load_info image_load_info = {};
>> +    int rc = 0;
>> +
>> +    if (enclave_fd < 0)
>> +        return -EINVAL;
>> +
>> +    image_load_info.flags = NE_EIF_IMAGE;
>> +
>> +    rc = ioctl(enclave_fd, NE_GET_IMAGE_LOAD_INFO, &image_load_info);
>> +    if (rc < 0) {
>> +        printf("Error in get image load info [rc=%d]\n", rc);
>> +
>> +        return rc;
>> +    }
>> +
>> +    printf("Enclave image offset in enclave memory is %lld\n",
>> +           image_load_info.memory_offset);
>> +
>> +    /*
>> +     * TODO: Copy enclave image in enclave memory starting from the 
>> given
>> +     * offset.
>> +     */
>
> Just open and read into the buffer at the given offset? :)

Aham, there is no big complexity in this. :) I just wanted to have it 
together with the updated functionality on the heartbeat logic below.

>
>> +
>> +    return 0;
>> +}
>> +
>> +/* Wait for a hearbeat from the enclave to check it has booted. */
>> +static int ne_check_enclave_booted(void)
>> +{
>> +    struct sockaddr_vm client_vsock_addr = {};
>> +    socklen_t client_vsock_len = sizeof(client_vsock_addr);
>> +    struct pollfd fds[1] = {};
>> +    int rc = 0;
>> +    unsigned char recv_buf = 0;
>> +    struct sockaddr_vm server_vsock_addr = {
>> +        .svm_family = AF_VSOCK,
>> +        .svm_cid = NE_IMAGE_LOAD_VSOCK_CID,
>> +        .svm_port = NE_IMAGE_LOAD_VSOCK_PORT,
>> +    };
>> +    int server_vsock_fd = 0;
>> +
>> +    server_vsock_fd = socket(AF_VSOCK, SOCK_STREAM, 0);
>> +    if (server_vsock_fd < 0) {
>> +        rc = server_vsock_fd;
>> +
>> +        printf("Error in socket [rc=%d]\n", rc);
>> +
>> +        return rc;
>> +    }
>> +
>> +    rc = bind(server_vsock_fd, (struct sockaddr *)&server_vsock_addr,
>> +          sizeof(server_vsock_addr));
>> +    if (rc < 0) {
>> +        printf("Error in bind [rc=%d]\n", rc);
>> +
>> +        goto out;
>> +    }
>> +
>> +    rc = listen(server_vsock_fd, 1);
>> +    if (rc < 0) {
>> +        printf("Error in listen [rc=%d]\n", rc);
>> +
>> +        goto out;
>> +    }
>> +
>> +    fds[0].fd = server_vsock_fd;
>> +    fds[0].events = POLLIN;
>> +
>> +    rc = poll(fds, 1, NE_POLL_WAIT_TIME_MS);
>> +    if (rc < 0) {
>> +        printf("Error in poll [%m]\n");
>> +
>> +        goto out;
>> +    }
>> +
>> +    if (!rc) {
>> +        printf("Poll timeout, %d seconds elapsed\n", 
>> NE_POLL_WAIT_TIME);
>> +
>> +        rc = -ETIMEDOUT;
>> +
>> +        goto out;
>> +    }
>> +
>> +    if ((fds[0].revents & POLLIN) == 0) {
>> +        printf("Poll received value %d\n", fds[0].revents);
>> +
>> +        rc = -EINVAL;
>> +
>> +        goto out;
>> +    }
>> +
>> +    rc = accept(server_vsock_fd, (struct sockaddr *)&client_vsock_addr,
>> +            &client_vsock_len);
>> +    if (rc < 0) {
>> +        printf("Error in accept [rc=%d]\n", rc);
>> +
>> +        goto out;
>> +    }
>> +
>> +    /*
>> +     * Read the heartbeat value that the init process in the enclave 
>> sends
>> +     * after vsock connect.
>> +     */
>> +    rc = read(server_vsock_fd, &recv_buf, sizeof(recv_buf));
>> +    if (rc < 0) {
>> +        printf("Error in read [rc=%d]\n", rc);
>> +
>> +        goto out;
>> +    }
>> +
>> +    if (rc != sizeof(recv_buf) ||
>> +        recv_buf != NE_IMAGE_LOAD_HEARTBEAT_VALUE) {
>> +        printf("Read %d instead of %d\n", recv_buf,
>> +               NE_IMAGE_LOAD_HEARTBEAT_VALUE);
>> +
>> +        goto out;
>> +    }
>> +
>> +    close(server_vsock_fd);
>> +
>> +    return 0;
>> +
>> +out:
>> +    close(server_vsock_fd);
>> +
>> +    return rc;
>> +}
>> +
>> +/* Set memory region for the given enclave. */
>> +static int ne_set_mem_region(int enclave_fd, struct ne_mem_region 
>> ne_mem_region)
>> +{
>> +    struct ne_user_memory_region mem_region = {};
>> +    int rc = 0;
>> +
>> +    if (enclave_fd < 0)
>> +        return -EINVAL;
>> +
>> +    mem_region.memory_size = ne_mem_region.mem_size;
>> +    mem_region.userspace_addr = (__u64)ne_mem_region.mem_addr;
>> +
>> +    rc = ioctl(enclave_fd, NE_SET_USER_MEMORY_REGION, &mem_region);
>> +    if (rc < 0) {
>> +        printf("Error in set user memory region [rc=%d]\n", rc);
>> +
>> +        return rc;
>> +    }
>> +
>> +    return 0;
>> +}
>> +
>> +/* Unmap all the memory regions that were set aside for the enclave. */
>> +static void ne_free_mem_regions(struct ne_mem_region ne_mem_regions[])
>> +{
>> +    unsigned int i = 0;
>> +
>> +    for (i = 0; i < NE_DEFAULT_NR_MEM_REGIONS; i++)
>> +        munmap(ne_mem_regions[i].mem_addr, ne_mem_regions[i].mem_size);
>> +}
>> +
>> +/* Create enclave vCPU. */
>> +static int ne_create_vcpu(int enclave_fd, struct ne_vcpu *ne_vcpu)
>> +{
>> +    if (enclave_fd < 0)
>> +        return -EINVAL;
>> +
>> +    if (!ne_vcpu)
>> +        return -EINVAL;
>> +
>> +    ne_vcpu->vcpu_fd = ioctl(enclave_fd, NE_CREATE_VCPU, 
>> &ne_vcpu->vcpu_id);
>> +    if (ne_vcpu->vcpu_fd < 0) {
>> +        printf("Error in create vcpu [rc=%d]\n", ne_vcpu->vcpu_fd);
>> +
>> +        return ne_vcpu->vcpu_fd;
>> +    }
>> +
>> +    return 0;
>> +}
>> +
>> +/* Release enclave vCPU fd(s). */
>> +static void ne_release_vcpus(struct ne_vcpu ne_vcpus[])
>> +{
>> +    unsigned int i = 0;
>> +
>> +    for (i = 0; i < NE_DEFAULT_NR_VCPUS; i++)
>> +        if (ne_vcpus[i].vcpu_fd > 0)
>> +            close(ne_vcpus[i].vcpu_fd);
>> +}
>> +
>> +int main(int argc, char *argv[])
>> +{
>> +    int enclave_fd = 0;
>> +    char enclave_image_path[PATH_MAX] = {};
>> +    struct ne_enclave_start_info enclave_start_info = {};
>> +    unsigned int i = 0;
>> +    int ne_api_version = 0;
>> +    int ne_dev_fd = 0;
>> +    struct ne_mem_region ne_mem_regions[NE_DEFAULT_NR_MEM_REGIONS] = 
>> {};
>> +    struct ne_vcpu ne_vcpus[NE_DEFAULT_NR_VCPUS] = {};
>> +    int rc = 0;
>> +    unsigned long slot_uid = 0;
>> +    pthread_t thread_id = 0;
>> +
>> +    if (argc != 2) {
>> +        printf("Usage: %s <path_to_enclave_image>\n", argv[0]);
>> +
>> +        exit(EXIT_FAILURE);
>> +    }
>> +
>> +    strncpy(enclave_image_path, argv[1], sizeof(enclave_image_path) 
>> - 1);
>
> Why can you not just pass argv[1] as path?

I just wanted to limit to PATH_MAX size, but I can have this check on 
argv[1] and then pass it as path.

>
>> +
>> +    ne_dev_fd = open(NE_DEV_NAME, O_RDWR | O_CLOEXEC);
>> +    if (ne_dev_fd < 0) {
>> +        printf("Error in open NE device [rc=%d]\n", ne_dev_fd);
>> +
>> +        exit(EXIT_FAILURE);
>> +    }
>> +
>> +    ne_api_version = ioctl(ne_dev_fd, NE_GET_API_VERSION);
>> +    if (ne_api_version != NE_EXPECTED_API_VERSION) {
>> +        printf("Expected API version %d, provided API version %d\n",
>> +               NE_EXPECTED_API_VERSION, ne_api_version);
>> +
>> +        close(ne_dev_fd);
>> +
>> +        exit(EXIT_FAILURE);
>> +    }
>> +
>> +    printf("Creating enclave slot ...\n");
>> +
>> +    enclave_fd = ioctl(ne_dev_fd, NE_CREATE_VM, &slot_uid);
>> +
>> +    close(ne_dev_fd);
>> +
>> +    if (enclave_fd < 0) {
>> +        printf("Error in create enclave slot [rc=%d]\n", enclave_fd);
>> +
>> +        exit(EXIT_FAILURE);
>> +    }
>> +
>> +    printf("Enclave fd %d\n", enclave_fd);
>> +
>> +    rc = pthread_create(&thread_id, NULL, ne_poll_enclave_fd,
>> +                (void *)&enclave_fd);
>> +    if (rc < 0) {
>> +        printf("Error in thread create [rc=%d]\n", rc);
>> +
>> +        close(enclave_fd);
>> +
>> +        exit(EXIT_FAILURE);
>> +    }
>> +
>> +    for (i = 0; i < NE_DEFAULT_NR_MEM_REGIONS; i++) {
>> +        ne_mem_regions[i].mem_size = NE_MIN_MEM_REGION_SIZE;
>> +        rc = ne_alloc_mem_region(&ne_mem_regions[i]);
>> +        if (rc < 0) {
>> +            printf("Error in alloc mem region, iter %d [rc=%d]\n",
>> +                   i, rc);
>> +
>> +            goto release_enclave_fd;
>> +        }
>> +    }
>> +
>> +    rc = ne_load_enclave_image(enclave_fd, ne_mem_regions,
>> +                   enclave_image_path);
>> +    if (rc < 0) {
>> +        printf("Error in load enclave image [rc=%d]\n", rc);
>> +
>> +        goto release_enclave_fd;
>> +    }
>> +
>> +    for (i = 0; i < NE_DEFAULT_NR_MEM_REGIONS; i++) {
>> +        rc = ne_set_mem_region(enclave_fd, ne_mem_regions[i]);
>> +        if (rc < 0) {
>> +            printf("Error in set mem region, iter %d [rc=%d]\n",
>> +                   i, rc);
>> +
>> +            goto release_enclave_fd;
>> +        }
>> +    }
>> +
>> +    printf("Enclave memory regions were added\n");
>> +
>> +    for (i = 0; i < NE_DEFAULT_NR_VCPUS; i++) {
>> +        /*
>> +         * The vCPU is chosen from the enclave vCPU pool, if the value
>> +         * of the vcpu_id is 0.
>> +         */
>> +        ne_vcpus[i].vcpu_id = 0;
>> +        rc = ne_create_vcpu(enclave_fd, &ne_vcpus[i]);
>> +        if (rc < 0) {
>> +            printf("Error in create vcpu, iter %d [rc=%d]\n",
>> +                   i, rc);
>> +
>> +            goto release_enclave_vcpu_fds;
>> +        }
>> +    }
>> +
>> +    printf("Enclave vCPUs were created\n");
>> +
>> +    rc = ioctl(enclave_fd, NE_START_ENCLAVE, &enclave_start_info);
>> +    if (rc < 0) {
>> +        printf("Error in start enclave [rc=%d]\n", rc);
>> +
>> +        goto release_enclave_vcpu_fds;
>> +    }
>> +
>> +    printf("Enclave started, CID %llu\n", 
>> enclave_start_info.enclave_cid);
>> +
>> +    /*
>> +     * TODO: Check for enclave hearbeat after it has started to see 
>> if it
>> +     * has booted.
>> +     */
>
> So you wrote the function to check for the heartbeat, but don't call 
> it? Why?
>

The logic flow (in the NE user space tooling, not from this sample) was 
in review at the time I added it here and recently has been updated. Now 
that we have completed the reviews, I will update this logic in the 
sample, together with including the enclave image loading in memory code 
bits mentioned above.

Thanks,
Andra

>
>> +
>> +    printf("Entering sleep for %d seconds ...\n", NE_SLEEP_TIME);
>> +
>> +    sleep(NE_SLEEP_TIME);
>> +
>> +    ne_release_vcpus(ne_vcpus);
>> +
>> +    close(enclave_fd);
>> +
>> +    ne_free_mem_regions(ne_mem_regions);
>> +
>> +    exit(EXIT_SUCCESS);
>> +
>> +release_enclave_vcpu_fds:
>> +    ne_release_vcpus(ne_vcpus);
>> +release_enclave_fd:
>> +    close(enclave_fd);
>> +    ne_free_mem_regions(ne_mem_regions);
>> +
>> +    exit(EXIT_FAILURE);
>> +}
>>




Amazon Development Center (Romania) S.R.L. registered office: 27A Sf. Lazar Street, UBC5, floor 2, Iasi, Iasi County, 700045, Romania. Registered in Romania. Registration number J22/2621/2005.
diff mbox series

Patch

diff --git a/samples/nitro_enclaves/.gitignore b/samples/nitro_enclaves/.gitignore
new file mode 100644
index 000000000000..827934129c90
--- /dev/null
+++ b/samples/nitro_enclaves/.gitignore
@@ -0,0 +1,2 @@ 
+# SPDX-License-Identifier: GPL-2.0
+ne_ioctl_sample
diff --git a/samples/nitro_enclaves/Makefile b/samples/nitro_enclaves/Makefile
new file mode 100644
index 000000000000..a3ec78fefb52
--- /dev/null
+++ b/samples/nitro_enclaves/Makefile
@@ -0,0 +1,16 @@ 
+# SPDX-License-Identifier: GPL-2.0
+#
+# Copyright 2020 Amazon.com, Inc. or its affiliates. All Rights Reserved.
+
+# Enclave lifetime management support for Nitro Enclaves (NE) - ioctl sample
+# usage.
+
+.PHONY: all clean
+
+CFLAGS += -Wall
+
+all:
+	$(CC) $(CFLAGS) -o ne_ioctl_sample ne_ioctl_sample.c -lpthread
+
+clean:
+	rm -f ne_ioctl_sample
diff --git a/samples/nitro_enclaves/ne_ioctl_sample.c b/samples/nitro_enclaves/ne_ioctl_sample.c
new file mode 100644
index 000000000000..572143d55d77
--- /dev/null
+++ b/samples/nitro_enclaves/ne_ioctl_sample.c
@@ -0,0 +1,520 @@ 
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright 2020 Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ */
+
+/**
+ * Sample flow of using the ioctl interface provided by the Nitro Enclaves (NE)
+ * kernel driver.
+ *
+ * Usage
+ * -----
+ *
+ * Load the nitro_enclaves module, setting also the enclave CPU pool. The
+ * enclave CPUs need to be full cores from the same NUMA node. CPU 0 and its
+ * siblings have to remain available for the primary / parent VM, so they
+ * cannot be included in the enclave CPU pool.
+ *
+ * See the cpu list section from the kernel documentation.
+ * https://www.kernel.org/doc/html/latest/admin-guide/kernel-parameters.html
+ *
+ *	insmod drivers/virt/nitro_enclaves/nitro_enclaves.ko
+ *	lsmod
+ *
+ *	The CPU pool can be set at runtime, after the kernel module is loaded.
+ *
+ *	echo <cpu-list> > /sys/module/nitro_enclaves/parameters/ne_cpus
+ *
+ *	NUMA and CPU siblings information can be found using
+ *
+ *	lscpu
+ *	/proc/cpuinfo
+ *
+ * Check the online / offline CPU list. The CPUs from the pool should be
+ * offlined.
+ *
+ *	lscpu
+ *
+ * Check dmesg for any warnings / errors through the NE driver lifetime / usage.
+ * The NE logs contain the "nitro_enclaves" or "pci 0000:00:02.0" pattern.
+ *
+ *	dmesg
+ *
+ * Setup hugetlbfs huge pages. The memory needs to be from the same NUMA node as
+ * the enclave CPUs.
+ * https://www.kernel.org/doc/Documentation/vm/hugetlbpage.txt
+ *
+ *	echo <nr_hugepages> > /proc/sys/vm/nr_hugepages
+ *
+ *	or set the number of 2 MiB / 1 GiB hugepages using
+ *
+ *	/sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
+ *	/sys/kernel/mm/hugepages/hugepages-1048576kB/nr_hugepages
+ *
+ *	In this example 256 hugepages of 2 MiB are used.
+ *
+ * Build and run the NE sample.
+ *
+ *	make -C samples/nitro_enclaves clean
+ *	make -C samples/nitro_enclaves
+ *	./samples/nitro_enclaves/ne_ioctl_sample <path_to_enclave_image>
+ *
+ * Unload the nitro_enclaves module.
+ *
+ *	rmmod nitro_enclaves
+ *	lsmod
+ */
+
+#include <stdio.h>
+#include <stdlib.h>
+#include <errno.h>
+#include <fcntl.h>
+#include <limits.h>
+#include <poll.h>
+#include <pthread.h>
+#include <string.h>
+#include <sys/ioctl.h>
+#include <sys/eventfd.h>
+#include <sys/mman.h>
+#include <sys/socket.h>
+#include <sys/types.h>
+#include <unistd.h>
+
+#include <linux/nitro_enclaves.h>
+#include <linux/vm_sockets.h>
+
+/* Nitro Enclaves (NE) misc device that provides the ioctl interface. */
+#define NE_DEV_NAME "/dev/nitro_enclaves"
+#define NE_EXPECTED_API_VERSION (1)
+
+/* Timeout in seconds / milliseconds for each poll event. */
+#define NE_POLL_WAIT_TIME (60)
+#define NE_POLL_WAIT_TIME_MS (NE_POLL_WAIT_TIME * 1000)
+
+/* Amount of time in seconds for the process to keep the enclave alive. */
+#define NE_SLEEP_TIME (300)
+
+/* Enclave vCPUs metadata. */
+#define NE_DEFAULT_NR_VCPUS (2)
+
+/* Enclave memory metadata */
+
+/* Min memory size - 2 MiB */
+#define NE_MIN_MEM_REGION_SIZE (2 * 1024 * 1024)
+
+/* 256 memory regions of 2 MiB */
+#define NE_DEFAULT_NR_MEM_REGIONS (256)
+
+/* Vsock addressing for enclave image loading heartbeat. */
+#define NE_IMAGE_LOAD_VSOCK_CID (3)
+#define NE_IMAGE_LOAD_VSOCK_PORT (9000)
+#define NE_IMAGE_LOAD_HEARTBEAT_VALUE (0xb7)
+
+struct ne_mem_region {
+	void *mem_addr;
+	size_t mem_size;
+};
+
+struct ne_vcpu {
+	int vcpu_fd;
+	unsigned int vcpu_id;
+};
+
+/* Thread function for polling the enclave fd. */
+void *ne_poll_enclave_fd(void *data)
+{
+	int enclave_fd = *(int *)data;
+	struct pollfd fds[1] = {};
+	int i = 0;
+	int rc = 0;
+
+	printf("Running from poll thread, enclave fd %d\n", enclave_fd);
+
+	fds[0].fd = enclave_fd;
+	fds[0].events = POLLIN | POLLERR | POLLHUP;
+
+	/* Keep on polling until the current process is terminated. */
+	while (1) {
+		printf("[iter %d] Polling ...\n", i);
+
+		rc = poll(fds, 1, NE_POLL_WAIT_TIME_MS);
+		if (rc < 0) {
+			printf("Error in poll [%m]\n");
+
+			return NULL;
+		}
+
+		i++;
+
+		if (!rc) {
+			printf("Poll: %d seconds elapsed\n",
+			       i * NE_POLL_WAIT_TIME);
+
+			continue;
+		}
+
+		printf("Poll received value %d\n", fds[0].revents);
+	}
+
+	return NULL;
+}
+
+/* Allocate memory region that will be used for the enclave. */
+static int ne_alloc_mem_region(struct ne_mem_region *ne_mem_region)
+{
+	if (!ne_mem_region)
+		return -EINVAL;
+
+	if (!ne_mem_region->mem_size)
+		return -EINVAL;
+
+	ne_mem_region->mem_addr = mmap(NULL, ne_mem_region->mem_size,
+				       PROT_READ | PROT_WRITE,
+				       MAP_PRIVATE | MAP_ANONYMOUS |
+				       MAP_HUGETLB, -1, 0);
+	if (ne_mem_region->mem_addr == MAP_FAILED) {
+		printf("Error in mmap memory [%m]\n");
+
+		return -1;
+	}
+
+	return 0;
+}
+
+/* Place enclave image in enclave memory. */
+static int ne_load_enclave_image(int enclave_fd,
+	struct ne_mem_region ne_mem_regions[], char enclave_image_path[])
+{
+	struct ne_image_load_info image_load_info = {};
+	int rc = 0;
+
+	if (enclave_fd < 0)
+		return -EINVAL;
+
+	image_load_info.flags = NE_EIF_IMAGE;
+
+	rc = ioctl(enclave_fd, NE_GET_IMAGE_LOAD_INFO, &image_load_info);
+	if (rc < 0) {
+		printf("Error in get image load info [rc=%d]\n", rc);
+
+		return rc;
+	}
+
+	printf("Enclave image offset in enclave memory is %lld\n",
+	       image_load_info.memory_offset);
+
+	/*
+	 * TODO: Copy enclave image in enclave memory starting from the given
+	 * offset.
+	 */
+
+	return 0;
+}
+
+/* Wait for a hearbeat from the enclave to check it has booted. */
+static int ne_check_enclave_booted(void)
+{
+	struct sockaddr_vm client_vsock_addr = {};
+	socklen_t client_vsock_len = sizeof(client_vsock_addr);
+	struct pollfd fds[1] = {};
+	int rc = 0;
+	unsigned char recv_buf = 0;
+	struct sockaddr_vm server_vsock_addr = {
+		.svm_family = AF_VSOCK,
+		.svm_cid = NE_IMAGE_LOAD_VSOCK_CID,
+		.svm_port = NE_IMAGE_LOAD_VSOCK_PORT,
+	};
+	int server_vsock_fd = 0;
+
+	server_vsock_fd = socket(AF_VSOCK, SOCK_STREAM, 0);
+	if (server_vsock_fd < 0) {
+		rc = server_vsock_fd;
+
+		printf("Error in socket [rc=%d]\n", rc);
+
+		return rc;
+	}
+
+	rc = bind(server_vsock_fd, (struct sockaddr *)&server_vsock_addr,
+		  sizeof(server_vsock_addr));
+	if (rc < 0) {
+		printf("Error in bind [rc=%d]\n", rc);
+
+		goto out;
+	}
+
+	rc = listen(server_vsock_fd, 1);
+	if (rc < 0) {
+		printf("Error in listen [rc=%d]\n", rc);
+
+		goto out;
+	}
+
+	fds[0].fd = server_vsock_fd;
+	fds[0].events = POLLIN;
+
+	rc = poll(fds, 1, NE_POLL_WAIT_TIME_MS);
+	if (rc < 0) {
+		printf("Error in poll [%m]\n");
+
+		goto out;
+	}
+
+	if (!rc) {
+		printf("Poll timeout, %d seconds elapsed\n", NE_POLL_WAIT_TIME);
+
+		rc = -ETIMEDOUT;
+
+		goto out;
+	}
+
+	if ((fds[0].revents & POLLIN) == 0) {
+		printf("Poll received value %d\n", fds[0].revents);
+
+		rc = -EINVAL;
+
+		goto out;
+	}
+
+	rc = accept(server_vsock_fd, (struct sockaddr *)&client_vsock_addr,
+		    &client_vsock_len);
+	if (rc < 0) {
+		printf("Error in accept [rc=%d]\n", rc);
+
+		goto out;
+	}
+
+	/*
+	 * Read the heartbeat value that the init process in the enclave sends
+	 * after vsock connect.
+	 */
+	rc = read(server_vsock_fd, &recv_buf, sizeof(recv_buf));
+	if (rc < 0) {
+		printf("Error in read [rc=%d]\n", rc);
+
+		goto out;
+	}
+
+	if (rc != sizeof(recv_buf) ||
+	    recv_buf != NE_IMAGE_LOAD_HEARTBEAT_VALUE) {
+		printf("Read %d instead of %d\n", recv_buf,
+		       NE_IMAGE_LOAD_HEARTBEAT_VALUE);
+
+		goto out;
+	}
+
+	close(server_vsock_fd);
+
+	return 0;
+
+out:
+	close(server_vsock_fd);
+
+	return rc;
+}
+
+/* Set memory region for the given enclave. */
+static int ne_set_mem_region(int enclave_fd, struct ne_mem_region ne_mem_region)
+{
+	struct ne_user_memory_region mem_region = {};
+	int rc = 0;
+
+	if (enclave_fd < 0)
+		return -EINVAL;
+
+	mem_region.memory_size = ne_mem_region.mem_size;
+	mem_region.userspace_addr = (__u64)ne_mem_region.mem_addr;
+
+	rc = ioctl(enclave_fd, NE_SET_USER_MEMORY_REGION, &mem_region);
+	if (rc < 0) {
+		printf("Error in set user memory region [rc=%d]\n", rc);
+
+		return rc;
+	}
+
+	return 0;
+}
+
+/* Unmap all the memory regions that were set aside for the  enclave. */
+static void ne_free_mem_regions(struct ne_mem_region ne_mem_regions[])
+{
+	unsigned int i = 0;
+
+	for (i = 0; i < NE_DEFAULT_NR_MEM_REGIONS; i++)
+		munmap(ne_mem_regions[i].mem_addr, ne_mem_regions[i].mem_size);
+}
+
+/* Create enclave vCPU. */
+static int ne_create_vcpu(int enclave_fd, struct ne_vcpu *ne_vcpu)
+{
+	if (enclave_fd < 0)
+		return -EINVAL;
+
+	if (!ne_vcpu)
+		return -EINVAL;
+
+	ne_vcpu->vcpu_fd = ioctl(enclave_fd, NE_CREATE_VCPU, &ne_vcpu->vcpu_id);
+	if (ne_vcpu->vcpu_fd < 0) {
+		printf("Error in create vcpu [rc=%d]\n", ne_vcpu->vcpu_fd);
+
+		return ne_vcpu->vcpu_fd;
+	}
+
+	return 0;
+}
+
+/* Release enclave vCPU fd(s). */
+static void ne_release_vcpus(struct ne_vcpu ne_vcpus[])
+{
+	unsigned int i = 0;
+
+	for (i = 0; i < NE_DEFAULT_NR_VCPUS; i++)
+		if (ne_vcpus[i].vcpu_fd > 0)
+			close(ne_vcpus[i].vcpu_fd);
+}
+
+int main(int argc, char *argv[])
+{
+	int enclave_fd = 0;
+	char enclave_image_path[PATH_MAX] = {};
+	struct ne_enclave_start_info enclave_start_info = {};
+	unsigned int i = 0;
+	int ne_api_version = 0;
+	int ne_dev_fd = 0;
+	struct ne_mem_region ne_mem_regions[NE_DEFAULT_NR_MEM_REGIONS] = {};
+	struct ne_vcpu ne_vcpus[NE_DEFAULT_NR_VCPUS] = {};
+	int rc = 0;
+	unsigned long slot_uid = 0;
+	pthread_t thread_id = 0;
+
+	if (argc != 2) {
+		printf("Usage: %s <path_to_enclave_image>\n", argv[0]);
+
+		exit(EXIT_FAILURE);
+	}
+
+	strncpy(enclave_image_path, argv[1], sizeof(enclave_image_path) - 1);
+
+	ne_dev_fd = open(NE_DEV_NAME, O_RDWR | O_CLOEXEC);
+	if (ne_dev_fd < 0) {
+		printf("Error in open NE device [rc=%d]\n", ne_dev_fd);
+
+		exit(EXIT_FAILURE);
+	}
+
+	ne_api_version = ioctl(ne_dev_fd, NE_GET_API_VERSION);
+	if (ne_api_version != NE_EXPECTED_API_VERSION) {
+		printf("Expected API version %d, provided API version %d\n",
+		       NE_EXPECTED_API_VERSION, ne_api_version);
+
+		close(ne_dev_fd);
+
+		exit(EXIT_FAILURE);
+	}
+
+	printf("Creating enclave slot ...\n");
+
+	enclave_fd = ioctl(ne_dev_fd, NE_CREATE_VM, &slot_uid);
+
+	close(ne_dev_fd);
+
+	if (enclave_fd < 0) {
+		printf("Error in create enclave slot [rc=%d]\n", enclave_fd);
+
+		exit(EXIT_FAILURE);
+	}
+
+	printf("Enclave fd %d\n", enclave_fd);
+
+	rc = pthread_create(&thread_id, NULL, ne_poll_enclave_fd,
+			    (void *)&enclave_fd);
+	if (rc < 0) {
+		printf("Error in thread create [rc=%d]\n", rc);
+
+		close(enclave_fd);
+
+		exit(EXIT_FAILURE);
+	}
+
+	for (i = 0; i < NE_DEFAULT_NR_MEM_REGIONS; i++) {
+		ne_mem_regions[i].mem_size = NE_MIN_MEM_REGION_SIZE;
+		rc = ne_alloc_mem_region(&ne_mem_regions[i]);
+		if (rc < 0) {
+			printf("Error in alloc mem region, iter %d [rc=%d]\n",
+			       i, rc);
+
+			goto release_enclave_fd;
+		}
+	}
+
+	rc = ne_load_enclave_image(enclave_fd, ne_mem_regions,
+				   enclave_image_path);
+	if (rc < 0) {
+		printf("Error in load enclave image [rc=%d]\n", rc);
+
+		goto release_enclave_fd;
+	}
+
+	for (i = 0; i < NE_DEFAULT_NR_MEM_REGIONS; i++) {
+		rc = ne_set_mem_region(enclave_fd, ne_mem_regions[i]);
+		if (rc < 0) {
+			printf("Error in set mem region, iter %d [rc=%d]\n",
+			       i, rc);
+
+			goto release_enclave_fd;
+		}
+	}
+
+	printf("Enclave memory regions were added\n");
+
+	for (i = 0; i < NE_DEFAULT_NR_VCPUS; i++) {
+		/*
+		 * The vCPU is chosen from the enclave vCPU pool, if the value
+		 * of the vcpu_id is 0.
+		 */
+		ne_vcpus[i].vcpu_id = 0;
+		rc = ne_create_vcpu(enclave_fd, &ne_vcpus[i]);
+		if (rc < 0) {
+			printf("Error in create vcpu, iter %d [rc=%d]\n",
+			       i, rc);
+
+			goto release_enclave_vcpu_fds;
+		}
+	}
+
+	printf("Enclave vCPUs were created\n");
+
+	rc = ioctl(enclave_fd, NE_START_ENCLAVE, &enclave_start_info);
+	if (rc < 0) {
+		printf("Error in start enclave [rc=%d]\n", rc);
+
+		goto release_enclave_vcpu_fds;
+	}
+
+	printf("Enclave started, CID %llu\n", enclave_start_info.enclave_cid);
+
+	/*
+	 * TODO: Check for enclave hearbeat after it has started to see if it
+	 * has booted.
+	 */
+
+	printf("Entering sleep for %d seconds ...\n", NE_SLEEP_TIME);
+
+	sleep(NE_SLEEP_TIME);
+
+	ne_release_vcpus(ne_vcpus);
+
+	close(enclave_fd);
+
+	ne_free_mem_regions(ne_mem_regions);
+
+	exit(EXIT_SUCCESS);
+
+release_enclave_vcpu_fds:
+	ne_release_vcpus(ne_vcpus);
+release_enclave_fd:
+	close(enclave_fd);
+	ne_free_mem_regions(ne_mem_regions);
+
+	exit(EXIT_FAILURE);
+}