Message ID | 20190619222401.14942-2-sean.j.christopherson@intel.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | security: x86/sgx: SGX vs. LSM | expand |
On Wed, Jun 19, 2019 at 03:23:50PM -0700, Sean Christopherson wrote: > Using per-vma refcounting to track mm_structs associated with an enclave > requires hooking .vm_close(), which in turn prevents the mm from merging > vmas (precisely to allow refcounting). Why having sgx_vma_close() prevents that? I do not understand the problem statement. > Avoid refcounting encl_mm altogether by registering an mmu_notifier at > .mmap(), removing the dying encl_mm at mmu_notifier.release() and > protecting mm_list during reclaim via a per-enclave SRCU. Right, there is the potential collision with my changes: 1. Your patch: enclave life-cycle equals life-cycle of all processes that are associated with the enclave. 2. My (yet be sent) patch: enclave life-cycle equals the life cycle. I won't rush with my patch. I rather merge neither at this point and you can review mine after you come back from your vacation. > Removing refcounting/vm_close() allows merging of enclave vmas, at the > cost of delaying removal of encl_mm structs from mm_list, i.e. an mm is > disassociated from an enclave when the mm exits or the enclave dies, as > opposed to when the last vma (in a given mm) is closed. > > The impact of delying encl_mm removal is its memory footprint and > whatever overhead is incurred during EPC reclaim (to walk an mm's vmas). > Practically speaking, a stale encl_mm will exist for a meaningful amount > of time if and only if the enclave is mapped in a long-lived process and > then passed off to another long-lived process. It is expected that the > vast majority of use cases will not encounter this condition, e.g. even > using a daemon to build enclaves should not result in a stale encl_mm as > the builder should never need to mmap() the enclave. This paragraph speaks only about "well behaving" software. > Even if there are scenarios that lead to defunct encl_mms, the cost is > likely far outweighed by the benefits of reducing the number of vmas > across all enclaves. > > Note, using SRCU to protect mm_list is not strictly necessary, i.e. the > existing walker with encl_mm refcounting could be massaged to work with > mmu_notifier.release(), but the resulting code is subtle and fragile (I > never actually got it working). The primary issue is that an encl_mm > can't be moved off the list until its refcount goes to zero, otherwise > the custom walker goes off into the weeds. The refcount requirement > then prevents using mm_list to identify if an mmu_notifier.release() > has fired, i.e. another mechanism is needed to guard against races > between exit_mmap() and sgx_release(). Is it really impossible to send a separate SRCU patch? I fully agree with the SRCU whereas rest of this patch is still under debate. If you could do that, I can merge it in no time. It is a small step into better direction. > Cc: Dave Hansen <dave.hansen@intel.com> > Cc: Andy Lutomirski <luto@kernel.org> > Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com> Needs to be rebased because the master missing your earlier bug fix. > --- > arch/x86/Kconfig | 2 + > arch/x86/kernel/cpu/sgx/driver/ioctl.c | 14 -- > arch/x86/kernel/cpu/sgx/driver/main.c | 38 ++++ > arch/x86/kernel/cpu/sgx/encl.c | 234 +++++++++++-------------- > arch/x86/kernel/cpu/sgx/encl.h | 19 +- > arch/x86/kernel/cpu/sgx/reclaim.c | 71 +++----- > 6 files changed, 182 insertions(+), 196 deletions(-) > > diff --git a/arch/x86/Kconfig b/arch/x86/Kconfig > index a0fd17c32521..940c52762f24 100644 > --- a/arch/x86/Kconfig > +++ b/arch/x86/Kconfig > @@ -1918,6 +1918,8 @@ config X86_INTEL_MEMORY_PROTECTION_KEYS > config INTEL_SGX > bool "Intel SGX core functionality" > depends on X86_64 && CPU_SUP_INTEL > + select MMU_NOTIFIER > + select SRCU > ---help--- > Intel(R) SGX is a set of CPU instructions that can be used by > applications to set aside private regions of code and data, referred > diff --git a/arch/x86/kernel/cpu/sgx/driver/ioctl.c b/arch/x86/kernel/cpu/sgx/driver/ioctl.c > index d17c60dca114..3552d642b26f 100644 > --- a/arch/x86/kernel/cpu/sgx/driver/ioctl.c > +++ b/arch/x86/kernel/cpu/sgx/driver/ioctl.c > @@ -276,7 +276,6 @@ static int sgx_encl_create(struct sgx_encl *encl, struct sgx_secs *secs) > { > unsigned long encl_size = secs->size + PAGE_SIZE; > struct sgx_epc_page *secs_epc; > - struct sgx_encl_mm *encl_mm; > unsigned long ssaframesize; > struct sgx_pageinfo pginfo; > struct sgx_secinfo secinfo; > @@ -311,12 +310,6 @@ static int sgx_encl_create(struct sgx_encl *encl, struct sgx_secs *secs) > > INIT_WORK(&encl->work, sgx_add_page_worker); > > - encl_mm = sgx_encl_mm_add(encl, current->mm); > - if (IS_ERR(encl_mm)) { > - ret = PTR_ERR(encl_mm); > - goto err_out; > - } > - > secs_epc = sgx_alloc_page(&encl->secs, true); > if (IS_ERR(secs_epc)) { > ret = PTR_ERR(secs_epc); > @@ -369,13 +362,6 @@ static int sgx_encl_create(struct sgx_encl *encl, struct sgx_secs *secs) > encl->backing = NULL; > } > > - if (!list_empty(&encl->mm_list)) { > - encl_mm = list_first_entry(&encl->mm_list, struct sgx_encl_mm, > - list); > - list_del(&encl_mm->list); > - kfree(encl_mm); > - } > - > mutex_unlock(&encl->lock); > return ret; > } > diff --git a/arch/x86/kernel/cpu/sgx/driver/main.c b/arch/x86/kernel/cpu/sgx/driver/main.c > index 0c831ee5e2de..07aa5f91b2dd 100644 > --- a/arch/x86/kernel/cpu/sgx/driver/main.c > +++ b/arch/x86/kernel/cpu/sgx/driver/main.c > @@ -25,6 +25,7 @@ u32 sgx_xsave_size_tbl[64]; > static int sgx_open(struct inode *inode, struct file *file) > { > struct sgx_encl *encl; > + int ret; > > encl = kzalloc(sizeof(*encl), GFP_KERNEL); > if (!encl) > @@ -38,6 +39,12 @@ static int sgx_open(struct inode *inode, struct file *file) > INIT_LIST_HEAD(&encl->mm_list); > spin_lock_init(&encl->mm_lock); > > + ret = init_srcu_struct(&encl->srcu); > + if (ret) { > + kfree(encl); > + return ret; > + } > + > file->private_data = encl; > > return 0; > @@ -46,6 +53,32 @@ static int sgx_open(struct inode *inode, struct file *file) > static int sgx_release(struct inode *inode, struct file *file) > { > struct sgx_encl *encl = file->private_data; > + struct sgx_encl_mm *encl_mm; > + > + /* > + * Objects can't be *moved* off an RCU protected list (deletion is ok), > + * nor can the object be freed until after synchronize_srcu(). > + */ > +restart: > + spin_lock(&encl->mm_lock); > + if (list_empty(&encl->mm_list)) { > + encl_mm = NULL; > + } else { > + encl_mm = list_first_entry(&encl->mm_list, struct sgx_encl_mm, > + list); > + list_del_rcu(&encl_mm->list); > + } > + spin_unlock(&encl->mm_lock); > + > + if (encl_mm) { > + synchronize_srcu(&encl->srcu); > + > + mmu_notifier_unregister(&encl_mm->mmu_notifier, encl_mm->mm); > + > + sgx_encl_mm_release(encl_mm); > + > + goto restart; > + } > > kref_put(&encl->refcount, sgx_encl_release); > > @@ -63,6 +96,11 @@ static long sgx_compat_ioctl(struct file *filep, unsigned int cmd, > static int sgx_mmap(struct file *file, struct vm_area_struct *vma) > { > struct sgx_encl *encl = file->private_data; > + int ret; > + > + ret = sgx_encl_mm_add(encl, vma->vm_mm); > + if (ret) > + return ret; > > vma->vm_ops = &sgx_vm_ops; > vma->vm_flags |= VM_PFNMAP | VM_DONTEXPAND | VM_DONTDUMP | VM_IO; > diff --git a/arch/x86/kernel/cpu/sgx/encl.c b/arch/x86/kernel/cpu/sgx/encl.c > index 9566eb72d417..c6436bbd4a68 100644 > --- a/arch/x86/kernel/cpu/sgx/encl.c > +++ b/arch/x86/kernel/cpu/sgx/encl.c > @@ -132,103 +132,125 @@ static struct sgx_encl_page *sgx_encl_load_page(struct sgx_encl *encl, > return entry; > } > > -struct sgx_encl_mm *sgx_encl_mm_add(struct sgx_encl *encl, > - struct mm_struct *mm) > +static void sgx_encl_mm_release_wq(struct work_struct *work) > +{ > + struct sgx_encl_mm *encl_mm = > + container_of(work, struct sgx_encl_mm, release_work); > + > + sgx_encl_mm_release(encl_mm); > +} > + > +/* > + * Being a call_srcu() callback, this needs to be short, and sgx_encl_release() > + * is anything but short. Do the final freeing in yet another async callback. > + */ > +static void sgx_encl_mm_release_delayed(struct rcu_head *rcu) Would rename this either as *_tail() or *_deferred(). > +{ > + struct sgx_encl_mm *encl_mm = > + container_of(rcu, struct sgx_encl_mm, rcu); > + > + INIT_WORK(&encl_mm->release_work, sgx_encl_mm_release_wq); > + schedule_work(&encl_mm->release_work); > +} > + > +static void sgx_mmu_notifier_release(struct mmu_notifier *mn, > + struct mm_struct *mm) > +{ > + struct sgx_encl_mm *encl_mm = > + container_of(mn, struct sgx_encl_mm, mmu_notifier); > + struct sgx_encl_mm *tmp = NULL; > + > + /* > + * The enclave itself can remove encl_mm. Note, objects can't be moved > + * off an RCU protected list, but deletion is ok. > + */ > + spin_lock(&encl_mm->encl->mm_lock); > + list_for_each_entry(tmp, &encl_mm->encl->mm_list, list) { > + if (tmp == encl_mm) { > + list_del_rcu(&encl_mm->list); > + break; > + } > + } > + spin_unlock(&encl_mm->encl->mm_lock); > + > + if (tmp == encl_mm) { > + synchronize_srcu(&encl_mm->encl->srcu); > + > + /* > + * Delay freeing encl_mm until after mmu_notifier releases any > + * SRCU locks. synchronize_srcu() must be called from process > + * context, i.e. we can't throw mmu_notifier_unregister() in a > + * work queue and be done with it. > + */ > + mmu_notifier_unregister_no_release(mn, mm); > + mmu_notifier_call_srcu(&encl_mm->rcu, > + &sgx_encl_mm_release_delayed); > + } > +} > + > +static const struct mmu_notifier_ops sgx_mmu_notifier_ops = { > + .release = sgx_mmu_notifier_release, > +}; > + > +static struct sgx_encl_mm *sgx_encl_find_mm(struct sgx_encl *encl, > + struct mm_struct *mm) > +{ > + struct sgx_encl_mm *encl_mm = NULL; > + struct sgx_encl_mm *tmp; > + int idx; > + > + idx = srcu_read_lock(&encl->srcu); > + > + list_for_each_entry_rcu(tmp, &encl->mm_list, list) { > + if (tmp->mm == mm) { > + encl_mm = tmp; > + break; > + } > + } > + > + srcu_read_unlock(&encl->srcu, idx); > + > + return encl_mm; > +} > + > +int sgx_encl_mm_add(struct sgx_encl *encl, struct mm_struct *mm) > { > struct sgx_encl_mm *encl_mm; > + int ret; > + > + lockdep_assert_held_exclusive(&mm->mmap_sem); > + > + /* > + * mm_structs are kept on mm_list until the mm or the enclave dies, > + * i.e. once an mm is off the list, it's gone for good, therefore it's > + * impossible to get a false positive on @mm due to a stale mm_list. > + */ > + if (sgx_encl_find_mm(encl, mm)) > + return 0; > > encl_mm = kzalloc(sizeof(*encl_mm), GFP_KERNEL); > if (!encl_mm) > - return ERR_PTR(-ENOMEM); > + return -ENOMEM; > > encl_mm->encl = encl; > encl_mm->mm = mm; > - kref_init(&encl_mm->refcount); > + encl_mm->mmu_notifier.ops = &sgx_mmu_notifier_ops; > + > + ret = __mmu_notifier_register(&encl_mm->mmu_notifier, mm); > + if (ret) { > + kfree(encl_mm); > + return ret; > + } > + > + kref_get(&encl->refcount); > > spin_lock(&encl->mm_lock); > - list_add(&encl_mm->list, &encl->mm_list); > + list_add_rcu(&encl_mm->list, &encl->mm_list); > spin_unlock(&encl->mm_lock); > > - return encl_mm; > -} > + synchronize_srcu(&encl->srcu); > > -void sgx_encl_mm_release(struct kref *ref) > -{ > - struct sgx_encl_mm *encl_mm = > - container_of(ref, struct sgx_encl_mm, refcount); > - > - spin_lock(&encl_mm->encl->mm_lock); > - list_del(&encl_mm->list); > - spin_unlock(&encl_mm->encl->mm_lock); > - > - kfree(encl_mm); > -} > - > -static struct sgx_encl_mm *sgx_encl_get_mm(struct sgx_encl *encl, > - struct mm_struct *mm) > -{ > - struct sgx_encl_mm *encl_mm = NULL; > - struct sgx_encl_mm *prev_mm = NULL; > - int iter; > - > - while (true) { > - encl_mm = sgx_encl_next_mm(encl, prev_mm, &iter); > - if (prev_mm) > - kref_put(&prev_mm->refcount, sgx_encl_mm_release); > - prev_mm = encl_mm; > - > - if (iter == SGX_ENCL_MM_ITER_DONE) > - break; > - > - if (iter == SGX_ENCL_MM_ITER_RESTART) > - continue; > - > - if (mm == encl_mm->mm) > - return encl_mm; > - } > - > - return NULL; > -} > - > -static void sgx_vma_open(struct vm_area_struct *vma) > -{ > - struct sgx_encl *encl = vma->vm_private_data; > - struct sgx_encl_mm *encl_mm; > - > - if (!encl) > - return; > - > - if (encl->flags & SGX_ENCL_DEAD) > - goto error; > - > - encl_mm = sgx_encl_get_mm(encl, vma->vm_mm); > - if (!encl_mm) { > - encl_mm = sgx_encl_mm_add(encl, vma->vm_mm); > - if (IS_ERR(encl_mm)) > - goto error; > - } > - > - return; > - > -error: > - vma->vm_private_data = NULL; > -} > - > -static void sgx_vma_close(struct vm_area_struct *vma) > -{ > - struct sgx_encl *encl = vma->vm_private_data; > - struct sgx_encl_mm *encl_mm; > - > - if (!encl) > - return; > - > - encl_mm = sgx_encl_get_mm(encl, vma->vm_mm); > - if (encl_mm) { > - kref_put(&encl_mm->refcount, sgx_encl_mm_release); > - > - /* Release kref for the VMA. */ > - kref_put(&encl_mm->refcount, sgx_encl_mm_release); > - } > + return 0; > } > > static unsigned int sgx_vma_fault(struct vm_fault *vmf) > @@ -366,8 +388,6 @@ static int sgx_vma_access(struct vm_area_struct *vma, unsigned long addr, > } > > const struct vm_operations_struct sgx_vm_ops = { > - .close = sgx_vma_close, > - .open = sgx_vma_open, > .fault = sgx_vma_fault, > .access = sgx_vma_access, > }; > @@ -465,7 +485,7 @@ void sgx_encl_release(struct kref *ref) > if (encl->backing) > fput(encl->backing); > > - WARN(!list_empty(&encl->mm_list), "sgx: mm_list non-empty"); > + WARN_ONCE(!list_empty(&encl->mm_list), "sgx: mm_list non-empty"); > > kfree(encl); > } > @@ -503,46 +523,6 @@ struct page *sgx_encl_get_backing_page(struct sgx_encl *encl, pgoff_t index) > return shmem_read_mapping_page_gfp(mapping, index, gfpmask); > } > > -/** > - * sgx_encl_next_mm() - Iterate to the next mm > - * @encl: an enclave > - * @mm: an mm list entry > - * @iter: iterator status > - * > - * Return: the enclave mm or NULL > - */ > -struct sgx_encl_mm *sgx_encl_next_mm(struct sgx_encl *encl, > - struct sgx_encl_mm *encl_mm, int *iter) > -{ > - struct list_head *entry; > - > - WARN(!encl, "%s: encl is NULL", __func__); > - WARN(!iter, "%s: iter is NULL", __func__); > - > - spin_lock(&encl->mm_lock); > - > - entry = encl_mm ? encl_mm->list.next : encl->mm_list.next; > - WARN(!entry, "%s: entry is NULL", __func__); > - > - if (entry == &encl->mm_list) { > - spin_unlock(&encl->mm_lock); > - *iter = SGX_ENCL_MM_ITER_DONE; > - return NULL; > - } > - > - encl_mm = list_entry(entry, struct sgx_encl_mm, list); > - > - if (!kref_get_unless_zero(&encl_mm->refcount)) { > - spin_unlock(&encl->mm_lock); > - *iter = SGX_ENCL_MM_ITER_RESTART; > - return NULL; > - } > - > - spin_unlock(&encl->mm_lock); > - *iter = SGX_ENCL_MM_ITER_NEXT; > - return encl_mm; > -} > - > static int sgx_encl_test_and_clear_young_cb(pte_t *ptep, pgtable_t token, > unsigned long addr, void *data) > { > diff --git a/arch/x86/kernel/cpu/sgx/encl.h b/arch/x86/kernel/cpu/sgx/encl.h > index c557f0374d74..0904b3c20ed0 100644 > --- a/arch/x86/kernel/cpu/sgx/encl.h > +++ b/arch/x86/kernel/cpu/sgx/encl.h > @@ -9,9 +9,11 @@ > #include <linux/kref.h> > #include <linux/list.h> > #include <linux/mm_types.h> > +#include <linux/mmu_notifier.h> > #include <linux/mutex.h> > #include <linux/notifier.h> > #include <linux/radix-tree.h> > +#include <linux/srcu.h> > #include <linux/workqueue.h> > > /** > @@ -57,8 +59,10 @@ enum sgx_encl_flags { > struct sgx_encl_mm { > struct sgx_encl *encl; > struct mm_struct *mm; > - struct kref refcount; > struct list_head list; > + struct mmu_notifier mmu_notifier; > + struct work_struct release_work; > + struct rcu_head rcu; > }; > > struct sgx_encl { > @@ -72,6 +76,7 @@ struct sgx_encl { > spinlock_t mm_lock; > struct file *backing; > struct kref refcount; > + struct srcu_struct srcu; > unsigned long base; > unsigned long size; > unsigned long ssaframesize; > @@ -118,11 +123,13 @@ void sgx_encl_destroy(struct sgx_encl *encl); > void sgx_encl_release(struct kref *ref); > pgoff_t sgx_encl_get_index(struct sgx_encl *encl, struct sgx_encl_page *page); > struct page *sgx_encl_get_backing_page(struct sgx_encl *encl, pgoff_t index); > -struct sgx_encl_mm *sgx_encl_next_mm(struct sgx_encl *encl, > - struct sgx_encl_mm *encl_mm, int *iter); > -struct sgx_encl_mm *sgx_encl_mm_add(struct sgx_encl *encl, > - struct mm_struct *mm); > -void sgx_encl_mm_release(struct kref *ref); > +int sgx_encl_mm_add(struct sgx_encl *encl, struct mm_struct *mm); > +static inline void sgx_encl_mm_release(struct sgx_encl_mm *encl_mm) > +{ > + kref_put(&encl_mm->encl->refcount, sgx_encl_release); > + > + kfree(encl_mm); > +} Please just open code this to the two call sites. Makes the code hard to follow. Right now I did not find anything else questionable from the code changes. Repeating myself but if it is by any means possible before going away, can you construct a pure SRCU patch. I could then reconstruct my changes on top off that, which would make evalution of both heck a lot easier. /Jarkko
On Fri, Jun 21, 2019 at 12:03:36AM +0300, Jarkko Sakkinen wrote: > On Wed, Jun 19, 2019 at 03:23:50PM -0700, Sean Christopherson wrote: > > Using per-vma refcounting to track mm_structs associated with an enclave > > requires hooking .vm_close(), which in turn prevents the mm from merging > > vmas (precisely to allow refcounting). > > Why having sgx_vma_close() prevents that? I do not understand the > problem statement. vmas that define .vm_close() cannot be merged. /* * If the vma has a ->close operation then the driver probably needs to release * per-vma resources, so we don't attempt to merge those. */ static inline int is_mergeable_vma(struct vm_area_struct *vma, struct file *file, unsigned long vm_flags, struct vm_userfaultfd_ctx vm_userfaultfd_ctx) { ... if (vma->vm_ops && vma->vm_ops->close) return 0; if (!is_mergeable_vm_userfaultfd_ctx(vma, vm_userfaultfd_ctx)) return 0; return 1; } > > > Avoid refcounting encl_mm altogether by registering an mmu_notifier at > > .mmap(), removing the dying encl_mm at mmu_notifier.release() and > > protecting mm_list during reclaim via a per-enclave SRCU. > > Right, there is the potential collision with my changes: > > 1. Your patch: enclave life-cycle equals life-cycle of all processes > that are associated with the enclave. > 2. My (yet be sent) patch: enclave life-cycle equals the life cycle. > > I won't rush with my patch. I rather merge neither at this point and > you can review mine after you come back from your vacation. > > > Removing refcounting/vm_close() allows merging of enclave vmas, at the > > cost of delaying removal of encl_mm structs from mm_list, i.e. an mm is > > disassociated from an enclave when the mm exits or the enclave dies, as > > opposed to when the last vma (in a given mm) is closed. > > > > The impact of delying encl_mm removal is its memory footprint and > > whatever overhead is incurred during EPC reclaim (to walk an mm's vmas). > > Practically speaking, a stale encl_mm will exist for a meaningful amount > > of time if and only if the enclave is mapped in a long-lived process and > > then passed off to another long-lived process. It is expected that the > > vast majority of use cases will not encounter this condition, e.g. even > > using a daemon to build enclaves should not result in a stale encl_mm as > > the builder should never need to mmap() the enclave. > > This paragraph speaks only about "well behaving" software. Malicious software isn't all that interesting as there are far easier ways to waste system resources. That being said, the encl_mm allocation can use GFP_KERNEL_ACCOUNT. > > Even if there are scenarios that lead to defunct encl_mms, the cost is > > likely far outweighed by the benefits of reducing the number of vmas > > across all enclaves. > > > > Note, using SRCU to protect mm_list is not strictly necessary, i.e. the > > existing walker with encl_mm refcounting could be massaged to work with > > mmu_notifier.release(), but the resulting code is subtle and fragile (I > > never actually got it working). The primary issue is that an encl_mm > > can't be moved off the list until its refcount goes to zero, otherwise > > the custom walker goes off into the weeds. The refcount requirement > > then prevents using mm_list to identify if an mmu_notifier.release() > > has fired, i.e. another mechanism is needed to guard against races > > between exit_mmap() and sgx_release(). > > Is it really impossible to send a separate SRCU patch? I can split out the SRCU as a precursor. It'll likely take me a few days to get it sent. > I fully agree with the SRCU whereas rest of this patch is still > under debate. > > If you could do that, I can merge it in no time. It is a small > step into better direction. > > > Cc: Dave Hansen <dave.hansen@intel.com> > > Cc: Andy Lutomirski <luto@kernel.org> > > Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com> > > Needs to be rebased because the master missing your earlier bug fix. ... > > diff --git a/arch/x86/kernel/cpu/sgx/encl.c b/arch/x86/kernel/cpu/sgx/encl.c > > index 9566eb72d417..c6436bbd4a68 100644 > > --- a/arch/x86/kernel/cpu/sgx/encl.c > > +++ b/arch/x86/kernel/cpu/sgx/encl.c > > @@ -132,103 +132,125 @@ static struct sgx_encl_page *sgx_encl_load_page(struct sgx_encl *encl, > > return entry; > > } > > > > -struct sgx_encl_mm *sgx_encl_mm_add(struct sgx_encl *encl, > > - struct mm_struct *mm) > > +static void sgx_encl_mm_release_wq(struct work_struct *work) > > +{ > > + struct sgx_encl_mm *encl_mm = > > + container_of(work, struct sgx_encl_mm, release_work); > > + > > + sgx_encl_mm_release(encl_mm); > > +} > > + > > +/* > > + * Being a call_srcu() callback, this needs to be short, and sgx_encl_release() > > + * is anything but short. Do the final freeing in yet another async callback. > > + */ > > +static void sgx_encl_mm_release_delayed(struct rcu_head *rcu) > > Would rename this either as *_tail() or *_deferred(). Deferred works for me. > > +{ > > + struct sgx_encl_mm *encl_mm = > > + container_of(rcu, struct sgx_encl_mm, rcu); > > + > > + INIT_WORK(&encl_mm->release_work, sgx_encl_mm_release_wq); > > + schedule_work(&encl_mm->release_work); > > +} > > + ... > > @@ -118,11 +123,13 @@ void sgx_encl_destroy(struct sgx_encl *encl); > > void sgx_encl_release(struct kref *ref); > > pgoff_t sgx_encl_get_index(struct sgx_encl *encl, struct sgx_encl_page *page); > > struct page *sgx_encl_get_backing_page(struct sgx_encl *encl, pgoff_t index); > > -struct sgx_encl_mm *sgx_encl_next_mm(struct sgx_encl *encl, > > - struct sgx_encl_mm *encl_mm, int *iter); > > -struct sgx_encl_mm *sgx_encl_mm_add(struct sgx_encl *encl, > > - struct mm_struct *mm); > > -void sgx_encl_mm_release(struct kref *ref); > > +int sgx_encl_mm_add(struct sgx_encl *encl, struct mm_struct *mm); > > +static inline void sgx_encl_mm_release(struct sgx_encl_mm *encl_mm) > > +{ > > + kref_put(&encl_mm->encl->refcount, sgx_encl_release); > > + > > + kfree(encl_mm); > > +} > > Please just open code this to the two call sites. Makes the code hard to > follow. Heh, I waffled between a helper and open coding. I chose poorly :-) > Right now I did not find anything else questionable from the code > changes. Repeating myself but if it is by any means possible before > going away, can you construct a pure SRCU patch. > > I could then reconstruct my changes on top off that, which would > make evalution of both heck a lot easier. > > /Jarkko
On Mon, Jul 08, 2019 at 07:57:07AM -0700, Sean Christopherson wrote: > On Fri, Jun 21, 2019 at 12:03:36AM +0300, Jarkko Sakkinen wrote: > > On Wed, Jun 19, 2019 at 03:23:50PM -0700, Sean Christopherson wrote: > > > Using per-vma refcounting to track mm_structs associated with an enclave > > > requires hooking .vm_close(), which in turn prevents the mm from merging > > > vmas (precisely to allow refcounting). > > > > Why having sgx_vma_close() prevents that? I do not understand the > > problem statement. > > vmas that define .vm_close() cannot be merged. Ugh, did not know that :-) Thank you. /Jarkko
diff --git a/arch/x86/Kconfig b/arch/x86/Kconfig index a0fd17c32521..940c52762f24 100644 --- a/arch/x86/Kconfig +++ b/arch/x86/Kconfig @@ -1918,6 +1918,8 @@ config X86_INTEL_MEMORY_PROTECTION_KEYS config INTEL_SGX bool "Intel SGX core functionality" depends on X86_64 && CPU_SUP_INTEL + select MMU_NOTIFIER + select SRCU ---help--- Intel(R) SGX is a set of CPU instructions that can be used by applications to set aside private regions of code and data, referred diff --git a/arch/x86/kernel/cpu/sgx/driver/ioctl.c b/arch/x86/kernel/cpu/sgx/driver/ioctl.c index d17c60dca114..3552d642b26f 100644 --- a/arch/x86/kernel/cpu/sgx/driver/ioctl.c +++ b/arch/x86/kernel/cpu/sgx/driver/ioctl.c @@ -276,7 +276,6 @@ static int sgx_encl_create(struct sgx_encl *encl, struct sgx_secs *secs) { unsigned long encl_size = secs->size + PAGE_SIZE; struct sgx_epc_page *secs_epc; - struct sgx_encl_mm *encl_mm; unsigned long ssaframesize; struct sgx_pageinfo pginfo; struct sgx_secinfo secinfo; @@ -311,12 +310,6 @@ static int sgx_encl_create(struct sgx_encl *encl, struct sgx_secs *secs) INIT_WORK(&encl->work, sgx_add_page_worker); - encl_mm = sgx_encl_mm_add(encl, current->mm); - if (IS_ERR(encl_mm)) { - ret = PTR_ERR(encl_mm); - goto err_out; - } - secs_epc = sgx_alloc_page(&encl->secs, true); if (IS_ERR(secs_epc)) { ret = PTR_ERR(secs_epc); @@ -369,13 +362,6 @@ static int sgx_encl_create(struct sgx_encl *encl, struct sgx_secs *secs) encl->backing = NULL; } - if (!list_empty(&encl->mm_list)) { - encl_mm = list_first_entry(&encl->mm_list, struct sgx_encl_mm, - list); - list_del(&encl_mm->list); - kfree(encl_mm); - } - mutex_unlock(&encl->lock); return ret; } diff --git a/arch/x86/kernel/cpu/sgx/driver/main.c b/arch/x86/kernel/cpu/sgx/driver/main.c index 0c831ee5e2de..07aa5f91b2dd 100644 --- a/arch/x86/kernel/cpu/sgx/driver/main.c +++ b/arch/x86/kernel/cpu/sgx/driver/main.c @@ -25,6 +25,7 @@ u32 sgx_xsave_size_tbl[64]; static int sgx_open(struct inode *inode, struct file *file) { struct sgx_encl *encl; + int ret; encl = kzalloc(sizeof(*encl), GFP_KERNEL); if (!encl) @@ -38,6 +39,12 @@ static int sgx_open(struct inode *inode, struct file *file) INIT_LIST_HEAD(&encl->mm_list); spin_lock_init(&encl->mm_lock); + ret = init_srcu_struct(&encl->srcu); + if (ret) { + kfree(encl); + return ret; + } + file->private_data = encl; return 0; @@ -46,6 +53,32 @@ static int sgx_open(struct inode *inode, struct file *file) static int sgx_release(struct inode *inode, struct file *file) { struct sgx_encl *encl = file->private_data; + struct sgx_encl_mm *encl_mm; + + /* + * Objects can't be *moved* off an RCU protected list (deletion is ok), + * nor can the object be freed until after synchronize_srcu(). + */ +restart: + spin_lock(&encl->mm_lock); + if (list_empty(&encl->mm_list)) { + encl_mm = NULL; + } else { + encl_mm = list_first_entry(&encl->mm_list, struct sgx_encl_mm, + list); + list_del_rcu(&encl_mm->list); + } + spin_unlock(&encl->mm_lock); + + if (encl_mm) { + synchronize_srcu(&encl->srcu); + + mmu_notifier_unregister(&encl_mm->mmu_notifier, encl_mm->mm); + + sgx_encl_mm_release(encl_mm); + + goto restart; + } kref_put(&encl->refcount, sgx_encl_release); @@ -63,6 +96,11 @@ static long sgx_compat_ioctl(struct file *filep, unsigned int cmd, static int sgx_mmap(struct file *file, struct vm_area_struct *vma) { struct sgx_encl *encl = file->private_data; + int ret; + + ret = sgx_encl_mm_add(encl, vma->vm_mm); + if (ret) + return ret; vma->vm_ops = &sgx_vm_ops; vma->vm_flags |= VM_PFNMAP | VM_DONTEXPAND | VM_DONTDUMP | VM_IO; diff --git a/arch/x86/kernel/cpu/sgx/encl.c b/arch/x86/kernel/cpu/sgx/encl.c index 9566eb72d417..c6436bbd4a68 100644 --- a/arch/x86/kernel/cpu/sgx/encl.c +++ b/arch/x86/kernel/cpu/sgx/encl.c @@ -132,103 +132,125 @@ static struct sgx_encl_page *sgx_encl_load_page(struct sgx_encl *encl, return entry; } -struct sgx_encl_mm *sgx_encl_mm_add(struct sgx_encl *encl, - struct mm_struct *mm) +static void sgx_encl_mm_release_wq(struct work_struct *work) +{ + struct sgx_encl_mm *encl_mm = + container_of(work, struct sgx_encl_mm, release_work); + + sgx_encl_mm_release(encl_mm); +} + +/* + * Being a call_srcu() callback, this needs to be short, and sgx_encl_release() + * is anything but short. Do the final freeing in yet another async callback. + */ +static void sgx_encl_mm_release_delayed(struct rcu_head *rcu) +{ + struct sgx_encl_mm *encl_mm = + container_of(rcu, struct sgx_encl_mm, rcu); + + INIT_WORK(&encl_mm->release_work, sgx_encl_mm_release_wq); + schedule_work(&encl_mm->release_work); +} + +static void sgx_mmu_notifier_release(struct mmu_notifier *mn, + struct mm_struct *mm) +{ + struct sgx_encl_mm *encl_mm = + container_of(mn, struct sgx_encl_mm, mmu_notifier); + struct sgx_encl_mm *tmp = NULL; + + /* + * The enclave itself can remove encl_mm. Note, objects can't be moved + * off an RCU protected list, but deletion is ok. + */ + spin_lock(&encl_mm->encl->mm_lock); + list_for_each_entry(tmp, &encl_mm->encl->mm_list, list) { + if (tmp == encl_mm) { + list_del_rcu(&encl_mm->list); + break; + } + } + spin_unlock(&encl_mm->encl->mm_lock); + + if (tmp == encl_mm) { + synchronize_srcu(&encl_mm->encl->srcu); + + /* + * Delay freeing encl_mm until after mmu_notifier releases any + * SRCU locks. synchronize_srcu() must be called from process + * context, i.e. we can't throw mmu_notifier_unregister() in a + * work queue and be done with it. + */ + mmu_notifier_unregister_no_release(mn, mm); + mmu_notifier_call_srcu(&encl_mm->rcu, + &sgx_encl_mm_release_delayed); + } +} + +static const struct mmu_notifier_ops sgx_mmu_notifier_ops = { + .release = sgx_mmu_notifier_release, +}; + +static struct sgx_encl_mm *sgx_encl_find_mm(struct sgx_encl *encl, + struct mm_struct *mm) +{ + struct sgx_encl_mm *encl_mm = NULL; + struct sgx_encl_mm *tmp; + int idx; + + idx = srcu_read_lock(&encl->srcu); + + list_for_each_entry_rcu(tmp, &encl->mm_list, list) { + if (tmp->mm == mm) { + encl_mm = tmp; + break; + } + } + + srcu_read_unlock(&encl->srcu, idx); + + return encl_mm; +} + +int sgx_encl_mm_add(struct sgx_encl *encl, struct mm_struct *mm) { struct sgx_encl_mm *encl_mm; + int ret; + + lockdep_assert_held_exclusive(&mm->mmap_sem); + + /* + * mm_structs are kept on mm_list until the mm or the enclave dies, + * i.e. once an mm is off the list, it's gone for good, therefore it's + * impossible to get a false positive on @mm due to a stale mm_list. + */ + if (sgx_encl_find_mm(encl, mm)) + return 0; encl_mm = kzalloc(sizeof(*encl_mm), GFP_KERNEL); if (!encl_mm) - return ERR_PTR(-ENOMEM); + return -ENOMEM; encl_mm->encl = encl; encl_mm->mm = mm; - kref_init(&encl_mm->refcount); + encl_mm->mmu_notifier.ops = &sgx_mmu_notifier_ops; + + ret = __mmu_notifier_register(&encl_mm->mmu_notifier, mm); + if (ret) { + kfree(encl_mm); + return ret; + } + + kref_get(&encl->refcount); spin_lock(&encl->mm_lock); - list_add(&encl_mm->list, &encl->mm_list); + list_add_rcu(&encl_mm->list, &encl->mm_list); spin_unlock(&encl->mm_lock); - return encl_mm; -} + synchronize_srcu(&encl->srcu); -void sgx_encl_mm_release(struct kref *ref) -{ - struct sgx_encl_mm *encl_mm = - container_of(ref, struct sgx_encl_mm, refcount); - - spin_lock(&encl_mm->encl->mm_lock); - list_del(&encl_mm->list); - spin_unlock(&encl_mm->encl->mm_lock); - - kfree(encl_mm); -} - -static struct sgx_encl_mm *sgx_encl_get_mm(struct sgx_encl *encl, - struct mm_struct *mm) -{ - struct sgx_encl_mm *encl_mm = NULL; - struct sgx_encl_mm *prev_mm = NULL; - int iter; - - while (true) { - encl_mm = sgx_encl_next_mm(encl, prev_mm, &iter); - if (prev_mm) - kref_put(&prev_mm->refcount, sgx_encl_mm_release); - prev_mm = encl_mm; - - if (iter == SGX_ENCL_MM_ITER_DONE) - break; - - if (iter == SGX_ENCL_MM_ITER_RESTART) - continue; - - if (mm == encl_mm->mm) - return encl_mm; - } - - return NULL; -} - -static void sgx_vma_open(struct vm_area_struct *vma) -{ - struct sgx_encl *encl = vma->vm_private_data; - struct sgx_encl_mm *encl_mm; - - if (!encl) - return; - - if (encl->flags & SGX_ENCL_DEAD) - goto error; - - encl_mm = sgx_encl_get_mm(encl, vma->vm_mm); - if (!encl_mm) { - encl_mm = sgx_encl_mm_add(encl, vma->vm_mm); - if (IS_ERR(encl_mm)) - goto error; - } - - return; - -error: - vma->vm_private_data = NULL; -} - -static void sgx_vma_close(struct vm_area_struct *vma) -{ - struct sgx_encl *encl = vma->vm_private_data; - struct sgx_encl_mm *encl_mm; - - if (!encl) - return; - - encl_mm = sgx_encl_get_mm(encl, vma->vm_mm); - if (encl_mm) { - kref_put(&encl_mm->refcount, sgx_encl_mm_release); - - /* Release kref for the VMA. */ - kref_put(&encl_mm->refcount, sgx_encl_mm_release); - } + return 0; } static unsigned int sgx_vma_fault(struct vm_fault *vmf) @@ -366,8 +388,6 @@ static int sgx_vma_access(struct vm_area_struct *vma, unsigned long addr, } const struct vm_operations_struct sgx_vm_ops = { - .close = sgx_vma_close, - .open = sgx_vma_open, .fault = sgx_vma_fault, .access = sgx_vma_access, }; @@ -465,7 +485,7 @@ void sgx_encl_release(struct kref *ref) if (encl->backing) fput(encl->backing); - WARN(!list_empty(&encl->mm_list), "sgx: mm_list non-empty"); + WARN_ONCE(!list_empty(&encl->mm_list), "sgx: mm_list non-empty"); kfree(encl); } @@ -503,46 +523,6 @@ struct page *sgx_encl_get_backing_page(struct sgx_encl *encl, pgoff_t index) return shmem_read_mapping_page_gfp(mapping, index, gfpmask); } -/** - * sgx_encl_next_mm() - Iterate to the next mm - * @encl: an enclave - * @mm: an mm list entry - * @iter: iterator status - * - * Return: the enclave mm or NULL - */ -struct sgx_encl_mm *sgx_encl_next_mm(struct sgx_encl *encl, - struct sgx_encl_mm *encl_mm, int *iter) -{ - struct list_head *entry; - - WARN(!encl, "%s: encl is NULL", __func__); - WARN(!iter, "%s: iter is NULL", __func__); - - spin_lock(&encl->mm_lock); - - entry = encl_mm ? encl_mm->list.next : encl->mm_list.next; - WARN(!entry, "%s: entry is NULL", __func__); - - if (entry == &encl->mm_list) { - spin_unlock(&encl->mm_lock); - *iter = SGX_ENCL_MM_ITER_DONE; - return NULL; - } - - encl_mm = list_entry(entry, struct sgx_encl_mm, list); - - if (!kref_get_unless_zero(&encl_mm->refcount)) { - spin_unlock(&encl->mm_lock); - *iter = SGX_ENCL_MM_ITER_RESTART; - return NULL; - } - - spin_unlock(&encl->mm_lock); - *iter = SGX_ENCL_MM_ITER_NEXT; - return encl_mm; -} - static int sgx_encl_test_and_clear_young_cb(pte_t *ptep, pgtable_t token, unsigned long addr, void *data) { diff --git a/arch/x86/kernel/cpu/sgx/encl.h b/arch/x86/kernel/cpu/sgx/encl.h index c557f0374d74..0904b3c20ed0 100644 --- a/arch/x86/kernel/cpu/sgx/encl.h +++ b/arch/x86/kernel/cpu/sgx/encl.h @@ -9,9 +9,11 @@ #include <linux/kref.h> #include <linux/list.h> #include <linux/mm_types.h> +#include <linux/mmu_notifier.h> #include <linux/mutex.h> #include <linux/notifier.h> #include <linux/radix-tree.h> +#include <linux/srcu.h> #include <linux/workqueue.h> /** @@ -57,8 +59,10 @@ enum sgx_encl_flags { struct sgx_encl_mm { struct sgx_encl *encl; struct mm_struct *mm; - struct kref refcount; struct list_head list; + struct mmu_notifier mmu_notifier; + struct work_struct release_work; + struct rcu_head rcu; }; struct sgx_encl { @@ -72,6 +76,7 @@ struct sgx_encl { spinlock_t mm_lock; struct file *backing; struct kref refcount; + struct srcu_struct srcu; unsigned long base; unsigned long size; unsigned long ssaframesize; @@ -118,11 +123,13 @@ void sgx_encl_destroy(struct sgx_encl *encl); void sgx_encl_release(struct kref *ref); pgoff_t sgx_encl_get_index(struct sgx_encl *encl, struct sgx_encl_page *page); struct page *sgx_encl_get_backing_page(struct sgx_encl *encl, pgoff_t index); -struct sgx_encl_mm *sgx_encl_next_mm(struct sgx_encl *encl, - struct sgx_encl_mm *encl_mm, int *iter); -struct sgx_encl_mm *sgx_encl_mm_add(struct sgx_encl *encl, - struct mm_struct *mm); -void sgx_encl_mm_release(struct kref *ref); +int sgx_encl_mm_add(struct sgx_encl *encl, struct mm_struct *mm); +static inline void sgx_encl_mm_release(struct sgx_encl_mm *encl_mm) +{ + kref_put(&encl_mm->encl->refcount, sgx_encl_release); + + kfree(encl_mm); +} int sgx_encl_test_and_clear_young(struct mm_struct *mm, struct sgx_encl_page *page); struct sgx_encl_page *sgx_encl_reserve_page(struct sgx_encl *encl, diff --git a/arch/x86/kernel/cpu/sgx/reclaim.c b/arch/x86/kernel/cpu/sgx/reclaim.c index f192ade93245..e9427220415b 100644 --- a/arch/x86/kernel/cpu/sgx/reclaim.c +++ b/arch/x86/kernel/cpu/sgx/reclaim.c @@ -140,23 +140,13 @@ static bool sgx_reclaimer_evict(struct sgx_epc_page *epc_page) { struct sgx_encl_page *page = epc_page->owner; struct sgx_encl *encl = page->encl; - struct sgx_encl_mm *encl_mm = NULL; - struct sgx_encl_mm *prev_mm = NULL; + struct sgx_encl_mm *encl_mm; bool ret = true; - int iter; + int idx; - while (true) { - encl_mm = sgx_encl_next_mm(encl, prev_mm, &iter); - if (prev_mm) - kref_put(&prev_mm->refcount, sgx_encl_mm_release); - prev_mm = encl_mm; - - if (iter == SGX_ENCL_MM_ITER_DONE) - break; - - if (iter == SGX_ENCL_MM_ITER_RESTART) - continue; + idx = srcu_read_lock(&encl->srcu); + list_for_each_entry_rcu(encl_mm, &encl->mm_list, list) { if (!mmget_not_zero(encl_mm->mm)) continue; @@ -164,14 +154,14 @@ static bool sgx_reclaimer_evict(struct sgx_epc_page *epc_page) ret = !sgx_encl_test_and_clear_young(encl_mm->mm, page); up_read(&encl_mm->mm->mmap_sem); - mmput(encl_mm->mm); + mmput_async(encl_mm->mm); - if (!ret || (encl->flags & SGX_ENCL_DEAD)) { - kref_put(&encl_mm->refcount, sgx_encl_mm_release); + if (!ret || (encl->flags & SGX_ENCL_DEAD)) break; - } } + srcu_read_unlock(&encl->srcu, idx); + /* * Do not reclaim this page if it has been recently accessed by any * mm_struct *and* if the enclave is still alive. No need to take @@ -195,24 +185,13 @@ static void sgx_reclaimer_block(struct sgx_epc_page *epc_page) struct sgx_encl_page *page = epc_page->owner; unsigned long addr = SGX_ENCL_PAGE_ADDR(page); struct sgx_encl *encl = page->encl; - struct sgx_encl_mm *encl_mm = NULL; - struct sgx_encl_mm *prev_mm = NULL; + struct sgx_encl_mm *encl_mm; struct vm_area_struct *vma; - int iter; - int ret; + int idx, ret; - while (true) { - encl_mm = sgx_encl_next_mm(encl, prev_mm, &iter); - if (prev_mm) - kref_put(&prev_mm->refcount, sgx_encl_mm_release); - prev_mm = encl_mm; - - if (iter == SGX_ENCL_MM_ITER_DONE) - break; - - if (iter == SGX_ENCL_MM_ITER_RESTART) - continue; + idx = srcu_read_lock(&encl->srcu); + list_for_each_entry_rcu(encl_mm, &encl->mm_list, list) { if (!mmget_not_zero(encl_mm->mm)) continue; @@ -224,9 +203,11 @@ static void sgx_reclaimer_block(struct sgx_epc_page *epc_page) up_read(&encl_mm->mm->mmap_sem); - mmput(encl_mm->mm); + mmput_async(encl_mm->mm); } + srcu_read_unlock(&encl->srcu, idx); + mutex_lock(&encl->lock); if (!(encl->flags & SGX_ENCL_DEAD)) { @@ -289,32 +270,24 @@ static void sgx_ipi_cb(void *info) static const cpumask_t *sgx_encl_ewb_cpumask(struct sgx_encl *encl) { cpumask_t *cpumask = &encl->cpumask; - struct sgx_encl_mm *encl_mm = NULL; - struct sgx_encl_mm *prev_mm = NULL; - int iter; + struct sgx_encl_mm *encl_mm; + int idx; cpumask_clear(cpumask); - while (true) { - encl_mm = sgx_encl_next_mm(encl, prev_mm, &iter); - if (prev_mm) - kref_put(&prev_mm->refcount, sgx_encl_mm_release); - prev_mm = encl_mm; - - if (iter == SGX_ENCL_MM_ITER_DONE) - break; - - if (iter == SGX_ENCL_MM_ITER_RESTART) - continue; + idx = srcu_read_lock(&encl->srcu); + list_for_each_entry_rcu(encl_mm, &encl->mm_list, list) { if (!mmget_not_zero(encl_mm->mm)) continue; cpumask_or(cpumask, cpumask, mm_cpumask(encl_mm->mm)); - mmput(encl_mm->mm); + mmput_async(encl_mm->mm); } + srcu_read_unlock(&encl->srcu, idx); + return cpumask; }
Using per-vma refcounting to track mm_structs associated with an enclave requires hooking .vm_close(), which in turn prevents the mm from merging vmas (precisely to allow refcounting). Avoid refcounting encl_mm altogether by registering an mmu_notifier at .mmap(), removing the dying encl_mm at mmu_notifier.release() and protecting mm_list during reclaim via a per-enclave SRCU. Removing refcounting/vm_close() allows merging of enclave vmas, at the cost of delaying removal of encl_mm structs from mm_list, i.e. an mm is disassociated from an enclave when the mm exits or the enclave dies, as opposed to when the last vma (in a given mm) is closed. The impact of delying encl_mm removal is its memory footprint and whatever overhead is incurred during EPC reclaim (to walk an mm's vmas). Practically speaking, a stale encl_mm will exist for a meaningful amount of time if and only if the enclave is mapped in a long-lived process and then passed off to another long-lived process. It is expected that the vast majority of use cases will not encounter this condition, e.g. even using a daemon to build enclaves should not result in a stale encl_mm as the builder should never need to mmap() the enclave. Even if there are scenarios that lead to defunct encl_mms, the cost is likely far outweighed by the benefits of reducing the number of vmas across all enclaves. Note, using SRCU to protect mm_list is not strictly necessary, i.e. the existing walker with encl_mm refcounting could be massaged to work with mmu_notifier.release(), but the resulting code is subtle and fragile (I never actually got it working). The primary issue is that an encl_mm can't be moved off the list until its refcount goes to zero, otherwise the custom walker goes off into the weeds. The refcount requirement then prevents using mm_list to identify if an mmu_notifier.release() has fired, i.e. another mechanism is needed to guard against races between exit_mmap() and sgx_release(). Cc: Dave Hansen <dave.hansen@intel.com> Cc: Andy Lutomirski <luto@kernel.org> Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com> --- arch/x86/Kconfig | 2 + arch/x86/kernel/cpu/sgx/driver/ioctl.c | 14 -- arch/x86/kernel/cpu/sgx/driver/main.c | 38 ++++ arch/x86/kernel/cpu/sgx/encl.c | 234 +++++++++++-------------- arch/x86/kernel/cpu/sgx/encl.h | 19 +- arch/x86/kernel/cpu/sgx/reclaim.c | 71 +++----- 6 files changed, 182 insertions(+), 196 deletions(-)