@@ -35,7 +35,6 @@
#include <asm/setup.h>
#include <asm/bzimage.h> /* for bzimage_parse */
#include <asm/io_apic.h>
-#include <asm/hap.h>
#include <asm/hpet.h>
#include <public/version.h>
@@ -1383,15 +1382,25 @@ int __init construct_dom0(
nr_pages);
}
- if ( is_pvh_domain(d) )
- hap_set_alloc_for_pvh_dom0(d, dom0_paging_pages(d, nr_pages));
-
/*
- * We enable paging mode again so guest_physmap_add_page will do the
- * right thing for us.
+ * We enable paging mode again so guest_physmap_add_page and
+ * paging_set_allocation will do the right thing for us.
*/
d->arch.paging.mode = save_pvh_pg_mode;
+ if ( is_pvh_domain(d) )
+ {
+ bool preempted;
+
+ do {
+ preempted = false;
+ paging_set_allocation(d, dom0_paging_pages(d, nr_pages),
+ &preempted);
+ process_pending_softirqs();
+ } while ( preempted );
+ }
+
+
/* Write the phys->machine and machine->phys table entries. */
for ( pfn = 0; pfn < count; pfn++ )
{
@@ -334,8 +334,7 @@ hap_get_allocation(struct domain *d)
/* Set the pool of pages to the required number of pages.
* Returns 0 for success, non-zero for failure. */
-static int
-hap_set_allocation(struct domain *d, unsigned int pages, int *preempted)
+int hap_set_allocation(struct domain *d, unsigned int pages, bool *preempted)
{
struct page_info *pg;
@@ -381,7 +380,7 @@ hap_set_allocation(struct domain *d, unsigned int pages, int *preempted)
/* Check to see if we need to yield and try again */
if ( preempted && general_preempt_check() )
{
- *preempted = 1;
+ *preempted = true;
return 0;
}
}
@@ -561,7 +560,7 @@ void hap_final_teardown(struct domain *d)
paging_unlock(d);
}
-void hap_teardown(struct domain *d, int *preempted)
+void hap_teardown(struct domain *d, bool *preempted)
{
struct vcpu *v;
mfn_t mfn;
@@ -609,7 +608,8 @@ out:
int hap_domctl(struct domain *d, xen_domctl_shadow_op_t *sc,
XEN_GUEST_HANDLE_PARAM(void) u_domctl)
{
- int rc, preempted = 0;
+ int rc;
+ bool preempted = false;
switch ( sc->op )
{
@@ -636,18 +636,6 @@ int hap_domctl(struct domain *d, xen_domctl_shadow_op_t *sc,
}
}
-void __init hap_set_alloc_for_pvh_dom0(struct domain *d,
- unsigned long hap_pages)
-{
- int rc;
-
- paging_lock(d);
- rc = hap_set_allocation(d, hap_pages, NULL);
- paging_unlock(d);
-
- BUG_ON(rc);
-}
-
static const struct paging_mode hap_paging_real_mode;
static const struct paging_mode hap_paging_protected_mode;
static const struct paging_mode hap_paging_pae_mode;
@@ -809,7 +809,8 @@ long paging_domctl_continuation(XEN_GUEST_HANDLE_PARAM(xen_domctl_t) u_domctl)
/* Call when destroying a domain */
int paging_teardown(struct domain *d)
{
- int rc, preempted = 0;
+ int rc;
+ bool preempted = false;
if ( hap_enabled(d) )
hap_teardown(d, &preempted);
@@ -954,6 +955,22 @@ void paging_write_p2m_entry(struct p2m_domain *p2m, unsigned long gfn,
safe_write_pte(p, new);
}
+int paging_set_allocation(struct domain *d, unsigned int pages, bool *preempted)
+{
+ int rc;
+
+ ASSERT(paging_mode_enabled(d));
+
+ paging_lock(d);
+ if ( hap_enabled(d) )
+ rc = hap_set_allocation(d, pages, preempted);
+ else
+ rc = shadow_set_allocation(d, pages, preempted);
+ paging_unlock(d);
+
+ return rc;
+}
+
/*
* Local variables:
* mode: C
@@ -1611,13 +1611,7 @@ shadow_free_p2m_page(struct domain *d, struct page_info *pg)
paging_unlock(d);
}
-/* Set the pool of shadow pages to the required number of pages.
- * Input will be rounded up to at least shadow_min_acceptable_pages(),
- * plus space for the p2m table.
- * Returns 0 for success, non-zero for failure. */
-static int sh_set_allocation(struct domain *d,
- unsigned int pages,
- int *preempted)
+int shadow_set_allocation(struct domain *d, unsigned int pages, bool *preempted)
{
struct page_info *sp;
unsigned int lower_bound;
@@ -1683,7 +1677,7 @@ static int sh_set_allocation(struct domain *d,
/* Check to see if we need to yield and try again */
if ( preempted && general_preempt_check() )
{
- *preempted = 1;
+ *preempted = true;
return 0;
}
}
@@ -3154,10 +3148,10 @@ int shadow_enable(struct domain *d, u32 mode)
if ( old_pages == 0 )
{
paging_lock(d);
- rv = sh_set_allocation(d, 1024, NULL); /* Use at least 4MB */
+ rv = shadow_set_allocation(d, 1024, NULL); /* Use at least 4MB */
if ( rv != 0 )
{
- sh_set_allocation(d, 0, NULL);
+ shadow_set_allocation(d, 0, NULL);
goto out_locked;
}
paging_unlock(d);
@@ -3239,7 +3233,7 @@ int shadow_enable(struct domain *d, u32 mode)
return rv;
}
-void shadow_teardown(struct domain *d, int *preempted)
+void shadow_teardown(struct domain *d, bool *preempted)
/* Destroy the shadow pagetables of this domain and free its shadow memory.
* Should only be called for dying domains. */
{
@@ -3301,7 +3295,7 @@ void shadow_teardown(struct domain *d, int *preempted)
if ( d->arch.paging.shadow.total_pages != 0 )
{
/* Destroy all the shadows and release memory to domheap */
- sh_set_allocation(d, 0, preempted);
+ shadow_set_allocation(d, 0, preempted);
if ( preempted && *preempted )
goto out;
@@ -3366,7 +3360,7 @@ void shadow_final_teardown(struct domain *d)
p2m_teardown(p2m_get_hostp2m(d));
/* Free any shadow memory that the p2m teardown released */
paging_lock(d);
- sh_set_allocation(d, 0, NULL);
+ shadow_set_allocation(d, 0, NULL);
SHADOW_PRINTK("dom %u final teardown done."
" Shadow pages total = %u, free = %u, p2m=%u\n",
d->domain_id,
@@ -3392,9 +3386,9 @@ static int shadow_one_bit_enable(struct domain *d, u32 mode)
if ( d->arch.paging.shadow.total_pages == 0 )
{
/* Init the shadow memory allocation if the user hasn't done so */
- if ( sh_set_allocation(d, 1, NULL) != 0 )
+ if ( shadow_set_allocation(d, 1, NULL) != 0 )
{
- sh_set_allocation(d, 0, NULL);
+ shadow_set_allocation(d, 0, NULL);
return -ENOMEM;
}
}
@@ -3463,7 +3457,7 @@ static int shadow_one_bit_disable(struct domain *d, u32 mode)
}
/* Pull down the memory allocation */
- if ( sh_set_allocation(d, 0, NULL) != 0 )
+ if ( shadow_set_allocation(d, 0, NULL) != 0 )
BUG(); /* In fact, we will have BUG()ed already */
shadow_hash_teardown(d);
SHADOW_PRINTK("un-shadowing of domain %u done."
@@ -3876,7 +3870,8 @@ int shadow_domctl(struct domain *d,
xen_domctl_shadow_op_t *sc,
XEN_GUEST_HANDLE_PARAM(void) u_domctl)
{
- int rc, preempted = 0;
+ int rc;
+ bool preempted = false;
switch ( sc->op )
{
@@ -3907,7 +3902,7 @@ int shadow_domctl(struct domain *d,
paging_unlock(d);
return -EINVAL;
}
- rc = sh_set_allocation(d, sc->mb << (20 - PAGE_SHIFT), &preempted);
+ rc = shadow_set_allocation(d, sc->mb << (20 - PAGE_SHIFT), &preempted);
paging_unlock(d);
if ( preempted )
/* Not finished. Set up to re-run the call. */
@@ -38,7 +38,7 @@ int hap_domctl(struct domain *d, xen_domctl_shadow_op_t *sc,
XEN_GUEST_HANDLE_PARAM(void) u_domctl);
int hap_enable(struct domain *d, u32 mode);
void hap_final_teardown(struct domain *d);
-void hap_teardown(struct domain *d, int *preempted);
+void hap_teardown(struct domain *d, bool *preempted);
void hap_vcpu_init(struct vcpu *v);
int hap_track_dirty_vram(struct domain *d,
unsigned long begin_pfn,
@@ -46,7 +46,7 @@ int hap_track_dirty_vram(struct domain *d,
XEN_GUEST_HANDLE_64(uint8) dirty_bitmap);
extern const struct paging_mode *hap_paging_get_mode(struct vcpu *);
-void hap_set_alloc_for_pvh_dom0(struct domain *d, unsigned long num_pages);
+int hap_set_allocation(struct domain *d, unsigned int pages, bool *preempted);
#endif /* XEN_HAP_H */
@@ -347,6 +347,13 @@ void pagetable_dying(struct domain *d, paddr_t gpa);
void paging_dump_domain_info(struct domain *d);
void paging_dump_vcpu_info(struct vcpu *v);
+/* Set the pool of shadow pages to the required number of pages.
+ * Input might be rounded up to at minimum amount of pages, plus
+ * space for the p2m table.
+ * Returns 0 for success, non-zero for failure. */
+int paging_set_allocation(struct domain *d, unsigned int pages,
+ bool *preempted);
+
#endif /* XEN_PAGING_H */
/*
@@ -73,7 +73,7 @@ int shadow_domctl(struct domain *d,
XEN_GUEST_HANDLE_PARAM(void) u_domctl);
/* Call when destroying a domain */
-void shadow_teardown(struct domain *d, int *preempted);
+void shadow_teardown(struct domain *d, bool *preempted);
/* Call once all of the references to the domain have gone away */
void shadow_final_teardown(struct domain *d);
@@ -83,6 +83,13 @@ void sh_remove_shadows(struct domain *d, mfn_t gmfn, int fast, int all);
/* Discard _all_ mappings from the domain's shadows. */
void shadow_blow_tables_per_domain(struct domain *d);
+/* Set the pool of shadow pages to the required number of pages.
+ * Input will be rounded up to at least shadow_min_acceptable_pages(),
+ * plus space for the p2m table.
+ * Returns 0 for success, non-zero for failure. */
+int shadow_set_allocation(struct domain *d, unsigned int pages,
+ bool *preempted);
+
#else /* !CONFIG_SHADOW_PAGING */
#define shadow_teardown(d, p) ASSERT(is_pv_domain(d))
@@ -91,6 +98,8 @@ void shadow_blow_tables_per_domain(struct domain *d);
({ ASSERT(is_pv_domain(d)); -EOPNOTSUPP; })
#define shadow_track_dirty_vram(d, begin_pfn, nr, bitmap) \
({ ASSERT_UNREACHABLE(); -EOPNOTSUPP; })
+#define shadow_set_allocation(d, pages, preempted) \
+ ({ ASSERT_UNREACHABLE(); -EOPNOTSUPP; })
static inline void sh_remove_shadows(struct domain *d, mfn_t gmfn,
bool_t fast, bool_t all) {}