@@ -926,11 +926,58 @@ static int reserve_offlined_page(struct page_info *head)
return count;
}
+static bool_t can_merge(struct page_info *head, unsigned int node,
+ unsigned int order)
+{
+ if ( !mfn_valid(page_to_mfn(head)) ||
+ !page_state_is(head, free) ||
+ (PFN_ORDER(head) != order) ||
+ (phys_to_nid(page_to_maddr(head)) != node) )
+ return 0;
+
+ return 1;
+}
+
+static void merge_chunks(struct page_info *pg, unsigned int node,
+ unsigned int zone, unsigned int order)
+{
+ ASSERT(spin_is_locked(&heap_lock));
+
+ /* Merge chunks as far as possible. */
+ while ( order < MAX_ORDER )
+ {
+ unsigned int mask = 1UL << order;
+
+ if ( (page_to_mfn(pg) & mask) )
+ {
+ /* Merge with predecessor block? */
+ if ( !can_merge(pg - mask, node, order) )
+ break;
+
+ pg -= mask;
+ page_list_del(pg, &heap(node, zone, order));
+ }
+ else
+ {
+ /* Merge with successor block? */
+ if ( !can_merge(pg + mask, node, order) )
+ break;
+
+ page_list_del(pg + mask, &heap(node, zone, order));
+ }
+
+ order++;
+ }
+
+ PFN_ORDER(pg) = order;
+ page_list_add_tail(pg, &heap(node, zone, order));
+}
+
/* Free 2^@order set of pages. */
static void free_heap_pages(
struct page_info *pg, unsigned int order)
{
- unsigned long mask, mfn = page_to_mfn(pg);
+ unsigned long mfn = page_to_mfn(pg);
unsigned int i, node = phys_to_nid(page_to_maddr(pg)), tainted = 0;
unsigned int zone = page_to_zone(pg);
@@ -977,38 +1024,7 @@ static void free_heap_pages(
midsize_alloc_zone_pages = max(
midsize_alloc_zone_pages, total_avail_pages / MIDSIZE_ALLOC_FRAC);
- /* Merge chunks as far as possible. */
- while ( order < MAX_ORDER )
- {
- mask = 1UL << order;
-
- if ( (page_to_mfn(pg) & mask) )
- {
- /* Merge with predecessor block? */
- if ( !mfn_valid(page_to_mfn(pg-mask)) ||
- !page_state_is(pg-mask, free) ||
- (PFN_ORDER(pg-mask) != order) ||
- (phys_to_nid(page_to_maddr(pg-mask)) != node) )
- break;
- pg -= mask;
- page_list_del(pg, &heap(node, zone, order));
- }
- else
- {
- /* Merge with successor block? */
- if ( !mfn_valid(page_to_mfn(pg+mask)) ||
- !page_state_is(pg+mask, free) ||
- (PFN_ORDER(pg+mask) != order) ||
- (phys_to_nid(page_to_maddr(pg+mask)) != node) )
- break;
- page_list_del(pg + mask, &heap(node, zone, order));
- }
-
- order++;
- }
-
- PFN_ORDER(pg) = order;
- page_list_add_tail(pg, &heap(node, zone, order));
+ merge_chunks(pg, node, zone, order);
if ( tainted )
reserve_offlined_page(pg);
This is needed for subsequent changes to memory scrubbing. Signed-off-by: Boris Ostrovsky <boris.ostrovsky@oracle.com> --- xen/common/page_alloc.c | 82 ++++++++++++++++++++++++++++------------------- 1 files changed, 49 insertions(+), 33 deletions(-)