Mailing List Archive

[xen master] x86/paging: fold HAP and shadow memory alloc related fields
commit 5d58062192dc38daeb993589fe0d3c035ef6f9e6
Author: Jan Beulich <jbeulich@suse.com>
AuthorDate: Thu Dec 22 10:05:21 2022 +0100
Commit: Jan Beulich <jbeulich@suse.com>
CommitDate: Thu Dec 22 10:05:21 2022 +0100

x86/paging: fold HAP and shadow memory alloc related fields

Especially with struct shadow_domain and struct hap_domain not living in
a union inside struct paging_domain, let's avoid the duplication: The
fields are named and used in identical ways, and only one of HAP or
shadow can be in use for a domain. This then also renders involved
expressions slightly more legible.

Signed-off-by: Jan Beulich <jbeulich@suse.com>
Reviewed-by: Andrew Cooper <andrew.cooper3@citrix.com>
---
xen/arch/x86/include/asm/domain.h | 17 +++---
xen/arch/x86/include/asm/hap.h | 1 -
xen/arch/x86/include/asm/shadow.h | 4 --
xen/arch/x86/mm/hap/hap.c | 71 ++++++++++-------------
xen/arch/x86/mm/paging.c | 13 +++--
xen/arch/x86/mm/shadow/common.c | 116 +++++++++++++++-----------------------
xen/arch/x86/mm/shadow/hvm.c | 2 +-
7 files changed, 91 insertions(+), 133 deletions(-)

diff --git a/xen/arch/x86/include/asm/domain.h b/xen/arch/x86/include/asm/domain.h
index 43ace233d7..b5354c3677 100644
--- a/xen/arch/x86/include/asm/domain.h
+++ b/xen/arch/x86/include/asm/domain.h
@@ -99,12 +99,6 @@ struct shadow_domain {
unsigned int opt_flags; /* runtime tunable optimizations on/off */
struct page_list_head pinned_shadows;

- /* Memory allocation */
- struct page_list_head freelist;
- unsigned int total_pages; /* number of pages allocated */
- unsigned int free_pages; /* number of pages on freelists */
- unsigned int p2m_pages; /* number of pages allocated to p2m */
-
/* 1-to-1 map for use when HVM vcpus have paging disabled */
pagetable_t unpaged_pagetable;

@@ -179,10 +173,6 @@ struct shadow_vcpu {
/* hardware assisted paging */
/************************************************/
struct hap_domain {
- struct page_list_head freelist;
- unsigned int total_pages; /* number of pages allocated */
- unsigned int free_pages; /* number of pages on freelists */
- unsigned int p2m_pages; /* number of pages allocated to p2m */
};

/************************************************/
@@ -218,6 +208,13 @@ struct paging_domain {
struct shadow_domain shadow;
/* extension for hardware-assited paging */
struct hap_domain hap;
+
+ /* Memory allocation (common to shadow and HAP) */
+ struct page_list_head freelist;
+ unsigned int total_pages; /* number of pages allocated */
+ unsigned int free_pages; /* number of pages on freelists */
+ unsigned int p2m_pages; /* number of pages allocated to p2m */
+
/* log dirty support */
struct log_dirty_domain log_dirty;

diff --git a/xen/arch/x86/include/asm/hap.h b/xen/arch/x86/include/asm/hap.h
index 14d2f212da..90dece29de 100644
--- a/xen/arch/x86/include/asm/hap.h
+++ b/xen/arch/x86/include/asm/hap.h
@@ -47,7 +47,6 @@ int hap_track_dirty_vram(struct domain *d,
extern const struct paging_mode *hap_paging_get_mode(struct vcpu *);
int hap_set_allocation(struct domain *d, unsigned int pages, bool *preempted);
unsigned int hap_get_allocation(struct domain *d);
-int hap_get_allocation_bytes(struct domain *d, uint64_t *size);

#endif /* XEN_HAP_H */

diff --git a/xen/arch/x86/include/asm/shadow.h b/xen/arch/x86/include/asm/shadow.h
index dad876d294..1365fe4805 100644
--- a/xen/arch/x86/include/asm/shadow.h
+++ b/xen/arch/x86/include/asm/shadow.h
@@ -97,8 +97,6 @@ void shadow_blow_tables_per_domain(struct domain *d);
int shadow_set_allocation(struct domain *d, unsigned int pages,
bool *preempted);

-int shadow_get_allocation_bytes(struct domain *d, uint64_t *size);
-
#else /* !CONFIG_SHADOW_PAGING */

#define shadow_vcpu_teardown(v) ASSERT(is_pv_vcpu(v))
@@ -110,8 +108,6 @@ int shadow_get_allocation_bytes(struct domain *d, uint64_t *size);
({ ASSERT_UNREACHABLE(); -EOPNOTSUPP; })
#define shadow_set_allocation(d, pages, preempted) \
({ ASSERT_UNREACHABLE(); -EOPNOTSUPP; })
-#define shadow_get_allocation_bytes(d, size) \
- ({ ASSERT_UNREACHABLE(); -EOPNOTSUPP; })

static inline void sh_remove_shadows(struct domain *d, mfn_t gmfn,
int fast, int all) {}
diff --git a/xen/arch/x86/mm/hap/hap.c b/xen/arch/x86/mm/hap/hap.c
index 0fc1b1d9ac..84cb358df1 100644
--- a/xen/arch/x86/mm/hap/hap.c
+++ b/xen/arch/x86/mm/hap/hap.c
@@ -249,11 +249,11 @@ static struct page_info *hap_alloc(struct domain *d)
if ( unlikely(d->is_dying) )
return NULL;

- pg = page_list_remove_head(&d->arch.paging.hap.freelist);
+ pg = page_list_remove_head(&d->arch.paging.freelist);
if ( unlikely(!pg) )
return NULL;

- d->arch.paging.hap.free_pages--;
+ d->arch.paging.free_pages--;

clear_domain_page(page_to_mfn(pg));

@@ -274,12 +274,12 @@ static void hap_free(struct domain *d, mfn_t mfn)
if ( unlikely(d->is_dying) )
{
free_domheap_page(pg);
- d->arch.paging.hap.total_pages--;
+ d->arch.paging.total_pages--;
return;
}

- d->arch.paging.hap.free_pages++;
- page_list_add_tail(pg, &d->arch.paging.hap.freelist);
+ d->arch.paging.free_pages++;
+ page_list_add_tail(pg, &d->arch.paging.freelist);
}

static struct page_info *cf_check hap_alloc_p2m_page(struct domain *d)
@@ -293,8 +293,8 @@ static struct page_info *cf_check hap_alloc_p2m_page(struct domain *d)

if ( likely(pg != NULL) )
{
- d->arch.paging.hap.total_pages--;
- d->arch.paging.hap.p2m_pages++;
+ d->arch.paging.total_pages--;
+ d->arch.paging.p2m_pages++;
ASSERT(!page_get_owner(pg) && !(pg->count_info & PGC_count_mask));
}
else if ( !d->arch.paging.p2m_alloc_failed && !d->is_dying )
@@ -328,8 +328,8 @@ static void cf_check hap_free_p2m_page(struct domain *d, struct page_info *pg)
pg->count_info &= ~PGC_count_mask;
page_set_owner(pg, NULL);
}
- d->arch.paging.hap.p2m_pages--;
- d->arch.paging.hap.total_pages++;
+ d->arch.paging.p2m_pages--;
+ d->arch.paging.total_pages++;
hap_free(d, page_to_mfn(pg));

paging_unlock(d);
@@ -338,24 +338,13 @@ static void cf_check hap_free_p2m_page(struct domain *d, struct page_info *pg)
/* Return the size of the pool, rounded up to the nearest MB */
unsigned int hap_get_allocation(struct domain *d)
{
- unsigned int pg = d->arch.paging.hap.total_pages
- + d->arch.paging.hap.p2m_pages;
+ unsigned int pg = d->arch.paging.total_pages
+ + d->arch.paging.p2m_pages;

return ((pg >> (20 - PAGE_SHIFT))
+ ((pg & ((1 << (20 - PAGE_SHIFT)) - 1)) ? 1 : 0));
}

-int hap_get_allocation_bytes(struct domain *d, uint64_t *size)
-{
- unsigned long pages = d->arch.paging.hap.total_pages;
-
- pages += d->arch.paging.hap.p2m_pages;
-
- *size = pages << PAGE_SHIFT;
-
- return 0;
-}
-
/* Set the pool of pages to the required number of pages.
* Returns 0 for success, non-zero for failure. */
int hap_set_allocation(struct domain *d, unsigned int pages, bool *preempted)
@@ -364,14 +353,14 @@ int hap_set_allocation(struct domain *d, unsigned int pages, bool *preempted)

ASSERT(paging_locked_by_me(d));

- if ( pages < d->arch.paging.hap.p2m_pages )
+ if ( pages < d->arch.paging.p2m_pages )
pages = 0;
else
- pages -= d->arch.paging.hap.p2m_pages;
+ pages -= d->arch.paging.p2m_pages;

for ( ; ; )
{
- if ( d->arch.paging.hap.total_pages < pages )
+ if ( d->arch.paging.total_pages < pages )
{
/* Need to allocate more memory from domheap */
pg = alloc_domheap_page(d, MEMF_no_owner);
@@ -380,22 +369,22 @@ int hap_set_allocation(struct domain *d, unsigned int pages, bool *preempted)
HAP_PRINTK("failed to allocate hap pages.\n");
return -ENOMEM;
}
- d->arch.paging.hap.free_pages++;
- d->arch.paging.hap.total_pages++;
- page_list_add_tail(pg, &d->arch.paging.hap.freelist);
+ d->arch.paging.free_pages++;
+ d->arch.paging.total_pages++;
+ page_list_add_tail(pg, &d->arch.paging.freelist);
}
- else if ( d->arch.paging.hap.total_pages > pages )
+ else if ( d->arch.paging.total_pages > pages )
{
/* Need to return memory to domheap */
- if ( page_list_empty(&d->arch.paging.hap.freelist) )
+ if ( page_list_empty(&d->arch.paging.freelist) )
{
HAP_PRINTK("failed to free enough hap pages.\n");
return -ENOMEM;
}
- pg = page_list_remove_head(&d->arch.paging.hap.freelist);
+ pg = page_list_remove_head(&d->arch.paging.freelist);
ASSERT(pg);
- d->arch.paging.hap.free_pages--;
- d->arch.paging.hap.total_pages--;
+ d->arch.paging.free_pages--;
+ d->arch.paging.total_pages--;
free_domheap_page(pg);
}
else
@@ -462,8 +451,6 @@ void hap_domain_init(struct domain *d)
.clean = hap_clean_dirty_bitmap,
};

- INIT_PAGE_LIST_HEAD(&d->arch.paging.hap.freelist);
-
/* Use HAP logdirty mechanism. */
paging_log_dirty_init(d, &hap_ops);
}
@@ -484,7 +471,7 @@ int hap_enable(struct domain *d, u32 mode)

domain_pause(d);

- old_pages = d->arch.paging.hap.total_pages;
+ old_pages = d->arch.paging.total_pages;
if ( old_pages == 0 )
{
paging_lock(d);
@@ -566,16 +553,16 @@ void hap_final_teardown(struct domain *d)
p2m_teardown(d->arch.nested_p2m[i], true, NULL);
}

- if ( d->arch.paging.hap.total_pages != 0 )
+ if ( d->arch.paging.total_pages != 0 )
hap_teardown(d, NULL);

p2m_teardown(p2m_get_hostp2m(d), true, NULL);
/* Free any memory that the p2m teardown released */
paging_lock(d);
hap_set_allocation(d, 0, NULL);
- ASSERT(d->arch.paging.hap.p2m_pages == 0);
- ASSERT(d->arch.paging.hap.free_pages == 0);
- ASSERT(d->arch.paging.hap.total_pages == 0);
+ ASSERT(d->arch.paging.p2m_pages == 0);
+ ASSERT(d->arch.paging.free_pages == 0);
+ ASSERT(d->arch.paging.total_pages == 0);
paging_unlock(d);
}

@@ -644,14 +631,14 @@ void hap_teardown(struct domain *d, bool *preempted)

paging_lock(d); /* Keep various asserts happy */

- if ( d->arch.paging.hap.total_pages != 0 )
+ if ( d->arch.paging.total_pages != 0 )
{
hap_set_allocation(d, 0, preempted);

if ( preempted && *preempted )
goto out;

- ASSERT(d->arch.paging.hap.total_pages == 0);
+ ASSERT(d->arch.paging.total_pages == 0);
}

d->arch.paging.mode &= ~PG_log_dirty;
diff --git a/xen/arch/x86/mm/paging.c b/xen/arch/x86/mm/paging.c
index 8d579fa9a3..1c5f54f92a 100644
--- a/xen/arch/x86/mm/paging.c
+++ b/xen/arch/x86/mm/paging.c
@@ -651,6 +651,7 @@ int paging_domain_init(struct domain *d)
if ( (rc = p2m_init(d)) != 0 )
return rc;

+ INIT_PAGE_LIST_HEAD(&d->arch.paging.freelist);
mm_lock_init(&d->arch.paging.lock);

/* This must be initialized separately from the rest of the
@@ -979,17 +980,17 @@ int __init paging_set_allocation(struct domain *d, unsigned int pages,

int arch_get_paging_mempool_size(struct domain *d, uint64_t *size)
{
- int rc;
+ unsigned long pages;

if ( is_pv_domain(d) ) /* TODO: Relax in due course */
return -EOPNOTSUPP;

- if ( hap_enabled(d) )
- rc = hap_get_allocation_bytes(d, size);
- else
- rc = shadow_get_allocation_bytes(d, size);
+ pages = d->arch.paging.total_pages;
+ pages += d->arch.paging.p2m_pages;

- return rc;
+ *size = pages << PAGE_SHIFT;
+
+ return 0;
}

int arch_set_paging_mempool_size(struct domain *d, uint64_t size)
diff --git a/xen/arch/x86/mm/shadow/common.c b/xen/arch/x86/mm/shadow/common.c
index cae1d9190c..16a5a2bcf8 100644
--- a/xen/arch/x86/mm/shadow/common.c
+++ b/xen/arch/x86/mm/shadow/common.c
@@ -55,7 +55,6 @@ int shadow_domain_init(struct domain *d)
.clean = sh_clean_dirty_bitmap,
};

- INIT_PAGE_LIST_HEAD(&d->arch.paging.shadow.freelist);
INIT_PAGE_LIST_HEAD(&d->arch.paging.shadow.pinned_shadows);

/* Use shadow pagetables for log-dirty support */
@@ -936,7 +935,7 @@ static bool __must_check _shadow_prealloc(struct domain *d, unsigned int pages)
mfn_t smfn;
int i;

- if ( d->arch.paging.shadow.free_pages >= pages )
+ if ( d->arch.paging.free_pages >= pages )
return true;

if ( unlikely(d->is_dying) )
@@ -958,7 +957,7 @@ static bool __must_check _shadow_prealloc(struct domain *d, unsigned int pages)
sh_unpin(d, smfn);

/* See if that freed up enough space */
- if ( d->arch.paging.shadow.free_pages >= pages )
+ if ( d->arch.paging.free_pages >= pages )
return true;
}

@@ -979,7 +978,7 @@ static bool __must_check _shadow_prealloc(struct domain *d, unsigned int pages)
0);

/* See if that freed up enough space */
- if ( d->arch.paging.shadow.free_pages >= pages )
+ if ( d->arch.paging.free_pages >= pages )
{
guest_flush_tlb_mask(d, d->dirty_cpumask);
return true;
@@ -991,10 +990,8 @@ static bool __must_check _shadow_prealloc(struct domain *d, unsigned int pages)
* hold Xen mappings for some vcpu. This can never happen. */
printk(XENLOG_ERR "Can't pre-allocate %u shadow pages!\n"
" shadow pages total = %u, free = %u, p2m=%u\n",
- pages,
- d->arch.paging.shadow.total_pages,
- d->arch.paging.shadow.free_pages,
- d->arch.paging.shadow.p2m_pages);
+ pages, d->arch.paging.total_pages,
+ d->arch.paging.free_pages, d->arch.paging.p2m_pages);

ASSERT_UNREACHABLE();

@@ -1122,7 +1119,7 @@ mfn_t shadow_alloc(struct domain *d,
ASSERT(shadow_type != SH_type_none);
perfc_incr(shadow_alloc);

- if ( d->arch.paging.shadow.free_pages < pages )
+ if ( d->arch.paging.free_pages < pages )
{
/* If we get here, we failed to allocate. This should never
* happen. It means that we didn't call shadow_prealloc()
@@ -1132,7 +1129,7 @@ mfn_t shadow_alloc(struct domain *d,
printk(XENLOG_ERR "Can't allocate %u shadow pages!\n", pages);
BUG();
}
- d->arch.paging.shadow.free_pages -= pages;
+ d->arch.paging.free_pages -= pages;

/* Backpointers that are MFNs need to be packed into PDXs (PFNs don't) */
switch (shadow_type)
@@ -1153,7 +1150,7 @@ mfn_t shadow_alloc(struct domain *d,
/* Init page info fields and clear the pages */
for ( i = 0; i < pages ; i++ )
{
- sp = page_list_remove_head(&d->arch.paging.shadow.freelist);
+ sp = page_list_remove_head(&d->arch.paging.freelist);
/* Before we overwrite the old contents of this page,
* we need to be sure that no TLB holds a pointer to it. */
cpumask_copy(&mask, d->dirty_cpumask);
@@ -1249,15 +1246,15 @@ void shadow_free(struct domain *d, mfn_t smfn)
free_domheap_page(sp);
}
else
- page_list_add_tail(sp, &d->arch.paging.shadow.freelist);
+ page_list_add_tail(sp, &d->arch.paging.freelist);

sp = next;
}

if ( unlikely(dying) )
- d->arch.paging.shadow.total_pages -= pages;
+ d->arch.paging.total_pages -= pages;
else
- d->arch.paging.shadow.free_pages += pages;
+ d->arch.paging.free_pages += pages;
}

/* Divert a page from the pool to be used by the p2m mapping.
@@ -1276,7 +1273,7 @@ shadow_alloc_p2m_page(struct domain *d)
* paging lock) and the log-dirty code (which always does). */
paging_lock_recursive(d);

- if ( d->arch.paging.shadow.total_pages
+ if ( d->arch.paging.total_pages
< shadow_min_acceptable_pages(d) + 1 )
{
if ( !d->arch.paging.p2m_alloc_failed )
@@ -1284,8 +1281,8 @@ shadow_alloc_p2m_page(struct domain *d)
d->arch.paging.p2m_alloc_failed = 1;
dprintk(XENLOG_ERR,
"d%d failed to allocate from shadow pool (tot=%u p2m=%u min=%u)\n",
- d->domain_id, d->arch.paging.shadow.total_pages,
- d->arch.paging.shadow.p2m_pages,
+ d->domain_id, d->arch.paging.total_pages,
+ d->arch.paging.p2m_pages,
shadow_min_acceptable_pages(d));
}
goto out;
@@ -1295,8 +1292,8 @@ shadow_alloc_p2m_page(struct domain *d)
goto out;

pg = mfn_to_page(shadow_alloc(d, SH_type_p2m_table, 0));
- d->arch.paging.shadow.p2m_pages++;
- d->arch.paging.shadow.total_pages--;
+ d->arch.paging.p2m_pages++;
+ d->arch.paging.total_pages--;
ASSERT(!page_get_owner(pg) && !(pg->count_info & PGC_count_mask));

out:
@@ -1327,8 +1324,8 @@ shadow_free_p2m_page(struct domain *d, struct page_info *pg)
* paging lock) and the log-dirty code (which always does). */
paging_lock_recursive(d);

- d->arch.paging.shadow.p2m_pages--;
- d->arch.paging.shadow.total_pages++;
+ d->arch.paging.p2m_pages--;
+ d->arch.paging.total_pages++;
shadow_free(d, page_to_mfn(pg));

paging_unlock(d);
@@ -1345,7 +1342,7 @@ static unsigned int sh_min_allocation(const struct domain *d)
max(max(domain_tot_pages(d) / 256,
is_hvm_domain(d) ? CONFIG_PAGING_LEVELS + 2 : 0U) +
is_hvm_domain(d),
- d->arch.paging.shadow.p2m_pages);
+ d->arch.paging.p2m_pages);
}

int shadow_set_allocation(struct domain *d, unsigned int pages, bool *preempted)
@@ -1361,15 +1358,15 @@ int shadow_set_allocation(struct domain *d, unsigned int pages, bool *preempted)

if ( pages < lower_bound )
pages = lower_bound;
- pages -= d->arch.paging.shadow.p2m_pages;
+ pages -= d->arch.paging.p2m_pages;
}

SHADOW_PRINTK("current %i target %i\n",
- d->arch.paging.shadow.total_pages, pages);
+ d->arch.paging.total_pages, pages);

for ( ; ; )
{
- if ( d->arch.paging.shadow.total_pages < pages )
+ if ( d->arch.paging.total_pages < pages )
{
/* Need to allocate more memory from domheap */
sp = (struct page_info *)
@@ -1379,29 +1376,29 @@ int shadow_set_allocation(struct domain *d, unsigned int pages, bool *preempted)
SHADOW_PRINTK("failed to allocate shadow pages.\n");
return -ENOMEM;
}
- d->arch.paging.shadow.free_pages++;
- d->arch.paging.shadow.total_pages++;
+ d->arch.paging.free_pages++;
+ d->arch.paging.total_pages++;
sp->u.sh.type = 0;
sp->u.sh.pinned = 0;
sp->u.sh.count = 0;
sp->tlbflush_timestamp = 0; /* Not in any TLB */
- page_list_add_tail(sp, &d->arch.paging.shadow.freelist);
+ page_list_add_tail(sp, &d->arch.paging.freelist);
}
- else if ( d->arch.paging.shadow.total_pages > pages )
+ else if ( d->arch.paging.total_pages > pages )
{
/* Need to return memory to domheap */
if ( !_shadow_prealloc(d, 1) )
return -ENOMEM;

- sp = page_list_remove_head(&d->arch.paging.shadow.freelist);
+ sp = page_list_remove_head(&d->arch.paging.freelist);
ASSERT(sp);
/*
* The pages were allocated anonymously, but the owner field
* gets overwritten normally, so need to clear it here.
*/
page_set_owner(sp, NULL);
- d->arch.paging.shadow.free_pages--;
- d->arch.paging.shadow.total_pages--;
+ d->arch.paging.free_pages--;
+ d->arch.paging.total_pages--;
free_domheap_page(sp);
}
else
@@ -1421,23 +1418,12 @@ int shadow_set_allocation(struct domain *d, unsigned int pages, bool *preempted)
/* Return the size of the shadow pool, rounded up to the nearest MB */
static unsigned int shadow_get_allocation(struct domain *d)
{
- unsigned int pg = d->arch.paging.shadow.total_pages
- + d->arch.paging.shadow.p2m_pages;
+ unsigned int pg = d->arch.paging.total_pages
+ + d->arch.paging.p2m_pages;
return ((pg >> (20 - PAGE_SHIFT))
+ ((pg & ((1 << (20 - PAGE_SHIFT)) - 1)) ? 1 : 0));
}

-int shadow_get_allocation_bytes(struct domain *d, uint64_t *size)
-{
- unsigned long pages = d->arch.paging.shadow.total_pages;
-
- pages += d->arch.paging.shadow.p2m_pages;
-
- *size = pages << PAGE_SHIFT;
-
- return 0;
-}
-
/**************************************************************************/
/* Hash table for storing the guest->shadow mappings.
* The table itself is an array of pointers to shadows; the shadows are then
@@ -2708,7 +2694,7 @@ int shadow_enable(struct domain *d, u32 mode)
}

/* Init the shadow memory allocation if the user hasn't done so */
- old_pages = d->arch.paging.shadow.total_pages;
+ old_pages = d->arch.paging.total_pages;
if ( old_pages < sh_min_allocation(d) )
{
paging_lock(d);
@@ -2888,7 +2874,7 @@ void shadow_teardown(struct domain *d, bool *preempted)
}
#endif /* (SHADOW_OPTIMIZATIONS & (SHOPT_VIRTUAL_TLB|SHOPT_OUT_OF_SYNC)) */

- if ( d->arch.paging.shadow.total_pages != 0 )
+ if ( d->arch.paging.total_pages != 0 )
{
/* Destroy all the shadows and release memory to domheap */
shadow_set_allocation(d, 0, preempted);
@@ -2900,7 +2886,7 @@ void shadow_teardown(struct domain *d, bool *preempted)
if (d->arch.paging.shadow.hash_table)
shadow_hash_teardown(d);

- ASSERT(d->arch.paging.shadow.total_pages == 0);
+ ASSERT(d->arch.paging.total_pages == 0);
}

/* Free the non-paged-vcpus pagetable; must happen after we've
@@ -2959,15 +2945,13 @@ void shadow_final_teardown(struct domain *d)
{
SHADOW_PRINTK("dom %u final teardown starts."
" Shadow pages total = %u, free = %u, p2m=%u\n",
- d->domain_id,
- d->arch.paging.shadow.total_pages,
- d->arch.paging.shadow.free_pages,
- d->arch.paging.shadow.p2m_pages);
+ d->domain_id, d->arch.paging.total_pages,
+ d->arch.paging.free_pages, d->arch.paging.p2m_pages);

/* Double-check that the domain didn't have any shadow memory.
* It is possible for a domain that never got domain_kill()ed
* to get here with its shadow allocation intact. */
- if ( d->arch.paging.shadow.total_pages != 0 )
+ if ( d->arch.paging.total_pages != 0 )
shadow_teardown(d, NULL);

/* It is now safe to pull down the p2m map. */
@@ -2977,13 +2961,11 @@ void shadow_final_teardown(struct domain *d)
shadow_set_allocation(d, 0, NULL);
SHADOW_PRINTK("dom %u final teardown done."
" Shadow pages total = %u, free = %u, p2m=%u\n",
- d->domain_id,
- d->arch.paging.shadow.total_pages,
- d->arch.paging.shadow.free_pages,
- d->arch.paging.shadow.p2m_pages);
- ASSERT(!d->arch.paging.shadow.total_pages);
- ASSERT(!d->arch.paging.shadow.free_pages);
- ASSERT(!d->arch.paging.shadow.p2m_pages);
+ d->domain_id, d->arch.paging.total_pages,
+ d->arch.paging.free_pages, d->arch.paging.p2m_pages);
+ ASSERT(d->arch.paging.p2m_pages == 0);
+ ASSERT(d->arch.paging.free_pages == 0);
+ ASSERT(d->arch.paging.total_pages == 0);
paging_unlock(d);
}

@@ -3000,7 +2982,7 @@ static int shadow_one_bit_enable(struct domain *d, u32 mode)

mode |= PG_SH_enable;

- if ( d->arch.paging.shadow.total_pages < sh_min_allocation(d) )
+ if ( d->arch.paging.total_pages < sh_min_allocation(d) )
{
/* Init the shadow memory allocation if the user hasn't done so */
if ( shadow_set_allocation(d, 1, NULL) != 0 )
@@ -3049,10 +3031,8 @@ static int shadow_one_bit_disable(struct domain *d, u32 mode)
/* Get this domain off shadows */
SHADOW_PRINTK("un-shadowing of domain %u starts."
" Shadow pages total = %u, free = %u, p2m=%u\n",
- d->domain_id,
- d->arch.paging.shadow.total_pages,
- d->arch.paging.shadow.free_pages,
- d->arch.paging.shadow.p2m_pages);
+ d->domain_id, d->arch.paging.total_pages,
+ d->arch.paging.free_pages, d->arch.paging.p2m_pages);
for_each_vcpu(d, v)
{
if ( v->arch.paging.mode )
@@ -3082,10 +3062,8 @@ static int shadow_one_bit_disable(struct domain *d, u32 mode)
shadow_hash_teardown(d);
SHADOW_PRINTK("un-shadowing of domain %u done."
" Shadow pages total = %u, free = %u, p2m=%u\n",
- d->domain_id,
- d->arch.paging.shadow.total_pages,
- d->arch.paging.shadow.free_pages,
- d->arch.paging.shadow.p2m_pages);
+ d->domain_id, d->arch.paging.total_pages,
+ d->arch.paging.free_pages, d->arch.paging.p2m_pages);
}

return 0;
diff --git a/xen/arch/x86/mm/shadow/hvm.c b/xen/arch/x86/mm/shadow/hvm.c
index 29a58d9131..88c3c16322 100644
--- a/xen/arch/x86/mm/shadow/hvm.c
+++ b/xen/arch/x86/mm/shadow/hvm.c
@@ -809,7 +809,7 @@ static void cf_check sh_unshadow_for_p2m_change(
* If there are any shadows, update them. But if shadow_teardown()
* has already been called then it's not safe to try.
*/
- if ( unlikely(!d->arch.paging.shadow.total_pages) )
+ if ( unlikely(!d->arch.paging.total_pages) )
return;

/* Only previously present / valid entries need processing. */
--
generated by git-patchbot for /home/xen/git/xen.git#master