Mailing List Archive

[xen staging] x86: mem-access is HVM-only
commit 99192701e34e2e82276289efe4fd450fb91a703b
Author: Jan Beulich <jbeulich@suse.com>
AuthorDate: Mon Apr 19 15:28:00 2021 +0200
Commit: Jan Beulich <jbeulich@suse.com>
CommitDate: Mon Apr 19 15:28:00 2021 +0200

x86: mem-access is HVM-only

By excluding the file from being built for !HVM, #ifdef-ary can be
removed from it.

The new HVM dependency on the Kconfig option is benign for Arm.

Signed-off-by: Jan Beulich <jbeulich@suse.com>
Acked-by: Tamas K Lengyel <tamas@tklengyel.com>
Reviewed-by: Alexandru Isaila <aisaila@bitdefender.com>
---
xen/arch/x86/Kconfig | 2 +-
xen/arch/x86/mm/mem_access.c | 20 --------------------
xen/arch/x86/vm_event.c | 2 ++
xen/common/Kconfig | 1 +
4 files changed, 4 insertions(+), 21 deletions(-)

diff --git a/xen/arch/x86/Kconfig b/xen/arch/x86/Kconfig
index 57776d5106..db0de0a9ec 100644
--- a/xen/arch/x86/Kconfig
+++ b/xen/arch/x86/Kconfig
@@ -16,7 +16,6 @@ config X86
select HAS_FAST_MULTIPLY
select HAS_IOPORTS
select HAS_KEXEC
- select MEM_ACCESS_ALWAYS_ON
select HAS_MEM_PAGING
select HAS_NS16550
select HAS_PASSTHROUGH
@@ -95,6 +94,7 @@ config HVM
def_bool !PV_SHIM_EXCLUSIVE
select COMPAT
select IOREQ_SERVER
+ select MEM_ACCESS_ALWAYS_ON
prompt "HVM support"
---help---
Interfaces to support HVM domains. HVM domains require hardware
diff --git a/xen/arch/x86/mm/mem_access.c b/xen/arch/x86/mm/mem_access.c
index ede774fb50..7750b81cd4 100644
--- a/xen/arch/x86/mm/mem_access.c
+++ b/xen/arch/x86/mm/mem_access.c
@@ -139,7 +139,6 @@ bool p2m_mem_access_emulate_check(struct vcpu *v,
return violation;
}

-#ifdef CONFIG_HVM
bool p2m_mem_access_check(paddr_t gpa, unsigned long gla,
struct npfec npfec,
vm_event_request_t **req_ptr)
@@ -282,7 +281,6 @@ int p2m_set_altp2m_mem_access(struct domain *d, struct p2m_domain *hp2m,
*/
return ap2m->set_entry(ap2m, gfn, mfn, PAGE_ORDER_4K, t, a, -1);
}
-#endif

static int set_mem_access(struct domain *d, struct p2m_domain *p2m,
struct p2m_domain *ap2m, p2m_access_t a,
@@ -290,7 +288,6 @@ static int set_mem_access(struct domain *d, struct p2m_domain *p2m,
{
int rc = 0;

-#ifdef CONFIG_HVM
if ( ap2m )
{
rc = p2m_set_altp2m_mem_access(d, p2m, ap2m, a, gfn);
@@ -299,9 +296,6 @@ static int set_mem_access(struct domain *d, struct p2m_domain *p2m,
rc = 0;
}
else
-#else
- ASSERT(!ap2m);
-#endif
{
p2m_access_t _a;
p2m_type_t t;
@@ -362,7 +356,6 @@ long p2m_set_mem_access(struct domain *d, gfn_t gfn, uint32_t nr,
long rc = 0;

/* altp2m view 0 is treated as the hostp2m */
-#ifdef CONFIG_HVM
if ( altp2m_idx )
{
if ( altp2m_idx >= min(ARRAY_SIZE(d->arch.altp2m_p2m), MAX_EPTP) ||
@@ -372,9 +365,6 @@ long p2m_set_mem_access(struct domain *d, gfn_t gfn, uint32_t nr,

ap2m = array_access_nospec(d->arch.altp2m_p2m, altp2m_idx);
}
-#else
- ASSERT(!altp2m_idx);
-#endif

if ( !xenmem_access_to_p2m_access(p2m, access, &a) )
return -EINVAL;
@@ -422,7 +412,6 @@ long p2m_set_mem_access_multi(struct domain *d,
long rc = 0;

/* altp2m view 0 is treated as the hostp2m */
-#ifdef CONFIG_HVM
if ( altp2m_idx )
{
if ( altp2m_idx >= min(ARRAY_SIZE(d->arch.altp2m_p2m), MAX_EPTP) ||
@@ -432,9 +421,6 @@ long p2m_set_mem_access_multi(struct domain *d,

ap2m = array_access_nospec(d->arch.altp2m_p2m, altp2m_idx);
}
-#else
- ASSERT(!altp2m_idx);
-#endif

p2m_lock(p2m);
if ( ap2m )
@@ -484,7 +470,6 @@ int p2m_get_mem_access(struct domain *d, gfn_t gfn, xenmem_access_t *access,
{
struct p2m_domain *p2m = p2m_get_hostp2m(d);

-#ifdef CONFIG_HVM
if ( !altp2m_active(d) )
{
if ( altp2m_idx )
@@ -499,9 +484,6 @@ int p2m_get_mem_access(struct domain *d, gfn_t gfn, xenmem_access_t *access,

p2m = array_access_nospec(d->arch.altp2m_p2m, altp2m_idx);
}
-#else
- ASSERT(!altp2m_idx);
-#endif

return _p2m_get_mem_access(p2m, gfn, access);
}
@@ -512,7 +494,6 @@ void arch_p2m_set_access_required(struct domain *d, bool access_required)

p2m_get_hostp2m(d)->access_required = access_required;

-#ifdef CONFIG_HVM
if ( altp2m_active(d) )
{
unsigned int i;
@@ -524,7 +505,6 @@ void arch_p2m_set_access_required(struct domain *d, bool access_required)
p2m->access_required = access_required;
}
}
-#endif
}

bool p2m_mem_access_sanity_check(const struct domain *d)
diff --git a/xen/arch/x86/vm_event.c b/xen/arch/x86/vm_event.c
index 8f73a73e2e..ffcc64c40b 100644
--- a/xen/arch/x86/vm_event.c
+++ b/xen/arch/x86/vm_event.c
@@ -265,6 +265,7 @@ void vm_event_emulate_check(struct vcpu *v, vm_event_response_t *rsp)
return;
}

+#ifdef CONFIG_HVM
switch ( rsp->reason )
{
case VM_EVENT_REASON_MEM_ACCESS:
@@ -298,6 +299,7 @@ void vm_event_emulate_check(struct vcpu *v, vm_event_response_t *rsp)
default:
break;
};
+#endif
}

void vm_event_reset_vmtrace(struct vcpu *v)
diff --git a/xen/common/Kconfig b/xen/common/Kconfig
index 3064bf6b89..b6c9827a7b 100644
--- a/xen/common/Kconfig
+++ b/xen/common/Kconfig
@@ -61,6 +61,7 @@ config MEM_ACCESS_ALWAYS_ON
config MEM_ACCESS
def_bool MEM_ACCESS_ALWAYS_ON
prompt "Memory Access and VM events" if !MEM_ACCESS_ALWAYS_ON
+ depends on HVM
---help---

Framework to configure memory access types for guests and receive
--
generated by git-patchbot for /home/xen/git/xen.git#staging