git: 215d2fd53f6c - main - vmm svm: Refactor per-vCPU data.
- Go to: [ bottom of page ] [ top of archives ] [ this month ]
Date: Fri, 18 Nov 2022 18:26:23 UTC
The branch main has been updated by jhb:
URL: https://cgit.FreeBSD.org/src/commit/?id=215d2fd53f6c254cb900e1775abae86d3fdada65
commit 215d2fd53f6c254cb900e1775abae86d3fdada65
Author: John Baldwin <jhb@FreeBSD.org>
AuthorDate: 2022-11-18 17:57:48 +0000
Commit: John Baldwin <jhb@FreeBSD.org>
CommitDate: 2022-11-18 18:25:35 +0000
vmm svm: Refactor per-vCPU data.
- Allocate VMCBs separately to avoid excessive padding in struct
svm_vcpu.
- Allocate APIC pages dynamically directly in struct vlapic.
- Move vm_mtrr into struct svm_vcpu rather than using a separate
parallel array.
Reviewed by: corvink, markj
Differential Revision: https://reviews.freebsd.org/D37148
---
sys/amd64/vmm/amd/svm.c | 17 +++++++++++++----
sys/amd64/vmm/amd/svm_msr.c | 4 ++--
sys/amd64/vmm/amd/svm_softc.h | 19 ++++++-------------
3 files changed, 21 insertions(+), 19 deletions(-)
diff --git a/sys/amd64/vmm/amd/svm.c b/sys/amd64/vmm/amd/svm.c
index 1a7990383b09..0db1e45594fa 100644
--- a/sys/amd64/vmm/amd/svm.c
+++ b/sys/amd64/vmm/amd/svm.c
@@ -565,8 +565,6 @@ svm_init(struct vm *vm, pmap_t pmap)
uint16_t maxcpus;
svm_sc = malloc(sizeof (*svm_sc), M_SVM, M_WAITOK | M_ZERO);
- if (((uintptr_t)svm_sc & PAGE_MASK) != 0)
- panic("malloc of svm_softc not aligned on page boundary");
svm_sc->msr_bitmap = contigmalloc(SVM_MSR_BITMAP_SIZE, M_SVM,
M_WAITOK, 0, ~(vm_paddr_t)0, PAGE_SIZE, 0);
@@ -619,9 +617,11 @@ svm_init(struct vm *vm, pmap_t pmap)
maxcpus = vm_get_maxcpus(svm_sc->vm);
for (i = 0; i < maxcpus; i++) {
vcpu = svm_get_vcpu(svm_sc, i);
+ vcpu->vmcb = malloc_aligned(sizeof(struct vmcb), PAGE_SIZE,
+ M_SVM, M_WAITOK | M_ZERO);
vcpu->nextrip = ~0;
vcpu->lastcpu = NOCPU;
- vcpu->vmcb_pa = vtophys(&vcpu->vmcb);
+ vcpu->vmcb_pa = vtophys(vcpu->vmcb);
vmcb_init(svm_sc, i, iopm_pa, msrpm_pa, pml4_pa);
svm_msr_guest_init(svm_sc, i);
}
@@ -2149,7 +2149,14 @@ static void
svm_cleanup(void *arg)
{
struct svm_softc *sc = arg;
+ struct svm_vcpu *vcpu;
+ uint16_t i, maxcpus;
+ maxcpus = vm_get_maxcpus(sc->vm);
+ for (i = 0; i < maxcpus; i++) {
+ vcpu = svm_get_vcpu(sc, i);
+ free(vcpu->vmcb, M_SVM);
+ }
contigfree(sc->iopm_bitmap, SVM_IO_BITMAP_SIZE, M_SVM);
contigfree(sc->msr_bitmap, SVM_MSR_BITMAP_SIZE, M_SVM);
free(sc, M_SVM);
@@ -2400,7 +2407,8 @@ svm_vlapic_init(void *arg, int vcpuid)
vlapic = malloc(sizeof(struct vlapic), M_SVM_VLAPIC, M_WAITOK | M_ZERO);
vlapic->vm = svm_sc->vm;
vlapic->vcpuid = vcpuid;
- vlapic->apic_page = (struct LAPIC *)&svm_sc->apic_page[vcpuid];
+ vlapic->apic_page = malloc_aligned(PAGE_SIZE, PAGE_SIZE, M_SVM_VLAPIC,
+ M_WAITOK | M_ZERO);
vlapic_init(vlapic);
@@ -2412,6 +2420,7 @@ svm_vlapic_cleanup(void *arg, struct vlapic *vlapic)
{
vlapic_cleanup(vlapic);
+ free(vlapic->apic_page, M_SVM_VLAPIC);
free(vlapic, M_SVM_VLAPIC);
}
diff --git a/sys/amd64/vmm/amd/svm_msr.c b/sys/amd64/vmm/amd/svm_msr.c
index 1a22f16cf48e..f0cea633a0cf 100644
--- a/sys/amd64/vmm/amd/svm_msr.c
+++ b/sys/amd64/vmm/amd/svm_msr.c
@@ -124,7 +124,7 @@ svm_rdmsr(struct svm_softc *sc, int vcpu, u_int num, uint64_t *result,
case MSR_MTRR16kBase ... MSR_MTRR16kBase + 1:
case MSR_MTRR64kBase:
case MSR_MTRRVarBase ... MSR_MTRRVarBase + (VMM_MTRR_VAR_MAX * 2) - 1:
- if (vm_rdmtrr(&sc->mtrr[vcpu], num, result) != 0) {
+ if (vm_rdmtrr(&sc->vcpu[vcpu].mtrr, num, result) != 0) {
vm_inject_gp(sc->vm, vcpu);
}
break;
@@ -156,7 +156,7 @@ svm_wrmsr(struct svm_softc *sc, int vcpu, u_int num, uint64_t val, bool *retu)
case MSR_MTRR16kBase ... MSR_MTRR16kBase + 1:
case MSR_MTRR64kBase:
case MSR_MTRRVarBase ... MSR_MTRRVarBase + (VMM_MTRR_VAR_MAX * 2) - 1:
- if (vm_wrmtrr(&sc->mtrr[vcpu], num, val) != 0) {
+ if (vm_wrmtrr(&sc->vcpu[vcpu].mtrr, num, val) != 0) {
vm_inject_gp(sc->vm, vcpu);
}
break;
diff --git a/sys/amd64/vmm/amd/svm_softc.h b/sys/amd64/vmm/amd/svm_softc.h
index 5f6a267617d2..b9e53ac9d4a0 100644
--- a/sys/amd64/vmm/amd/svm_softc.h
+++ b/sys/amd64/vmm/amd/svm_softc.h
@@ -41,12 +41,8 @@ struct asid {
uint32_t num; /* range is [1, nasid - 1] */
};
-/*
- * XXX separate out 'struct vmcb' from 'svm_vcpu' to avoid wasting space
- * due to VMCB alignment requirements.
- */
struct svm_vcpu {
- struct vmcb vmcb; /* hardware saved vcpu context */
+ struct vmcb *vmcb; /* hardware saved vcpu context */
struct svm_regctx swctx; /* software saved vcpu context */
uint64_t vmcb_pa; /* VMCB physical address */
uint64_t nextrip; /* next instruction to be executed by guest */
@@ -54,23 +50,20 @@ struct svm_vcpu {
uint32_t dirty; /* state cache bits that must be cleared */
long eptgen; /* pmap->pm_eptgen when the vcpu last ran */
struct asid asid;
-} __aligned(PAGE_SIZE);
+ struct vm_mtrr mtrr;
+};
/*
* SVM softc, one per virtual machine.
*/
struct svm_softc {
- uint8_t apic_page[VM_MAXCPU][PAGE_SIZE];
struct svm_vcpu vcpu[VM_MAXCPU];
vm_offset_t nptp; /* nested page table */
uint8_t *iopm_bitmap; /* shared by all vcpus */
uint8_t *msr_bitmap; /* shared by all vcpus */
struct vm *vm;
- struct vm_mtrr mtrr[VM_MAXCPU];
};
-CTASSERT((offsetof(struct svm_softc, nptp) & PAGE_MASK) == 0);
-
static __inline struct svm_vcpu *
svm_get_vcpu(struct svm_softc *sc, int vcpu)
{
@@ -82,21 +75,21 @@ static __inline struct vmcb *
svm_get_vmcb(struct svm_softc *sc, int vcpu)
{
- return (&(sc->vcpu[vcpu].vmcb));
+ return ((sc->vcpu[vcpu].vmcb));
}
static __inline struct vmcb_state *
svm_get_vmcb_state(struct svm_softc *sc, int vcpu)
{
- return (&(sc->vcpu[vcpu].vmcb.state));
+ return (&(sc->vcpu[vcpu].vmcb->state));
}
static __inline struct vmcb_ctrl *
svm_get_vmcb_ctrl(struct svm_softc *sc, int vcpu)
{
- return (&(sc->vcpu[vcpu].vmcb.ctrl));
+ return (&(sc->vcpu[vcpu].vmcb->ctrl));
}
static __inline struct svm_regctx *