svn commit: r255669 - in projects/bhyve_npt_pmap/sys/amd64: amd64 vmm vmm/intel
Neel Natu
neel at FreeBSD.org
Wed Sep 18 17:18:20 UTC 2013
Author: neel
Date: Wed Sep 18 17:18:19 2013
New Revision: 255669
URL: http://svnweb.freebsd.org/changeset/base/255669
Log:
Make multi-line KASSERTs style(9) compliant by using 4 spaces to indent the
continuation lines.
Pointed out by: alc@
Modified:
projects/bhyve_npt_pmap/sys/amd64/amd64/pmap.c
projects/bhyve_npt_pmap/sys/amd64/vmm/intel/vmx.c
projects/bhyve_npt_pmap/sys/amd64/vmm/vmm.c
Modified: projects/bhyve_npt_pmap/sys/amd64/amd64/pmap.c
==============================================================================
--- projects/bhyve_npt_pmap/sys/amd64/amd64/pmap.c Wed Sep 18 16:39:01 2013 (r255668)
+++ projects/bhyve_npt_pmap/sys/amd64/amd64/pmap.c Wed Sep 18 17:18:19 2013 (r255669)
@@ -1064,7 +1064,7 @@ pmap_swap_pat(pmap_t pmap, pt_entry_t en
case PT_X86:
/* Verify that both PAT bits are not set at the same time */
KASSERT((entry & x86_pat_bits) != x86_pat_bits,
- ("Invalid PAT bits in entry %#lx", entry));
+ ("Invalid PAT bits in entry %#lx", entry));
/* Swap the PAT bits if one of them is set */
if ((entry & x86_pat_bits) != 0)
@@ -1278,7 +1278,7 @@ pmap_invalidate_ept(pmap_t pmap)
{
KASSERT(!CPU_ISSET(curcpu, &pmap->pm_active),
- ("pmap_invalidate_ept: absurd pm_active"));
+ ("pmap_invalidate_ept: absurd pm_active"));
/*
* The TLB mappings associated with a vcpu context are not
@@ -2738,7 +2738,8 @@ reclaim_pv_chunk(pmap_t locked_pmap, str
m = PHYS_TO_VM_PAGE(tpte & PG_FRAME);
if ((tpte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
KASSERT((tpte & PG_RO) == 0,
- ("readonly modified PTE %#lx", tpte));
+ ("readonly modified PTE %#lx",
+ tpte));
vm_page_dirty(m);
}
if ((tpte & PG_A) != 0)
@@ -3450,7 +3451,7 @@ pmap_remove_pde(pmap_t pmap, pd_entry_t
va < eva; va += PAGE_SIZE, m++) {
if ((oldpde & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
KASSERT((oldpde & PG_RO) == 0,
- ("readonly modified PDE %#lx", oldpde));
+ ("readonly modified PDE %#lx", oldpde));
vm_page_dirty(m);
}
if (oldpde & PG_A)
@@ -3500,7 +3501,7 @@ pmap_remove_pte(pmap_t pmap, pt_entry_t
m = PHYS_TO_VM_PAGE(oldpte & PG_FRAME);
if ((oldpte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
KASSERT((oldpte & PG_RO) == 0,
- ("readonly modified PTE %#lx", oldpte));
+ ("readonly modified PTE %#lx", oldpte));
vm_page_dirty(m);
}
if (oldpte & PG_A)
@@ -3750,7 +3751,7 @@ small_mappings:
*/
if ((tpte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
KASSERT((tpte & PG_RO) == 0,
- ("readonly modified PTE %#lx", tpte));
+ ("readonly modified PTE %#lx", tpte));
vm_page_dirty(m);
}
pmap_unuse_pt(pmap, pv->pv_va, *pde, &free);
@@ -3792,7 +3793,7 @@ retry:
va < eva; va += PAGE_SIZE, m++)
if ((oldpde & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
KASSERT((oldpde & PG_RO) == 0,
- ("readonly modified PDE %#lx", oldpde));
+ ("readonly modified PDE %#lx", oldpde));
vm_page_dirty(m);
}
}
@@ -3931,8 +3932,8 @@ retry:
if ((pbits & (PG_MANAGED | PG_M | PG_RW)) ==
(PG_MANAGED | PG_M | PG_RW)) {
KASSERT((pbits & PG_RO) == 0,
- ("readonly modified PTE %#lx",
- pbits));
+ ("readonly modified PTE %#lx",
+ pbits));
m = PHYS_TO_VM_PAGE(pbits & PG_FRAME);
vm_page_dirty(m);
}
@@ -4027,8 +4028,8 @@ setpte:
}
if ((oldpte & (PG_M | PG_RW)) == PG_RW) {
KASSERT(!pmap_emulate_ad_bits(pmap),
- ("invalid RW/M bits for dirty bit "
- "emulation %#lx", oldpte));
+ ("invalid RW/M bits for dirty bit emulation %#lx",
+ oldpte));
/*
* When PG_M is already clear, PG_RW can be cleared
* without a TLB invalidation.
@@ -4298,8 +4299,8 @@ validate:
if ((origpte & (PG_M | PG_RW)) == (PG_M |
PG_RW)) {
KASSERT((origpte & PG_RO) == 0,
- ("readonly modified PTE %#lx",
- origpte));
+ ("readonly modified PTE %#lx",
+ origpte));
vm_page_dirty(om);
}
if ((origpte & PG_A) != 0)
@@ -5269,8 +5270,8 @@ pmap_remove_pages(pmap_t pmap)
*/
if ((tpte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
KASSERT((tpte & PG_RO) == 0,
- ("readonly modified PTE %#lx",
- tpte));
+ ("readonly modified PTE %#lx",
+ tpte));
if (superpage) {
for (mt = m; mt < &m[NBPDR / PAGE_SIZE]; mt++)
vm_page_dirty(mt);
@@ -5558,7 +5559,7 @@ retry:
goto retry;
if ((oldpte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
KASSERT((oldpte & PG_RO) == 0,
- ("readonly modified PTE %#lx", oldpte));
+ ("readonly modified PTE %#lx", oldpte));
vm_page_dirty(m);
}
pmap_invalidate_page(pmap, pv->pv_va);
@@ -5986,8 +5987,8 @@ small_mappings:
("modified readonly pte %#lx", oldpte));
} else {
KASSERT((oldpte & (PG_M | PG_RW)) == 0,
- ("invalid RW/M bits for dirty bit "
- "emulation %#lx", oldpte));
+ ("invalid RW/M bits for dirty bit "
+ "emulation %#lx", oldpte));
}
}
if ((oldpte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
@@ -6678,8 +6679,8 @@ pmap_emulate_dirty(pmap_t pmap, vm_offse
pte = pmap_pde_to_pte(pde, va);
if ((*pte & (PG_V | PG_RO)) == PG_V) {
KASSERT((*pte & PG_A) != 0,
- ("pmap_emulate_dirty: accessed and valid bits "
- "mismatch %#lx", *pte));
+ ("pmap_emulate_dirty: accessed and valid bits "
+ "mismatch %#lx", *pte));
atomic_set_long(pte, PG_M | PG_RW);
rv = 0; /* success */
Modified: projects/bhyve_npt_pmap/sys/amd64/vmm/intel/vmx.c
==============================================================================
--- projects/bhyve_npt_pmap/sys/amd64/vmm/intel/vmx.c Wed Sep 18 16:39:01 2013 (r255668)
+++ projects/bhyve_npt_pmap/sys/amd64/vmm/intel/vmx.c Wed Sep 18 17:18:19 2013 (r255669)
@@ -1543,9 +1543,9 @@ vmx_run(void *arg, int vcpu, register_t
vmexit = vm_exitinfo(vmx->vm, vcpu);
KASSERT(vmxctx->pmap == pmap,
- ("pmap %p different than ctx pmap %p", pmap, vmxctx->pmap));
+ ("pmap %p different than ctx pmap %p", pmap, vmxctx->pmap));
KASSERT(vmxctx->eptp == vmx->eptp,
- ("eptp %p different than ctx eptp %#lx", eptp, vmxctx->eptp));
+ ("eptp %p different than ctx eptp %#lx", eptp, vmxctx->eptp));
/*
* XXX Can we avoid doing this every time we do a vm run?
Modified: projects/bhyve_npt_pmap/sys/amd64/vmm/vmm.c
==============================================================================
--- projects/bhyve_npt_pmap/sys/amd64/vmm/vmm.c Wed Sep 18 16:39:01 2013 (r255668)
+++ projects/bhyve_npt_pmap/sys/amd64/vmm/vmm.c Wed Sep 18 17:18:19 2013 (r255669)
@@ -462,8 +462,8 @@ vm_gpa_unwire(struct vm *vm)
seg->gpa, seg->gpa + seg->len,
VM_MAP_WIRE_USER | VM_MAP_WIRE_NOHOLES);
KASSERT(rv == KERN_SUCCESS, ("vm(%s) memory segment "
- "%#lx/%ld could not be unwired: %d",
- vm_name(vm), seg->gpa, seg->len, rv));
+ "%#lx/%ld could not be unwired: %d",
+ vm_name(vm), seg->gpa, seg->len, rv));
seg->wired = FALSE;
}
@@ -514,16 +514,15 @@ vm_iommu_modify(struct vm *vm, boolean_t
for (i = 0; i < vm->num_mem_segs; i++) {
seg = &vm->mem_segs[i];
- KASSERT(seg->wired,
- ("vm(%s) memory segment %#lx/%ld not wired",
- vm_name(vm), seg->gpa, seg->len));
-
+ KASSERT(seg->wired, ("vm(%s) memory segment %#lx/%ld not wired",
+ vm_name(vm), seg->gpa, seg->len));
+
gpa = seg->gpa;
while (gpa < seg->gpa + seg->len) {
vp = vm_gpa_hold(vm, gpa, PAGE_SIZE, VM_PROT_WRITE,
&cookie);
KASSERT(vp != NULL, ("vm(%s) could not map gpa %#lx",
- vm_name(vm), gpa));
+ vm_name(vm), gpa));
vm_gpa_release(cookie);
@@ -584,7 +583,7 @@ vm_assign_pptdev(struct vm *vm, int bus,
*/
if (ppt_num_devices(vm) == 0) {
KASSERT(vm->iommu == NULL,
- ("vm_assign_pptdev: iommu must be NULL"));
+ ("vm_assign_pptdev: iommu must be NULL"));
maxaddr = vmm_mem_maxaddr();
vm->iommu = iommu_create_domain(maxaddr);
@@ -906,10 +905,9 @@ vm_handle_paging(struct vm *vm, int vcpu
vme = &vcpu->exitinfo;
ftype = vme->u.paging.fault_type;
- KASSERT(ftype == VM_PROT_WRITE ||
- ftype == VM_PROT_EXECUTE ||
- ftype == VM_PROT_READ,
- ("vm_handle_paging: invalid fault_type %d", ftype));
+ KASSERT(ftype == VM_PROT_READ ||
+ ftype == VM_PROT_WRITE || ftype == VM_PROT_EXECUTE,
+ ("vm_handle_paging: invalid fault_type %d", ftype));
/*
* If the mapping exists then the write fault may be intentional
@@ -1007,7 +1005,7 @@ restart:
critical_enter();
KASSERT(!CPU_ISSET(curcpu, &pmap->pm_active),
- ("vm_run: absurd pm_active"));
+ ("vm_run: absurd pm_active"));
tscval = rdtsc();
More information about the svn-src-projects
mailing list