git: 9a6e9d779923 - main - VM_PAGE_TO_DMAP: Wrapper macro to return direct map address for a page
- Go to: [ bottom of page ] [ top of archives ] [ this month ]
Date: Thu, 23 Apr 2026 18:16:29 UTC
The branch main has been updated by jhb:
URL: https://cgit.FreeBSD.org/src/commit/?id=9a6e9d7799235d3e122bcc8065e865ae265a6ce2
commit 9a6e9d7799235d3e122bcc8065e865ae265a6ce2
Author: John Baldwin <jhb@FreeBSD.org>
AuthorDate: 2026-04-23 17:05:55 +0000
Commit: John Baldwin <jhb@FreeBSD.org>
CommitDate: 2026-04-23 17:05:55 +0000
VM_PAGE_TO_DMAP: Wrapper macro to return direct map address for a page
Effort: CHERI upstreaming
Reviewed by: kib
Sponsored by: AFRL, DARPA
Pull Request: https://github.com/freebsd/freebsd-src/pull/2068
---
sys/amd64/acpica/acpi_wakeup.c | 2 +-
sys/amd64/amd64/efirt_machdep.c | 2 +-
sys/amd64/amd64/mp_machdep.c | 12 ++++++------
sys/amd64/amd64/pmap.c | 41 ++++++++++++++++++++--------------------
sys/amd64/include/sf_buf.h | 2 +-
sys/arm64/arm64/efirt_machdep.c | 2 +-
sys/arm64/arm64/pmap.c | 24 +++++++++++------------
sys/arm64/include/sf_buf.h | 2 +-
sys/dev/hwt/hwt_vm.c | 6 +-----
sys/dev/iscsi/icl_soft.c | 4 ++--
sys/dev/vmm/vmm_mem.c | 2 +-
sys/kern/uipc_ktls.c | 2 +-
sys/opencrypto/criov.c | 22 ++++++++++-----------
sys/powerpc/aim/mmu_oea64.c | 22 +++++----------------
sys/powerpc/aim/mmu_radix.c | 42 ++++++++++++++++++-----------------------
sys/powerpc/aim/slb.c | 2 +-
sys/powerpc/booke/pmap_64.c | 20 ++++++++------------
sys/riscv/include/sf_buf.h | 2 +-
sys/riscv/riscv/pmap.c | 22 ++++++++++-----------
sys/sys/sf_buf.h | 2 +-
sys/vm/uma_core.c | 8 ++------
sys/vm/vm_page.h | 1 +
22 files changed, 105 insertions(+), 139 deletions(-)
diff --git a/sys/amd64/acpica/acpi_wakeup.c b/sys/amd64/acpica/acpi_wakeup.c
index ace559236de5..0dd670300da9 100644
--- a/sys/amd64/acpica/acpi_wakeup.c
+++ b/sys/amd64/acpica/acpi_wakeup.c
@@ -344,7 +344,7 @@ acpi_alloc_wakeup_handler(void **wakeaddr,
for (i = 0; i < ACPI_WAKEPT_PAGES - (la57 ? 0 : 1); i++) {
wakept_m[i] = pmap_page_alloc_below_4g(true);
- wakept_pages[i] = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(wakept_m[i]));
+ wakept_pages[i] = VM_PAGE_TO_DMAP(wakept_m[i]);
}
if (EVENTHANDLER_REGISTER(power_resume, acpi_stop_beep, NULL,
EVENTHANDLER_PRI_LAST) == NULL) {
diff --git a/sys/amd64/amd64/efirt_machdep.c b/sys/amd64/amd64/efirt_machdep.c
index 17088f69887d..f3e3db9cba9c 100644
--- a/sys/amd64/amd64/efirt_machdep.c
+++ b/sys/amd64/amd64/efirt_machdep.c
@@ -197,7 +197,7 @@ efi_create_1t1_map(struct efi_md *map, int ndesc, int descsz)
VM_OBJECT_WLOCK(obj_1t1_pt);
efi_pmltop_page = efi_1t1_page();
VM_OBJECT_WUNLOCK(obj_1t1_pt);
- pml = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(efi_pmltop_page));
+ pml = VM_PAGE_TO_DMAP(efi_pmltop_page);
if (la57) {
efi_pml5 = pml;
pmap_pinit_pml5(efi_pmltop_page);
diff --git a/sys/amd64/amd64/mp_machdep.c b/sys/amd64/amd64/mp_machdep.c
index 172b0b9e544e..69490b8e275f 100644
--- a/sys/amd64/amd64/mp_machdep.c
+++ b/sys/amd64/amd64/mp_machdep.c
@@ -355,29 +355,29 @@ start_all_aps(void)
/* Create a transient 1:1 mapping of low 4G */
if (la57) {
m_pml4 = pmap_page_alloc_below_4g(true);
- v_pml4 = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m_pml4));
+ v_pml4 = VM_PAGE_TO_DMAP(m_pml4);
} else {
v_pml4 = &kernel_pmap->pm_pmltop[0];
}
m_pdp = pmap_page_alloc_below_4g(true);
- v_pdp = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m_pdp));
+ v_pdp = VM_PAGE_TO_DMAP(m_pdp);
m_pd[0] = pmap_page_alloc_below_4g(false);
- v_pd = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m_pd[0]));
+ v_pd = VM_PAGE_TO_DMAP(m_pd[0]);
for (i = 0; i < NPDEPG; i++)
v_pd[i] = (i << PDRSHIFT) | X86_PG_V | X86_PG_RW | X86_PG_A |
X86_PG_M | PG_PS;
m_pd[1] = pmap_page_alloc_below_4g(false);
- v_pd = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m_pd[1]));
+ v_pd = VM_PAGE_TO_DMAP(m_pd[1]);
for (i = 0; i < NPDEPG; i++)
v_pd[i] = (NBPDP + (i << PDRSHIFT)) | X86_PG_V | X86_PG_RW |
X86_PG_A | X86_PG_M | PG_PS;
m_pd[2] = pmap_page_alloc_below_4g(false);
- v_pd = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m_pd[2]));
+ v_pd = VM_PAGE_TO_DMAP(m_pd[2]);
for (i = 0; i < NPDEPG; i++)
v_pd[i] = (2UL * NBPDP + (i << PDRSHIFT)) | X86_PG_V |
X86_PG_RW | X86_PG_A | X86_PG_M | PG_PS;
m_pd[3] = pmap_page_alloc_below_4g(false);
- v_pd = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m_pd[3]));
+ v_pd = VM_PAGE_TO_DMAP(m_pd[3]);
for (i = 0; i < NPDEPG; i++)
v_pd[i] = (3UL * NBPDP + (i << PDRSHIFT)) | X86_PG_V |
X86_PG_RW | X86_PG_A | X86_PG_M | PG_PS;
diff --git a/sys/amd64/amd64/pmap.c b/sys/amd64/amd64/pmap.c
index 6f8f767c40bd..691fac8c89b4 100644
--- a/sys/amd64/amd64/pmap.c
+++ b/sys/amd64/amd64/pmap.c
@@ -4233,7 +4233,7 @@ pmap_pinit_pml4(vm_page_t pml4pg)
pml4_entry_t *pm_pml4;
int i;
- pm_pml4 = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(pml4pg));
+ pm_pml4 = VM_PAGE_TO_DMAP(pml4pg);
/* Wire in kernel global address entries. */
for (i = 0; i < NKPML4E; i++) {
@@ -4276,7 +4276,7 @@ pmap_pinit_pml5(vm_page_t pml5pg)
pml5_entry_t *pm_pml5;
int i;
- pm_pml5 = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(pml5pg));
+ pm_pml5 = VM_PAGE_TO_DMAP(pml5pg);
for (i = 0; i < NPML5EPG / 2; i++)
pm_pml5[i] = 0;
for (; i < NPML5EPG; i++)
@@ -4289,7 +4289,7 @@ pmap_pinit_pml4_pti(vm_page_t pml4pgu)
pml4_entry_t *pm_pml4u;
int i;
- pm_pml4u = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(pml4pgu));
+ pm_pml4u = VM_PAGE_TO_DMAP(pml4pgu);
for (i = 0; i < NPML4EPG; i++)
pm_pml4u[i] = pti_pml4[i];
}
@@ -4299,7 +4299,7 @@ pmap_pinit_pml5_pti(vm_page_t pml5pgu)
{
pml5_entry_t *pm_pml5u;
- pm_pml5u = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(pml5pgu));
+ pm_pml5u = VM_PAGE_TO_DMAP(pml5pgu);
pagezero(pm_pml5u);
/*
@@ -4752,7 +4752,7 @@ retry:
else
return (NULL);
}
- pde = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(pdpg));
+ pde = VM_PAGE_TO_DMAP(pdpg);
pde = &pde[pmap_pde_index(va)];
} else
panic("pmap_alloc_pde: missing page table page for va %#lx",
@@ -7027,7 +7027,7 @@ restart:
NULL, va);
if (mp == NULL)
goto allocf;
- pdpe = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(mp));
+ pdpe = VM_PAGE_TO_DMAP(mp);
pdpe = &pdpe[pmap_pdpe_index(va)];
origpte = *pdpe;
MPASS(origpte == 0);
@@ -7048,7 +7048,7 @@ restart:
NULL, va);
if (mp == NULL)
goto allocf;
- pde = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(mp));
+ pde = VM_PAGE_TO_DMAP(mp);
pde = &pde[pmap_pde_index(va)];
origpte = *pde;
MPASS(origpte == 0);
@@ -7757,7 +7757,7 @@ pmap_enter_quick_locked(pmap_t pmap, vm_offset_t va, vm_page_t m,
return (NULL);
}
}
- pte = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(mpte));
+ pte = VM_PAGE_TO_DMAP(mpte);
pte = &pte[pmap_pte_index(va)];
} else {
mpte = NULL;
@@ -8192,7 +8192,7 @@ pmap_copy(pmap_t dst_pmap, pmap_t src_pmap, vm_offset_t dst_addr, vm_size_t len,
} else if ((dstmpte = pmap_allocpte(dst_pmap, addr,
NULL)) == NULL)
goto out;
- dst_pte = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(dstmpte));
+ dst_pte = VM_PAGE_TO_DMAP(dstmpte);
dst_pte = &dst_pte[pmap_pte_index(addr)];
if (*dst_pte == 0 &&
pmap_try_insert_pv_entry(dst_pmap, addr,
@@ -8260,7 +8260,7 @@ pmap_zero_page(vm_page_t m)
#ifdef TSLOG_PAGEZERO
TSENTER();
#endif
- va = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m));
+ va = VM_PAGE_TO_DMAP(m);
pagezero(va);
#ifdef TSLOG_PAGEZERO
TSEXIT();
@@ -8274,7 +8274,7 @@ pmap_zero_page(vm_page_t m)
void
pmap_zero_page_area(vm_page_t m, int off, int size)
{
- void *va = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m));
+ void *va = VM_PAGE_TO_DMAP(m);
if (off == 0 && size == PAGE_SIZE)
pagezero(va);
@@ -8288,8 +8288,8 @@ pmap_zero_page_area(vm_page_t m, int off, int size)
void
pmap_copy_page(vm_page_t msrc, vm_page_t mdst)
{
- void *src = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(msrc));
- void *dst = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(mdst));
+ void *src = VM_PAGE_TO_DMAP(msrc);
+ void *dst = VM_PAGE_TO_DMAP(mdst);
pagecopy(src, dst);
}
@@ -9613,8 +9613,7 @@ pmap_page_set_memattr(vm_page_t m, vm_memattr_t ma)
* required for data coherence.
*/
if ((m->flags & PG_FICTITIOUS) == 0 &&
- pmap_change_attr(PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)), PAGE_SIZE,
- m->md.pat_mode))
+ pmap_change_attr(VM_PAGE_TO_DMAP(m), PAGE_SIZE, m->md.pat_mode))
panic("memory attribute change on the direct map failed");
}
@@ -9631,8 +9630,8 @@ pmap_page_set_memattr_noflush(vm_page_t m, vm_memattr_t ma)
if ((m->flags & PG_FICTITIOUS) != 0)
return;
PMAP_LOCK(kernel_pmap);
- error = pmap_change_props_locked(PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)),
- PAGE_SIZE, PROT_NONE, m->md.pat_mode, 0);
+ error = pmap_change_props_locked(VM_PAGE_TO_DMAP(m), PAGE_SIZE,
+ PROT_NONE, m->md.pat_mode, 0);
PMAP_UNLOCK(kernel_pmap);
if (error != 0)
panic("memory attribute change on the direct map failed");
@@ -11150,7 +11149,7 @@ pmap_pti_init(void)
pti_obj = vm_pager_allocate(OBJT_PHYS, NULL, 0, VM_PROT_ALL, 0, NULL);
VM_OBJECT_WLOCK(pti_obj);
pml4_pg = pmap_pti_alloc_page();
- pti_pml4 = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(pml4_pg));
+ pti_pml4 = VM_PAGE_TO_DMAP(pml4_pg);
for (va = kva_layout.km_low; va <= kva_layout.km_high &&
va >= kva_layout.km_low && va > NBPML4; va += NBPML4) {
pdpe = pmap_pti_pdpe(va);
@@ -12279,7 +12278,7 @@ ptpages_show_pml4(vm_page_t pg4, int num_entries, uint64_t PG_V)
pd_entry_t *pd;
int i4, i3, i2;
- pml4 = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(pg4));
+ pml4 = VM_PAGE_TO_DMAP(pg4);
for (i4 = 0; i4 < num_entries; i4++) {
if ((pml4[i4] & PG_V) == 0)
continue;
@@ -12289,7 +12288,7 @@ ptpages_show_pml4(vm_page_t pg4, int num_entries, uint64_t PG_V)
continue;
}
ptpages_show_page(3, i4, pg3);
- pdp = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(pg3));
+ pdp = VM_PAGE_TO_DMAP(pg3);
for (i3 = 0; i3 < NPDPEPG; i3++) {
if ((pdp[i3] & PG_V) == 0)
continue;
@@ -12299,7 +12298,7 @@ ptpages_show_pml4(vm_page_t pg4, int num_entries, uint64_t PG_V)
continue;
}
ptpages_show_page(2, i3, pg2);
- pd = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(pg2));
+ pd = VM_PAGE_TO_DMAP(pg2);
for (i2 = 0; i2 < NPDEPG; i2++) {
if ((pd[i2] & PG_V) == 0)
continue;
diff --git a/sys/amd64/include/sf_buf.h b/sys/amd64/include/sf_buf.h
index 18f4b67a5a35..0f0d88783210 100644
--- a/sys/amd64/include/sf_buf.h
+++ b/sys/amd64/include/sf_buf.h
@@ -40,7 +40,7 @@ static inline void *
sf_buf_kva(struct sf_buf *sf)
{
- return (PHYS_TO_DMAP(VM_PAGE_TO_PHYS((vm_page_t)sf)));
+ return (VM_PAGE_TO_DMAP((vm_page_t)sf));
}
static inline vm_page_t
diff --git a/sys/arm64/arm64/efirt_machdep.c b/sys/arm64/arm64/efirt_machdep.c
index c8d23199583e..b079393cd2a9 100644
--- a/sys/arm64/arm64/efirt_machdep.c
+++ b/sys/arm64/arm64/efirt_machdep.c
@@ -179,7 +179,7 @@ efi_create_1t1_map(struct efi_md *map, int ndesc, int descsz)
VM_OBJECT_WLOCK(obj_1t1_pt);
efi_l0_page = efi_1t1_page();
VM_OBJECT_WUNLOCK(obj_1t1_pt);
- efi_l0 = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(efi_l0_page));
+ efi_l0 = VM_PAGE_TO_DMAP(efi_l0_page);
efi_ttbr0 = ASID_TO_OPERAND(ASID_RESERVED_FOR_EFI) |
VM_PAGE_TO_PHYS(efi_l0_page);
diff --git a/sys/arm64/arm64/pmap.c b/sys/arm64/arm64/pmap.c
index a0f8bbb4bed0..4ba2fb6f6678 100644
--- a/sys/arm64/arm64/pmap.c
+++ b/sys/arm64/arm64/pmap.c
@@ -3035,7 +3035,7 @@ retry:
else
return (NULL);
}
- l2 = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(l2pg));
+ l2 = VM_PAGE_TO_DMAP(l2pg);
l2 = &l2[pmap_l2_index(va)];
} else
panic("pmap_alloc_l2: missing page table page for va %#lx",
@@ -5309,7 +5309,7 @@ restart:
PMAP_LOCK(pmap);
goto restart;
}
- l2p = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(mp));
+ l2p = VM_PAGE_TO_DMAP(mp);
l2p = &l2p[pmap_l2_index(va)];
origpte = pmap_load(l2p);
} else {
@@ -5341,7 +5341,7 @@ restart:
goto restart;
}
mp->ref_count += L3C_ENTRIES - 1;
- l3p = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(mp));
+ l3p = VM_PAGE_TO_DMAP(mp);
l3p = &l3p[pmap_l3_index(va)];
} else {
l3p = pmap_l2_to_l3(l2p, va);
@@ -6112,7 +6112,7 @@ retry:
(*ml3p)->ref_count += L3C_ENTRIES - 1;
}
}
- l3p = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(*ml3p));
+ l3p = VM_PAGE_TO_DMAP(*ml3p);
} else {
*ml3p = NULL;
@@ -6368,7 +6368,7 @@ pmap_enter_quick_locked(pmap_t pmap, vm_offset_t va, vm_page_t m,
return (mpte);
}
}
- l3 = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(mpte));
+ l3 = VM_PAGE_TO_DMAP(mpte);
l3 = &l3[pmap_l3_index(va)];
} else {
mpte = NULL;
@@ -6788,7 +6788,7 @@ pmap_copy(pmap_t dst_pmap, pmap_t src_pmap, vm_offset_t dst_addr, vm_size_t len,
} else if ((dstmpte = pmap_alloc_l3(dst_pmap, addr,
NULL)) == NULL)
goto out;
- dst_pte = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(dstmpte));
+ dst_pte = VM_PAGE_TO_DMAP(dstmpte);
dst_pte = &dst_pte[pmap_l3_index(addr)];
if ((ptetemp & ATTR_CONTIGUOUS) != 0 && (addr &
L3C_OFFSET) == 0 && addr + L3C_OFFSET <=
@@ -6875,7 +6875,7 @@ pmap_vmspace_copy(pmap_t dst_pmap, pmap_t src_pmap)
void
pmap_zero_page(vm_page_t m)
{
- void *va = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m));
+ void *va = VM_PAGE_TO_DMAP(m);
pagezero(va);
}
@@ -6889,7 +6889,7 @@ pmap_zero_page(vm_page_t m)
void
pmap_zero_page_area(vm_page_t m, int off, int size)
{
- void *va = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m));
+ void *va = VM_PAGE_TO_DMAP(m);
if (off == 0 && size == PAGE_SIZE)
pagezero(va);
@@ -6906,8 +6906,8 @@ pmap_zero_page_area(vm_page_t m, int off, int size)
void
pmap_copy_page(vm_page_t msrc, vm_page_t mdst)
{
- void *src = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(msrc));
- void *dst = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(mdst));
+ void *src = VM_PAGE_TO_DMAP(msrc);
+ void *dst = VM_PAGE_TO_DMAP(mdst);
pagecopy(src, dst);
}
@@ -6954,7 +6954,7 @@ void *
pmap_quick_enter_page(vm_page_t m)
{
- return (PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)));
+ return (VM_PAGE_TO_DMAP(m));
}
void
@@ -8195,7 +8195,7 @@ pmap_page_set_memattr(vm_page_t m, vm_memattr_t ma)
* required for data coherence.
*/
if ((m->flags & PG_FICTITIOUS) == 0 &&
- pmap_change_attr(PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)), PAGE_SIZE,
+ pmap_change_attr(VM_PAGE_TO_DMAP(m), PAGE_SIZE,
m->md.pv_memattr) != 0)
panic("memory attribute change on the direct map failed");
}
diff --git a/sys/arm64/include/sf_buf.h b/sys/arm64/include/sf_buf.h
index 497789732c27..c476dc75852b 100644
--- a/sys/arm64/include/sf_buf.h
+++ b/sys/arm64/include/sf_buf.h
@@ -41,7 +41,7 @@ static inline void *
sf_buf_kva(struct sf_buf *sf)
{
- return (PHYS_TO_DMAP(VM_PAGE_TO_PHYS((vm_page_t)sf)));
+ return (VM_PAGE_TO_DMAP((vm_page_t)sf));
}
static inline vm_page_t
diff --git a/sys/dev/hwt/hwt_vm.c b/sys/dev/hwt/hwt_vm.c
index 0ea1fe44d4e3..42fc636e7abc 100644
--- a/sys/dev/hwt/hwt_vm.c
+++ b/sys/dev/hwt/hwt_vm.c
@@ -109,9 +109,6 @@ hwt_vm_alloc_pages(struct hwt_vm *vm, int kva_req)
{
vm_paddr_t low, high, boundary;
vm_memattr_t memattr;
-#ifdef __aarch64__
- void *va;
-#endif
int alignment;
vm_page_t m;
int pflags;
@@ -159,8 +156,7 @@ retry:
#endif
#ifdef __aarch64__
- va = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m));
- cpu_dcache_wb_range(va, PAGE_SIZE);
+ cpu_dcache_wb_range(VM_PAGE_TO_DMAP(m), PAGE_SIZE);
#endif
m->valid = VM_PAGE_BITS_ALL;
diff --git a/sys/dev/iscsi/icl_soft.c b/sys/dev/iscsi/icl_soft.c
index ea5b1e6c51c4..b03a88473187 100644
--- a/sys/dev/iscsi/icl_soft.c
+++ b/sys/dev/iscsi/icl_soft.c
@@ -1196,7 +1196,7 @@ icl_soft_conn_pdu_append_bio(struct icl_conn *ic, struct icl_pdu *request,
while (len > 0) {
todo = MIN(len, PAGE_SIZE - page_offset);
- vaddr = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(bp->bio_ma[i]));
+ vaddr = VM_PAGE_TO_DMAP(bp->bio_ma[i]);
do {
mtodo = min(todo, M_SIZE(m) - m->m_len);
@@ -1287,7 +1287,7 @@ icl_soft_conn_pdu_get_bio(struct icl_conn *ic, struct icl_pdu *ip,
while (len > 0) {
todo = MIN(len, PAGE_SIZE - page_offset);
- vaddr = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(bp->bio_ma[i]));
+ vaddr = VM_PAGE_TO_DMAP(bp->bio_ma[i]);
m_copydata(ip->ip_data_mbuf, pdu_off, todo, vaddr +
page_offset);
diff --git a/sys/dev/vmm/vmm_mem.c b/sys/dev/vmm/vmm_mem.c
index fc48756510ef..143020968de3 100644
--- a/sys/dev/vmm/vmm_mem.c
+++ b/sys/dev/vmm/vmm_mem.c
@@ -449,7 +449,7 @@ _vm_gpa_hold(struct vm *vm, vm_paddr_t gpa, size_t len, int reqprot,
if (count == 1) {
*cookie = m;
- return ((char *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)) + pageoff);
+ return ((char *)VM_PAGE_TO_DMAP(m) + pageoff);
} else {
*cookie = NULL;
return (NULL);
diff --git a/sys/kern/uipc_ktls.c b/sys/kern/uipc_ktls.c
index 41b18b29929b..61d1a171c467 100644
--- a/sys/kern/uipc_ktls.c
+++ b/sys/kern/uipc_ktls.c
@@ -449,7 +449,7 @@ ktls_buffer_import(void *arg, void **store, int count, int domain, int flags)
VM_MEMATTR_DEFAULT);
if (m == NULL)
break;
- store[i] = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m));
+ store[i] = VM_PAGE_TO_DMAP(m);
}
return (i);
}
diff --git a/sys/opencrypto/criov.c b/sys/opencrypto/criov.c
index ffe13bcca5bb..a77543485a8f 100644
--- a/sys/opencrypto/criov.c
+++ b/sys/opencrypto/criov.c
@@ -162,7 +162,7 @@ cvm_page_apply(vm_page_t *pages, int off, int len,
processed = 0;
CVM_PAGE_SKIP();
while (len > 0) {
- char *kaddr = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(*pages));
+ char *kaddr = VM_PAGE_TO_DMAP(*pages);
count = min(PAGE_SIZE - off, len);
rval = (*f)(arg, kaddr + off, count);
if (rval)
@@ -183,7 +183,7 @@ cvm_page_contiguous_segment(vm_page_t *pages, size_t skip, int len)
pages += (skip / PAGE_SIZE);
skip -= rounddown(skip, PAGE_SIZE);
- return (((char *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(*pages))) + skip);
+ return ((char *)VM_PAGE_TO_DMAP(*pages) + skip);
}
/*
@@ -200,8 +200,7 @@ cvm_page_copyback(vm_page_t *pages, int off, int len, c_caddr_t cp)
CVM_PAGE_SKIP();
while (len > 0) {
count = min(PAGE_SIZE - off, len);
- bcopy(cp, (char *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(*pages)) + off,
- count);
+ bcopy(cp, (char *)VM_PAGE_TO_DMAP(*pages) + off, count);
len -= count;
cp += count;
processed += count;
@@ -225,8 +224,7 @@ cvm_page_copydata(vm_page_t *pages, int off, int len, caddr_t cp)
CVM_PAGE_SKIP();
while (len > 0) {
count = min(PAGE_SIZE - off, len);
- bcopy(((char *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(*pages)) + off), cp,
- count);
+ bcopy((char *)VM_PAGE_TO_DMAP(*pages) + off, cp, count);
len -= count;
cp += count;
processed += count;
@@ -439,8 +437,8 @@ crypto_cursor_segment(struct crypto_buffer_cursor *cc, size_t *len)
return (mtod(cc->cc_mbuf, char *) + cc->cc_offset);
case CRYPTO_BUF_VMPAGE:
*len = PAGE_SIZE - cc->cc_offset;
- return ((char *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(
- *cc->cc_vmpage)) + cc->cc_offset);
+ return ((char *)VM_PAGE_TO_DMAP(*cc->cc_vmpage) +
+ cc->cc_offset);
case CRYPTO_BUF_UIO:
*len = cc->cc_iov->iov_len - cc->cc_offset;
return ((char *)cc->cc_iov->iov_base + cc->cc_offset);
@@ -494,8 +492,8 @@ crypto_cursor_copyback(struct crypto_buffer_cursor *cc, int size,
break;
case CRYPTO_BUF_VMPAGE:
for (;;) {
- dst = (char *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(
- *cc->cc_vmpage)) + cc->cc_offset;
+ dst = (char *)VM_PAGE_TO_DMAP(*cc->cc_vmpage) +
+ cc->cc_offset;
remain = MIN(PAGE_SIZE - cc->cc_offset, cc->cc_buf_len);
todo = MIN(remain, size);
memcpy(dst, src, todo);
@@ -583,8 +581,8 @@ crypto_cursor_copydata(struct crypto_buffer_cursor *cc, int size, void *vdst)
break;
case CRYPTO_BUF_VMPAGE:
for (;;) {
- src = (char *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(
- *cc->cc_vmpage)) + cc->cc_offset;
+ src = (char *)VM_PAGE_TO_DMAP(*cc->cc_vmpage) +
+ cc->cc_offset;
remain = MIN(PAGE_SIZE - cc->cc_offset, cc->cc_buf_len);
todo = MIN(remain, size);
memcpy(dst, src, todo);
diff --git a/sys/powerpc/aim/mmu_oea64.c b/sys/powerpc/aim/mmu_oea64.c
index 93787d31256b..d8e85a72f4e0 100644
--- a/sys/powerpc/aim/mmu_oea64.c
+++ b/sys/powerpc/aim/mmu_oea64.c
@@ -1446,13 +1446,7 @@ moea64_copy_page(vm_page_t msrc, vm_page_t mdst)
void
moea64_copy_page_dmap(vm_page_t msrc, vm_page_t mdst)
{
- vm_paddr_t dst;
- vm_paddr_t src;
-
- dst = VM_PAGE_TO_PHYS(mdst);
- src = VM_PAGE_TO_PHYS(msrc);
-
- bcopy(PHYS_TO_DMAP(src), PHYS_TO_DMAP(dst), PAGE_SIZE);
+ bcopy(VM_PAGE_TO_DMAP(msrc), VM_PAGE_TO_DMAP(mdst), PAGE_SIZE);
}
inline void
@@ -1466,13 +1460,11 @@ moea64_copy_pages_dmap(vm_page_t *ma, vm_offset_t a_offset,
while (xfersize > 0) {
a_pg_offset = a_offset & PAGE_MASK;
cnt = min(xfersize, PAGE_SIZE - a_pg_offset);
- a_cp = (char *)PHYS_TO_DMAP(
- VM_PAGE_TO_PHYS(ma[a_offset >> PAGE_SHIFT])) +
+ a_cp = (char *)VM_PAGE_TO_DMAP(ma[a_offset >> PAGE_SHIFT]) +
a_pg_offset;
b_pg_offset = b_offset & PAGE_MASK;
cnt = min(cnt, PAGE_SIZE - b_pg_offset);
- b_cp = (char *)PHYS_TO_DMAP(
- VM_PAGE_TO_PHYS(mb[b_offset >> PAGE_SHIFT])) +
+ b_cp = (char *)VM_PAGE_TO_DMAP(mb[b_offset >> PAGE_SHIFT]) +
b_pg_offset;
bcopy(a_cp, b_cp, cnt);
a_offset += cnt;
@@ -1549,11 +1541,7 @@ moea64_zero_page(vm_page_t m)
void
moea64_zero_page_dmap(vm_page_t m)
{
- vm_paddr_t pa = VM_PAGE_TO_PHYS(m);
- void *va;
-
- va = PHYS_TO_DMAP(pa);
- bzero(va, PAGE_SIZE);
+ bzero(VM_PAGE_TO_DMAP(m), PAGE_SIZE);
}
void *
@@ -1586,7 +1574,7 @@ void *
moea64_quick_enter_page_dmap(vm_page_t m)
{
- return (PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)));
+ return (VM_PAGE_TO_DMAP(m));
}
void
diff --git a/sys/powerpc/aim/mmu_radix.c b/sys/powerpc/aim/mmu_radix.c
index dd3cb5128fd2..7df54408838e 100644
--- a/sys/powerpc/aim/mmu_radix.c
+++ b/sys/powerpc/aim/mmu_radix.c
@@ -1224,7 +1224,7 @@ retry:
PV_STAT(atomic_add_int(&pc_chunk_count, 1));
PV_STAT(atomic_add_int(&pc_chunk_allocs, 1));
dump_add_page(m->phys_addr);
- pc = PHYS_TO_DMAP(m->phys_addr);
+ pc = VM_PAGE_TO_DMAP(m);
pc->pc_pmap = pmap;
pc->pc_map[0] = PC_FREE0;
pc->pc_map[1] = PC_FREE1;
@@ -1649,7 +1649,7 @@ retry:
PV_STAT(atomic_add_int(&pc_chunk_count, 1));
PV_STAT(atomic_add_int(&pc_chunk_allocs, 1));
dump_add_page(m->phys_addr);
- pc = PHYS_TO_DMAP(m->phys_addr);
+ pc = VM_PAGE_TO_DMAP(m);
pc->pc_pmap = pmap;
pc->pc_map[0] = PC_FREE0 & ~1ul; /* preallocated bit 0 */
pc->pc_map[1] = PC_FREE1;
@@ -2577,7 +2577,7 @@ mmu_radix_copy(pmap_t dst_pmap, pmap_t src_pmap, vm_offset_t dst_addr,
dst_pdpg = pmap_allocl3e(dst_pmap, addr, NULL);
if (dst_pdpg == NULL)
break;
- l3e = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(dst_pdpg));
+ l3e = VM_PAGE_TO_DMAP(dst_pdpg);
l3e = &l3e[pmap_pml3e_index(addr)];
if (be64toh(*l3e) == 0 && ((srcptepaddr & PG_MANAGED) == 0 ||
pmap_pv_insert_l3e(dst_pmap, addr, srcptepaddr,
@@ -2615,8 +2615,7 @@ mmu_radix_copy(pmap_t dst_pmap, pmap_t src_pmap, vm_offset_t dst_addr,
else if ((dstmpte = pmap_allocpte(dst_pmap,
addr, NULL)) == NULL)
goto out;
- dst_pte =
- PHYS_TO_DMAP(VM_PAGE_TO_PHYS(dstmpte));
+ dst_pte = VM_PAGE_TO_DMAP(dstmpte);
dst_pte = &dst_pte[pmap_pte_index(addr)];
if (be64toh(*dst_pte) == 0 &&
pmap_try_insert_pv_entry(dst_pmap, addr,
@@ -2670,8 +2669,8 @@ out:
static void
mmu_radix_copy_page(vm_page_t msrc, vm_page_t mdst)
{
- void *src = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(msrc));
- void *dst = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(mdst));
+ void *src = VM_PAGE_TO_DMAP(msrc);
+ void *dst = VM_PAGE_TO_DMAP(mdst);
CTR3(KTR_PMAP, "%s(%p, %p)", __func__, src, dst);
/*
@@ -2694,13 +2693,11 @@ mmu_radix_copy_pages(vm_page_t ma[], vm_offset_t a_offset, vm_page_t mb[],
while (xfersize > 0) {
a_pg_offset = a_offset & PAGE_MASK;
cnt = min(xfersize, PAGE_SIZE - a_pg_offset);
- a_cp = (char *)PHYS_TO_DMAP(
- VM_PAGE_TO_PHYS(ma[a_offset >> PAGE_SHIFT])) +
+ a_cp = (char *)VM_PAGE_TO_DMAP(ma[a_offset >> PAGE_SHIFT]) +
a_pg_offset;
b_pg_offset = b_offset & PAGE_MASK;
cnt = min(cnt, PAGE_SIZE - b_pg_offset);
- b_cp = (char *)PHYS_TO_DMAP(
- VM_PAGE_TO_PHYS(mb[b_offset >> PAGE_SHIFT])) +
+ b_cp = (char *)VM_PAGE_TO_DMAP(mb[b_offset >> PAGE_SHIFT]) +
b_pg_offset;
bcopy(a_cp, b_cp, cnt);
a_offset += cnt;
@@ -3216,7 +3213,7 @@ pmap_enter_l3e(pmap_t pmap, vm_offset_t va, pml3_entry_t newpde, u_int flags,
" in pmap %p", va, pmap);
return (KERN_RESOURCE_SHORTAGE);
}
- l3e = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(pdpg));
+ l3e = VM_PAGE_TO_DMAP(pdpg);
l3e = &l3e[pmap_pml3e_index(va)];
oldl3e = be64toh(*l3e);
if ((oldl3e & PG_V) != 0) {
@@ -3279,7 +3276,7 @@ pmap_enter_l3e(pmap_t pmap, vm_offset_t va, pml3_entry_t newpde, u_int flags,
}
pmap_resident_count_inc(pmap, 1);
uwptpg->ref_count = NPTEPG;
- pmap_fill_ptp(PHYS_TO_DMAP(VM_PAGE_TO_PHYS(uwptpg)), newpde);
+ pmap_fill_ptp(VM_PAGE_TO_DMAP(uwptpg), newpde);
}
if ((newpde & PG_MANAGED) != 0) {
/*
@@ -3422,7 +3419,7 @@ mmu_radix_enter_quick_locked(pmap_t pmap, vm_offset_t va, vm_page_t m,
return (mpte);
}
}
- pte = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(mpte));
+ pte = VM_PAGE_TO_DMAP(mpte);
pte = &pte[pmap_pte_index(va)];
} else {
mpte = NULL;
@@ -3629,7 +3626,7 @@ radix_pgd_import(void *arg __unused, void **store, int count, int domain __unuse
RADIX_PGD_SIZE / PAGE_SIZE,
0, (vm_paddr_t)-1, RADIX_PGD_SIZE, L1_PAGE_SIZE,
VM_MEMATTR_DEFAULT);
- store[i] = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m));
+ store[i] = VM_PAGE_TO_DMAP(m);
}
return (count);
}
@@ -4104,7 +4101,7 @@ mmu_radix_object_init_pt(pmap_t pmap, vm_offset_t addr,
addr += L3_PAGE_SIZE;
continue;
}
- l3e = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(pdpg));
+ l3e = VM_PAGE_TO_DMAP(pdpg);
l3e = &l3e[pmap_pml3e_index(addr)];
if ((be64toh(*l3e) & PG_V) == 0) {
pa |= PG_M | PG_A | PG_RW;
@@ -5783,7 +5780,7 @@ mmu_radix_zero_page(vm_page_t m)
void *addr;
CTR2(KTR_PMAP, "%s(%p)", __func__, m);
- addr = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m));
+ addr = VM_PAGE_TO_DMAP(m);
pagezero(addr);
}
@@ -5794,7 +5791,7 @@ mmu_radix_zero_page_area(vm_page_t m, int off, int size)
CTR4(KTR_PMAP, "%s(%p, %d, %d)", __func__, m, off, size);
MPASS(off + size <= PAGE_SIZE);
- addr = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m));
+ addr = VM_PAGE_TO_DMAP(m);
memset(addr + off, 0, size);
}
@@ -5945,8 +5942,8 @@ mmu_radix_page_set_memattr(vm_page_t m, vm_memattr_t ma)
* required for data coherence.
*/
if ((m->flags & PG_FICTITIOUS) == 0 &&
- mmu_radix_change_attr(PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)),
- PAGE_SIZE, m->md.mdpg_cache_attrs))
+ mmu_radix_change_attr(VM_PAGE_TO_DMAP(m), PAGE_SIZE,
+ m->md.mdpg_cache_attrs))
panic("memory attribute change on the direct map failed");
}
@@ -6178,11 +6175,8 @@ mmu_radix_dumpsys_map(vm_paddr_t pa, size_t sz,
void *
mmu_radix_quick_enter_page(vm_page_t m)
{
- vm_paddr_t paddr;
-
CTR2(KTR_PMAP, "%s(%p)", __func__, m);
- paddr = VM_PAGE_TO_PHYS(m);
- return (PHYS_TO_DMAP(paddr));
+ return (VM_PAGE_TO_DMAP(m));
}
void
diff --git a/sys/powerpc/aim/slb.c b/sys/powerpc/aim/slb.c
index 00e3577001cd..84527cf2b54e 100644
--- a/sys/powerpc/aim/slb.c
+++ b/sys/powerpc/aim/slb.c
@@ -505,7 +505,7 @@ slb_uma_real_alloc(uma_zone_t zone, vm_size_t bytes, int domain,
return (NULL);
if (hw_direct_map)
- va = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m));
+ va = VM_PAGE_TO_DMAP(m);
else {
va = (void *)(VM_PAGE_TO_PHYS(m) | DMAP_BASE_ADDRESS);
pmap_kenter((vm_offset_t)va, VM_PAGE_TO_PHYS(m));
diff --git a/sys/powerpc/booke/pmap_64.c b/sys/powerpc/booke/pmap_64.c
index 08449b9407ae..97f9a5967167 100644
--- a/sys/powerpc/booke/pmap_64.c
+++ b/sys/powerpc/booke/pmap_64.c
@@ -167,7 +167,7 @@ mmu_booke_alloc_page(pmap_t pmap, unsigned int idx, bool nosleep)
}
m->pindex = idx;
- return (PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)));
+ return (VM_PAGE_TO_DMAP(m));
}
/* Initialize pool of kva ptbl buffers. */
@@ -669,7 +669,7 @@ mmu_booke_zero_page_area(vm_page_t m, int off, int size)
/* XXX KASSERT off and size are within a single page? */
- va = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m));
+ va = VM_PAGE_TO_DMAP(m);
bzero(va + off, size);
}
@@ -679,11 +679,7 @@ mmu_booke_zero_page_area(vm_page_t m, int off, int size)
static void
mmu_booke_zero_page(vm_page_t m)
{
- void *va;
-
- va = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m));
-
- bzero(va, PAGE_SIZE);
+ bzero(VM_PAGE_TO_DMAP(m), PAGE_SIZE);
}
/*
@@ -696,8 +692,8 @@ mmu_booke_copy_page(vm_page_t sm, vm_page_t dm)
{
void *sva, *dva;
- sva = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(sm));
- dva = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(dm));
+ sva = VM_PAGE_TO_DMAP(sm);
+ dva = VM_PAGE_TO_DMAP(dm);
memcpy(dva, sva, PAGE_SIZE);
}
@@ -718,8 +714,8 @@ mmu_booke_copy_pages(vm_page_t *ma, vm_offset_t a_offset,
pb = mb[b_offset >> PAGE_SHIFT];
cnt = min(xfersize, PAGE_SIZE - a_pg_offset);
cnt = min(cnt, PAGE_SIZE - b_pg_offset);
- a_cp = (caddr_t)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(pa)) + a_pg_offset;
- b_cp = (caddr_t)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(pb)) + b_pg_offset;
+ a_cp = (caddr_t)VM_PAGE_TO_DMAP(pa) + a_pg_offset;
+ b_cp = (caddr_t)VM_PAGE_TO_DMAP(pb) + b_pg_offset;
bcopy(a_cp, b_cp, cnt);
a_offset += cnt;
b_offset += cnt;
@@ -730,7 +726,7 @@ mmu_booke_copy_pages(vm_page_t *ma, vm_offset_t a_offset,
static void *
mmu_booke_quick_enter_page(vm_page_t m)
{
- return (PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)));
+ return (VM_PAGE_TO_DMAP(m));
}
static void
diff --git a/sys/riscv/include/sf_buf.h b/sys/riscv/include/sf_buf.h
index 1ea67076e676..d00833238417 100644
--- a/sys/riscv/include/sf_buf.h
+++ b/sys/riscv/include/sf_buf.h
@@ -37,7 +37,7 @@ static inline void *
sf_buf_kva(struct sf_buf *sf)
{
- return (PHYS_TO_DMAP(VM_PAGE_TO_PHYS((vm_page_t)sf)));
+ return (VM_PAGE_TO_DMAP((vm_page_t)sf));
}
static inline vm_page_t
diff --git a/sys/riscv/riscv/pmap.c b/sys/riscv/riscv/pmap.c
index 0cac747334a9..1f3fd8c24b81 100644
--- a/sys/riscv/riscv/pmap.c
+++ b/sys/riscv/riscv/pmap.c
@@ -2194,7 +2194,7 @@ retry:
PV_STAT(atomic_add_int(&pc_chunk_count, 1));
PV_STAT(atomic_add_int(&pc_chunk_allocs, 1));
dump_add_page(m->phys_addr);
- pc = PHYS_TO_DMAP(m->phys_addr);
+ pc = VM_PAGE_TO_DMAP(m);
pc->pc_pmap = pmap;
pc->pc_map[0] = PC_FREEN & ~1ul; /* preallocated bit 0 */
pc->pc_map[1] = PC_FREEN;
@@ -2257,7 +2257,7 @@ retry:
PV_STAT(atomic_add_int(&pc_chunk_count, 1));
PV_STAT(atomic_add_int(&pc_chunk_allocs, 1));
dump_add_page(m->phys_addr);
- pc = PHYS_TO_DMAP(m->phys_addr);
+ pc = VM_PAGE_TO_DMAP(m);
pc->pc_pmap = pmap;
pc->pc_map[0] = PC_FREEN;
pc->pc_map[1] = PC_FREEN;
@@ -3585,7 +3585,7 @@ pmap_enter_l2(pmap_t pmap, vm_offset_t va, pd_entry_t new_l2, u_int flags,
return (KERN_RESOURCE_SHORTAGE);
}
- l2 = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(l2pg));
+ l2 = VM_PAGE_TO_DMAP(l2pg);
l2 = &l2[pmap_l2_index(va)];
if ((oldl2 = pmap_load(l2)) != 0) {
KASSERT(l2pg->ref_count > 1,
@@ -3827,7 +3827,7 @@ pmap_enter_quick_locked(pmap_t pmap, vm_offset_t va, vm_page_t m,
return (mpte);
}
}
- l3 = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(mpte));
+ l3 = VM_PAGE_TO_DMAP(mpte);
l3 = &l3[pmap_l3_index(va)];
} else {
mpte = NULL;
@@ -4032,9 +4032,7 @@ pmap_copy(pmap_t dst_pmap, pmap_t src_pmap, vm_offset_t dst_addr, vm_size_t len,
void
pmap_zero_page(vm_page_t m)
{
- void *va = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m));
-
- pagezero(va);
+ pagezero(VM_PAGE_TO_DMAP(m));
}
/*
@@ -4046,7 +4044,7 @@ pmap_zero_page(vm_page_t m)
void
pmap_zero_page_area(vm_page_t m, int off, int size)
{
- void *va = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m));
+ void *va = VM_PAGE_TO_DMAP(m);
if (off == 0 && size == PAGE_SIZE)
pagezero(va);
@@ -4063,8 +4061,8 @@ pmap_zero_page_area(vm_page_t m, int off, int size)
void
pmap_copy_page(vm_page_t msrc, vm_page_t mdst)
{
- void *src = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(msrc));
- void *dst = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(mdst));
+ void *src = VM_PAGE_TO_DMAP(msrc);
+ void *dst = VM_PAGE_TO_DMAP(mdst);
pagecopy(src, dst);
}
@@ -4111,7 +4109,7 @@ void *
pmap_quick_enter_page(vm_page_t m)
{
- return (PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)));
+ return (VM_PAGE_TO_DMAP(m));
}
void
@@ -4928,7 +4926,7 @@ pmap_page_set_memattr(vm_page_t m, vm_memattr_t ma)
* required for data coherence.
*/
if ((m->flags & PG_FICTITIOUS) == 0 &&
- pmap_change_attr(PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)), PAGE_SIZE,
+ pmap_change_attr(VM_PAGE_TO_DMAP(m), PAGE_SIZE,
m->md.pv_memattr) != 0)
panic("memory attribute change on the direct map failed");
}
diff --git a/sys/sys/sf_buf.h b/sys/sys/sf_buf.h
index b03615237110..ef4892e72b73 100644
--- a/sys/sys/sf_buf.h
+++ b/sys/sys/sf_buf.h
@@ -114,7 +114,7 @@ static inline void *
sf_buf_kva(struct sf_buf *sf)
{
if (PMAP_HAS_DMAP)
- return (PHYS_TO_DMAP(VM_PAGE_TO_PHYS((vm_page_t)sf)));
+ return (VM_PAGE_TO_DMAP((vm_page_t)sf));
return ((void *)sf->kva);
}
diff --git a/sys/vm/uma_core.c b/sys/vm/uma_core.c
index 8070db40c06d..410b4f5a5142 100644
--- a/sys/vm/uma_core.c
+++ b/sys/vm/uma_core.c
@@ -2083,19 +2083,15 @@ uma_small_alloc(uma_zone_t zone, vm_size_t bytes, int domain, uint8_t *flags,
int wait)
{
vm_page_t m;
- vm_paddr_t pa;
- void *va;
*flags = UMA_SLAB_PRIV;
*** 26 LINES SKIPPED ***