git: 19a18bb750e8 - main - powerpc64: initial conversion of oea64 to rwlocks
- Go to: [ bottom of page ] [ top of archives ] [ this month ]
Date: Sun, 22 Feb 2026 02:25:43 UTC
The branch main has been updated by adrian:
URL: https://cgit.FreeBSD.org/src/commit/?id=19a18bb750e8ed88207664b4959e433b7e68e926
commit 19a18bb750e8ed88207664b4959e433b7e68e926
Author: Adrian Chadd <adrian@FreeBSD.org>
AuthorDate: 2026-02-22 02:22:46 +0000
Commit: Adrian Chadd <adrian@FreeBSD.org>
CommitDate: 2026-02-22 02:22:46 +0000
powerpc64: initial conversion of oea64 to rwlocks
Convert the oea64 mmu code to rwlocks.
Reviewed by: jhibbits
Differential Revision: https://reviews.freebsd.org/D54936
---
sys/powerpc/aim/mmu_oea64.c | 50 +++++++++++++++++++++++++++++----------------
1 file changed, 32 insertions(+), 18 deletions(-)
diff --git a/sys/powerpc/aim/mmu_oea64.c b/sys/powerpc/aim/mmu_oea64.c
index 7dfda5776a37..22442c59fd30 100644
--- a/sys/powerpc/aim/mmu_oea64.c
+++ b/sys/powerpc/aim/mmu_oea64.c
@@ -123,7 +123,7 @@ uintptr_t moea64_get_unique_vsid(void);
*/
#define PV_LOCK_COUNT MAXCPU
-static struct mtx_padalign pv_lock[PV_LOCK_COUNT];
+static struct rwlock __exclusive_cache_line pv_lock[PV_LOCK_COUNT];
#define PV_LOCK_SHIFT HPT_SP_SHIFT
#define pa_index(pa) ((pa) >> PV_LOCK_SHIFT)
@@ -138,11 +138,17 @@ static struct mtx_padalign pv_lock[PV_LOCK_COUNT];
#else
#define PV_LOCK_IDX(pa) (pa_index(pa) % PV_LOCK_COUNT)
#endif
-#define PV_LOCKPTR(pa) ((struct mtx *)(&pv_lock[PV_LOCK_IDX(pa)]))
-#define PV_LOCK(pa) mtx_lock(PV_LOCKPTR(pa))
-#define PV_UNLOCK(pa) mtx_unlock(PV_LOCKPTR(pa))
-#define PV_LOCKASSERT(pa) mtx_assert(PV_LOCKPTR(pa), MA_OWNED)
-#define PV_PAGE_LOCK(m) PV_LOCK(VM_PAGE_TO_PHYS(m))
+#define PV_LOCKPTR(pa) ((struct rwlock *)(&pv_lock[PV_LOCK_IDX(pa)]))
+
+#define PV_WR_LOCK(pa) rw_wlock(PV_LOCKPTR(pa))
+#define PV_RD_LOCK(pa) rw_rlock(PV_LOCKPTR(pa))
+#define PV_UNLOCK(pa) rw_unlock(PV_LOCKPTR(pa))
+#define PV_LOCKASSERT(pa) rw_assert(PV_LOCKPTR(pa), RA_LOCKED)
+#define PV_LOCK_RD_ASSERT(pa) rw_assert(PV_LOCKPTR(pa), RA_RLOCKED)
+#define PV_LOCK_WR_ASSERT(pa) rw_assert(PV_LOCKPTR(pa), RA_WLOCKED)
+
+#define PV_PAGE_WR_LOCK(m) PV_WR_LOCK(VM_PAGE_TO_PHYS(m))
+#define PV_PAGE_RD_LOCK(m) PV_RD_LOCK(VM_PAGE_TO_PHYS(m))
#define PV_PAGE_UNLOCK(m) PV_UNLOCK(VM_PAGE_TO_PHYS(m))
#define PV_PAGE_LOCKASSERT(m) PV_LOCKASSERT(VM_PAGE_TO_PHYS(m))
@@ -500,7 +506,7 @@ static struct pvo_head *
vm_page_to_pvoh(vm_page_t m)
{
- mtx_assert(PV_LOCKPTR(VM_PAGE_TO_PHYS(m)), MA_OWNED);
+ rw_assert(PV_LOCKPTR(VM_PAGE_TO_PHYS(m)), RA_LOCKED);
return (&m->md.mdpg_pvoh);
}
@@ -1028,7 +1034,7 @@ moea64_mid_bootstrap(vm_offset_t kernelstart, vm_offset_t kernelend)
*/
mtx_init(&moea64_slb_mutex, "SLB table", NULL, MTX_DEF);
for (i = 0; i < PV_LOCK_COUNT; i++)
- mtx_init(&pv_lock[i], "page pv", NULL, MTX_DEF);
+ rw_init(&pv_lock[i], "pv lock");
/*
* Initialise the bootstrap pvo pool.
@@ -1644,7 +1650,7 @@ moea64_enter(pmap_t pmap, vm_offset_t va, vm_page_t m,
pvo->pvo_vaddr |= PVO_MANAGED;
}
- PV_LOCK(pa);
+ PV_WR_LOCK(pa);
PMAP_LOCK(pmap);
if (pvo->pvo_pmap == NULL)
init_pvo_entry(pvo, pmap, va);
@@ -2023,7 +2029,7 @@ moea64_remove_write(vm_page_t m)
return;
powerpc_sync();
- PV_PAGE_LOCK(m);
+ PV_PAGE_WR_LOCK(m);
refchg = 0;
LIST_FOREACH(pvo, vm_page_to_pvoh(m), pvo_vlink) {
pmap = pvo->pvo_pmap;
@@ -2096,7 +2102,7 @@ moea64_page_set_memattr(vm_page_t m, vm_memattr_t ma)
lo = moea64_calc_wimg(VM_PAGE_TO_PHYS(m), ma);
- PV_PAGE_LOCK(m);
+ PV_PAGE_WR_LOCK(m);
LIST_FOREACH(pvo, vm_page_to_pvoh(m), pvo_vlink) {
pmap = pvo->pvo_pmap;
PMAP_LOCK(pmap);
@@ -2355,7 +2361,7 @@ moea64_page_exists_quick(pmap_t pmap, vm_page_t m)
("moea64_page_exists_quick: page %p is not managed", m));
loops = 0;
rv = false;
- PV_PAGE_LOCK(m);
+ PV_PAGE_RD_LOCK(m);
LIST_FOREACH(pvo, vm_page_to_pvoh(m), pvo_vlink) {
if (!(pvo->pvo_vaddr & PVO_DEAD) && pvo->pvo_pmap == pmap) {
rv = true;
@@ -2390,7 +2396,7 @@ moea64_page_wired_mappings(vm_page_t m)
count = 0;
if ((m->oflags & VPO_UNMANAGED) != 0)
return (count);
- PV_PAGE_LOCK(m);
+ PV_PAGE_RD_LOCK(m);
LIST_FOREACH(pvo, vm_page_to_pvoh(m), pvo_vlink)
if ((pvo->pvo_vaddr & (PVO_DEAD | PVO_WIRED)) == PVO_WIRED)
count++;
@@ -2760,7 +2766,7 @@ moea64_remove_all(vm_page_t m)
LIST_INIT(&freequeue);
- PV_PAGE_LOCK(m);
+ PV_PAGE_WR_LOCK(m);
LIST_FOREACH_SAFE(pvo, vm_page_to_pvoh(m), pvo_vlink, next_pvo) {
pmap = pvo->pvo_pmap;
PMAP_LOCK(pmap);
@@ -2983,7 +2989,7 @@ moea64_pvo_remove_from_page(struct pvo_entry *pvo)
if (pvo->pvo_vaddr & PVO_MANAGED)
pg = PHYS_TO_VM_PAGE(PVO_PADDR(pvo));
- PV_LOCK(PVO_PADDR(pvo));
+ PV_WR_LOCK(PVO_PADDR(pvo));
moea64_pvo_remove_from_page_locked(pvo, pg);
PV_UNLOCK(PVO_PADDR(pvo));
}
@@ -3024,7 +3030,7 @@ moea64_query_bit(vm_page_t m, uint64_t ptebit)
*/
rv = false;
powerpc_sync();
- PV_PAGE_LOCK(m);
+ PV_PAGE_RD_LOCK(m);
LIST_FOREACH(pvo, vm_page_to_pvoh(m), pvo_vlink) {
if (PVO_IS_SP(pvo)) {
ret = moea64_sp_query(pvo, ptebit);
@@ -3084,7 +3090,7 @@ moea64_clear_bit(vm_page_t m, u_int64_t ptebit)
* For each pvo entry, clear the pte's ptebit.
*/
count = 0;
- PV_PAGE_LOCK(m);
+ PV_PAGE_WR_LOCK(m);
LIST_FOREACH(pvo, vm_page_to_pvoh(m), pvo_vlink) {
if (PVO_IS_SP(pvo)) {
if ((ret = moea64_sp_clear(pvo, m, ptebit)) != -1) {
@@ -3690,7 +3696,7 @@ moea64_sp_enter(pmap_t pmap, vm_offset_t va, vm_page_t m,
}
}
- PV_LOCK(spa);
+ PV_WR_LOCK(spa);
PMAP_LOCK(pmap);
/* Note: moea64_remove_locked() also clears cached REF/CHG bits. */
@@ -4094,6 +4100,8 @@ moea64_sp_query_locked(struct pvo_entry *pvo, uint64_t ptebit)
pmap_t pmap;
struct pvo_entry *sp;
+ PV_LOCKASSERT(PVO_PADDR(pvo));
+
pmap = pvo->pvo_pmap;
PMAP_LOCK_ASSERT(pmap, MA_OWNED);
@@ -4126,12 +4134,18 @@ moea64_sp_query_locked(struct pvo_entry *pvo, uint64_t ptebit)
return (refchg);
}
+/*
+ * Note: this assumes the vm_page represented by the given pvo
+ * is at least read locked.
+ */
static int64_t
moea64_sp_query(struct pvo_entry *pvo, uint64_t ptebit)
{
int64_t refchg;
pmap_t pmap;
+ PV_LOCKASSERT(PVO_PADDR(pvo));
+
pmap = pvo->pvo_pmap;
PMAP_LOCK(pmap);