git: 5d32157d4e5c - main - vm_object: Modify vm_object_allocate_anon() to return OBJT_SWAP objects
- Go to: [ bottom of page ] [ top of archives ] [ this month ]
Date: Sun, 17 Jul 2022 11:24:15 UTC
The branch main has been updated by markj:
URL: https://cgit.FreeBSD.org/src/commit/?id=5d32157d4e5c2dfda5ca39d40bf4620b04fc2b29
commit 5d32157d4e5c2dfda5ca39d40bf4620b04fc2b29
Author: Mark Johnston <markj@FreeBSD.org>
AuthorDate: 2022-07-16 15:27:27 +0000
Commit: Mark Johnston <markj@FreeBSD.org>
CommitDate: 2022-07-17 11:09:48 +0000
vm_object: Modify vm_object_allocate_anon() to return OBJT_SWAP objects
With this change, OBJT_DEFAULT objects are no longer allocated.
Instead, anonymous objects are always of type OBJT_SWAP and always have
OBJ_SWAP set.
Modify the page fault handler to check the swap block radix tree in
places where it checked for objects of type OBJT_DEFAULT. In
particular, there's no need to invoke getpages for an OBJT_SWAP object
with no swap blocks assigned.
Reviewed by: alc, kib
Tested by: pho
Sponsored by: The FreeBSD Foundation
Differential Revision: https://reviews.freebsd.org/D35785
---
sys/vm/vm_fault.c | 25 ++++++++++++++++++++-----
sys/vm/vm_object.c | 8 +++++---
sys/vm/vm_object.h | 3 ++-
3 files changed, 27 insertions(+), 9 deletions(-)
diff --git a/sys/vm/vm_fault.c b/sys/vm/vm_fault.c
index c379301f866f..d398be941e84 100644
--- a/sys/vm/vm_fault.c
+++ b/sys/vm/vm_fault.c
@@ -85,6 +85,7 @@ __FBSDID("$FreeBSD$");
#include <sys/lock.h>
#include <sys/mman.h>
#include <sys/mutex.h>
+#include <sys/pctrie.h>
#include <sys/proc.h>
#include <sys/racct.h>
#include <sys/refcount.h>
@@ -220,6 +221,21 @@ fault_page_free(vm_page_t *mp)
}
}
+/*
+ * Return true if a vm_pager_get_pages() call is needed in order to check
+ * whether the pager might have a particular page, false if it can be determined
+ * immediately that the pager can not have a copy. For swap objects, this can
+ * be checked quickly.
+ */
+static inline bool
+fault_object_needs_getpages(vm_object_t object)
+{
+ VM_OBJECT_ASSERT_LOCKED(object);
+
+ return ((object->flags & OBJ_SWAP) == 0 ||
+ !pctrie_is_empty(&object->un_pager.swp.swp_blks));
+}
+
static inline void
unlock_map(struct faultstate *fs)
{
@@ -1406,10 +1422,9 @@ vm_fault_object(struct faultstate *fs, int *behindp, int *aheadp)
/*
* Page is not resident. If the pager might contain the page
* or this is the beginning of the search, allocate a new
- * page. (Default objects are zero-fill, so there is no real
- * pager for them.)
+ * page.
*/
- if (fs->m == NULL && (fs->object->type != OBJT_DEFAULT ||
+ if (fs->m == NULL && (fault_object_needs_getpages(fs->object) ||
fs->object == fs->first_object)) {
res = vm_fault_allocate(fs);
if (res != FAULT_CONTINUE)
@@ -1422,7 +1437,7 @@ vm_fault_object(struct faultstate *fs, int *behindp, int *aheadp)
* object without dropping the lock to preserve atomicity of
* shadow faults.
*/
- if (fs->object->type != OBJT_DEFAULT) {
+ if (fault_object_needs_getpages(fs->object)) {
/*
* At this point, we have either allocated a new page
* or found an existing page that is only partially
@@ -1841,7 +1856,7 @@ vm_fault_prefault(const struct faultstate *fs, vm_offset_t addra,
if (!obj_locked)
VM_OBJECT_RLOCK(lobject);
while ((m = vm_page_lookup(lobject, pindex)) == NULL &&
- lobject->type == OBJT_DEFAULT &&
+ !fault_object_needs_getpages(lobject) &&
(backing_object = lobject->backing_object) != NULL) {
KASSERT((lobject->backing_object_offset & PAGE_MASK) ==
0, ("vm_fault_prefault: unaligned object offset"));
diff --git a/sys/vm/vm_object.c b/sys/vm/vm_object.c
index 4272a734b233..57904454eff7 100644
--- a/sys/vm/vm_object.c
+++ b/sys/vm/vm_object.c
@@ -244,8 +244,10 @@ _vm_object_allocate(objtype_t type, vm_pindex_t size, u_short flags,
object->type = type;
object->flags = flags;
- if ((flags & OBJ_SWAP) != 0)
+ if ((flags & OBJ_SWAP) != 0) {
pctrie_init(&object->un_pager.swp.swp_blks);
+ object->un_pager.swp.writemappings = 0;
+ }
/*
* Ensure that swap_pager_swapoff() iteration over object_list
@@ -473,8 +475,8 @@ vm_object_allocate_anon(vm_pindex_t size, vm_object_t backing_object,
else
handle = backing_object;
object = uma_zalloc(obj_zone, M_WAITOK);
- _vm_object_allocate(OBJT_DEFAULT, size, OBJ_ANON | OBJ_ONEMAPPING,
- object, handle);
+ _vm_object_allocate(OBJT_SWAP, size,
+ OBJ_ANON | OBJ_ONEMAPPING | OBJ_SWAP, object, handle);
object->cred = cred;
object->charge = cred != NULL ? charge : 0;
return (object);
diff --git a/sys/vm/vm_object.h b/sys/vm/vm_object.h
index 80e84dcabf98..719198637726 100644
--- a/sys/vm/vm_object.h
+++ b/sys/vm/vm_object.h
@@ -202,7 +202,8 @@ struct vm_object {
#define OBJ_SIZEVNLOCK 0x0040 /* lock vnode to check obj size */
#define OBJ_PG_DTOR 0x0080 /* dont reset object, leave that for dtor */
#define OBJ_SHADOWLIST 0x0100 /* Object is on the shadow list. */
-#define OBJ_SWAP 0x0200 /* object swaps */
+#define OBJ_SWAP 0x0200 /* object swaps, type will be OBJT_SWAP
+ or dynamically registered */
#define OBJ_SPLIT 0x0400 /* object is being split */
#define OBJ_COLLAPSING 0x0800 /* Parent of collapse. */
#define OBJ_COLORED 0x1000 /* pg_color is defined */