git: 3261dea72c24 - stable/13 - Revert "vm_pageout_scans: correct detection of active object"

From: Konstantin Belousov <kib_at_FreeBSD.org>
Date: Thu, 10 Feb 2022 14:56:33 UTC
The branch stable/13 has been updated by kib:

URL: https://cgit.FreeBSD.org/src/commit/?id=3261dea72c24aa7b33eb90aeae95d82078cfc5e4

commit 3261dea72c24aa7b33eb90aeae95d82078cfc5e4
Author:     Konstantin Belousov <kib@FreeBSD.org>
AuthorDate: 2022-02-10 14:50:42 +0000
Commit:     Konstantin Belousov <kib@FreeBSD.org>
CommitDate: 2022-02-10 14:56:15 +0000

    Revert "vm_pageout_scans: correct detection of active object"
    
    This reverts commit 3de96d664aaaf8e3fb1ca4fc4bd864d2cf734b24.
    
    PR:     261707
    
    (cherry picked from commit b51927b7b018d268c91b2127d82786caf68254de)
---
 sys/vm/vm_pageout.c | 56 +++++++++++++++++------------------------------------
 1 file changed, 18 insertions(+), 38 deletions(-)

diff --git a/sys/vm/vm_pageout.c b/sys/vm/vm_pageout.c
index 64a06dd7d55b..899d35cd43b6 100644
--- a/sys/vm/vm_pageout.c
+++ b/sys/vm/vm_pageout.c
@@ -712,38 +712,6 @@ unlock_mp:
 	return (error);
 }
 
-/*
- * Check if the object is active.  Non-anonymous swap objects are
- * always referenced by the owner, for them require ref_count > 1 in
- * order to ignore the ownership ref.
- *
- * Perform an unsynchronized object ref count check.  While
- * the page lock ensures that the page is not reallocated to
- * another object, in particular, one with unmanaged mappings
- * that cannot support pmap_ts_referenced(), two races are,
- * nonetheless, possible:
- * 1) The count was transitioning to zero, but we saw a non-
- *    zero value.  pmap_ts_referenced() will return zero
- *    because the page is not mapped.
- * 2) The count was transitioning to one, but we saw zero.
- *    This race delays the detection of a new reference.  At
- *    worst, we will deactivate and reactivate the page.
- */
-static bool
-vm_pageout_object_act(vm_object_t object)
-{
-	return (object->ref_count >
-	    ((object->flags & (OBJ_SWAP | OBJ_ANON)) == OBJ_SWAP ? 1 : 0));
-}
-
-static int
-vm_pageout_page_ts_referenced(vm_object_t object, vm_page_t m)
-{
-	if (!vm_pageout_object_act(object))
-		return (0);
-	return (pmap_ts_referenced(m));
-}
-
 /*
  * Attempt to launder the specified number of pages.
  *
@@ -838,7 +806,7 @@ scan:
 		if (vm_page_none_valid(m))
 			goto free_page;
 
-		refs = vm_pageout_page_ts_referenced(object, m);
+		refs = object->ref_count != 0 ? pmap_ts_referenced(m) : 0;
 
 		for (old = vm_page_astate_load(m);;) {
 			/*
@@ -858,7 +826,7 @@ scan:
 			}
 			if (act_delta == 0) {
 				;
-			} else if (vm_pageout_object_act(object)) {
+			} else if (object->ref_count != 0) {
 				/*
 				 * Increase the activation count if the page was
 				 * referenced while in the laundry queue.  This
@@ -1295,8 +1263,20 @@ act_scan:
 		 * Test PGA_REFERENCED after calling pmap_ts_referenced() so
 		 * that a reference from a concurrently destroyed mapping is
 		 * observed here and now.
+		 *
+		 * Perform an unsynchronized object ref count check.  While
+		 * the page lock ensures that the page is not reallocated to
+		 * another object, in particular, one with unmanaged mappings
+		 * that cannot support pmap_ts_referenced(), two races are,
+		 * nonetheless, possible:
+		 * 1) The count was transitioning to zero, but we saw a non-
+		 *    zero value.  pmap_ts_referenced() will return zero
+		 *    because the page is not mapped.
+		 * 2) The count was transitioning to one, but we saw zero.
+		 *    This race delays the detection of a new reference.  At
+		 *    worst, we will deactivate and reactivate the page.
 		 */
-		refs = vm_pageout_page_ts_referenced(object, m);
+		refs = object->ref_count != 0 ? pmap_ts_referenced(m) : 0;
 
 		old = vm_page_astate_load(m);
 		do {
@@ -1546,7 +1526,7 @@ vm_pageout_scan_inactive(struct vm_domain *vmd, int page_shortage)
 		if (vm_page_none_valid(m))
 			goto free_page;
 
-		refs = vm_pageout_page_ts_referenced(object, m);
+		refs = object->ref_count != 0 ? pmap_ts_referenced(m) : 0;
 
 		for (old = vm_page_astate_load(m);;) {
 			/*
@@ -1566,7 +1546,7 @@ vm_pageout_scan_inactive(struct vm_domain *vmd, int page_shortage)
 			}
 			if (act_delta == 0) {
 				;
-			} else if (vm_pageout_object_act(object)) {
+			} else if (object->ref_count != 0) {
 				/*
 				 * Increase the activation count if the
 				 * page was referenced while in the
@@ -1604,7 +1584,7 @@ vm_pageout_scan_inactive(struct vm_domain *vmd, int page_shortage)
 		 * mappings allow write access, then the page may still be
 		 * modified until the last of those mappings are removed.
 		 */
-		if (vm_pageout_object_act(object)) {
+		if (object->ref_count != 0) {
 			vm_page_test_dirty(m);
 			if (m->dirty == 0 && !vm_page_try_remove_all(m))
 				goto skip_page;