git: 490b09f24006 - main - uma_zalloc_domain: call uma_zalloc_debug in multi-domain path
- Go to: [ bottom of page ] [ top of archives ] [ this month ]
Date: Sat, 26 Mar 2022 01:11:47 UTC
The branch main has been updated by vangyzen:
URL: https://cgit.FreeBSD.org/src/commit/?id=490b09f240065d7ef61f68ec1bf134d729cfad28
commit 490b09f240065d7ef61f68ec1bf134d729cfad28
Author: Eric van Gyzen <vangyzen@FreeBSD.org>
AuthorDate: 2022-03-07 17:12:15 +0000
Commit: Eric van Gyzen <vangyzen@FreeBSD.org>
CommitDate: 2022-03-26 01:10:38 +0000
uma_zalloc_domain: call uma_zalloc_debug in multi-domain path
It was only called in the non-NUMA and single-domain paths.
Some of its assertions were duplicated in uma_zalloc_domain,
but some things were missed, especially memguard.
Reviewed by: markj, rstone
MFC after: 1 week
Sponsored by: Dell EMC Isilon
Differential Revision: https://reviews.freebsd.org/D34472
---
sys/vm/uma_core.c | 11 +++++------
1 file changed, 5 insertions(+), 6 deletions(-)
diff --git a/sys/vm/uma_core.c b/sys/vm/uma_core.c
index 9fd2fd5e5d03..77182e460a54 100644
--- a/sys/vm/uma_core.c
+++ b/sys/vm/uma_core.c
@@ -3816,12 +3816,6 @@ uma_zalloc_domain(uma_zone_t zone, void *udata, int domain, int flags)
CTR4(KTR_UMA, "uma_zalloc_domain zone %s(%p) domain %d flags %d",
zone->uz_name, zone, domain, flags);
- if (flags & M_WAITOK) {
- WITNESS_WARN(WARN_GIANTOK | WARN_SLEEPOK, NULL,
- "uma_zalloc_domain: zone \"%s\"", zone->uz_name);
- }
- KASSERT(curthread->td_critnest == 0 || SCHEDULER_STOPPED(),
- ("uma_zalloc_domain: called with spinlock or critical section held"));
KASSERT((zone->uz_flags & UMA_ZONE_SMR) == 0,
("uma_zalloc_domain: called with SMR zone."));
#ifdef NUMA
@@ -3831,6 +3825,11 @@ uma_zalloc_domain(uma_zone_t zone, void *udata, int domain, int flags)
if (vm_ndomains == 1)
return (uma_zalloc_arg(zone, udata, flags));
+#ifdef UMA_ZALLOC_DEBUG
+ if (uma_zalloc_debug(zone, &item, udata, flags) == EJUSTRETURN)
+ return (item);
+#endif
+
/*
* Try to allocate from the bucket cache before falling back to the keg.
* We could try harder and attempt to allocate from per-CPU caches or