git: efcb2ec8cb81 - main - callout: provide CALLOUT_TRYLOCK flag
- Go to: [ bottom of page ] [ top of archives ] [ this month ]
Date: Thu, 24 Oct 2024 17:14:37 UTC
The branch main has been updated by glebius:
URL: https://cgit.FreeBSD.org/src/commit/?id=efcb2ec8cb81ea44c58e41082c6d965f9493099f
commit efcb2ec8cb81ea44c58e41082c6d965f9493099f
Author: Gleb Smirnoff <glebius@FreeBSD.org>
AuthorDate: 2024-10-24 16:58:05 +0000
Commit: Gleb Smirnoff <glebius@FreeBSD.org>
CommitDate: 2024-10-24 17:14:03 +0000
callout: provide CALLOUT_TRYLOCK flag
If a callout was initialized with the flag, then the callout(9) system
will not drop the callwheel lock in softclock_call_cc() to obtain the
callout lock. Instead it will use try-lock semantic to obtain the
callout's lock. In case of a failure the callout will be rescheduled to
the 50% of the precision value. The main benefit of such behavior is not
the avoidance of the lock contention in the callout thread, but the fact
that callout with such flag can be actually stopped in a safe manner,
because the race window in the beginning of softclock_call_cc() is closed.
Call of callout_stop() on such a callout would guarantee that nothing will
be executed after callout_stop() returns, neither callout lock will be
dereferenced. A callout marked as CALLOUT_TRYLOCK |
CALLOUT_RETURNUNLOCKED can call callout_stop() from the callout function
itself (0, a failure to stop, will be returned), then unlock the lock and
then free the memory containing the callout structure.
Caveat: when calling callout_stop() from outside the callout function, the
return value from callout_stop() is still inconsistent. A race window at
the end of softclock_call_cc() still exists, so callout_stop() may report
failure to stop, which would not be true.
Reviewed by: jtl, kib
Differential Revision: https://reviews.freebsd.org/D45746
---
sys/kern/kern_timeout.c | 40 ++++++++++++++++++++++++++++++----------
sys/sys/callout.h | 2 +-
2 files changed, 31 insertions(+), 11 deletions(-)
diff --git a/sys/kern/kern_timeout.c b/sys/kern/kern_timeout.c
index 699a57d51f0e..0fa1af6ef5a7 100644
--- a/sys/kern/kern_timeout.c
+++ b/sys/kern/kern_timeout.c
@@ -649,6 +649,7 @@ softclock_call_cc(struct callout *c, struct callout_cpu *cc,
static callout_func_t *lastfunc;
#endif
+ CC_LOCK_ASSERT(cc);
KASSERT((c->c_iflags & CALLOUT_PENDING) == CALLOUT_PENDING,
("softclock_call_cc: pend %p %x", c, c->c_iflags));
KASSERT((c->c_flags & CALLOUT_ACTIVE) == CALLOUT_ACTIVE,
@@ -671,16 +672,29 @@ softclock_call_cc(struct callout *c, struct callout_cpu *cc,
cc_exec_last_func(cc, direct) = c_func;
cc_exec_last_arg(cc, direct) = c_arg;
cc_exec_cancel(cc, direct) = false;
- CC_UNLOCK(cc);
if (c_lock != NULL) {
- class->lc_lock(c_lock, lock_status);
- /*
- * The callout may have been cancelled
- * while we switched locks.
- */
- if (cc_exec_cancel(cc, direct)) {
- class->lc_unlock(c_lock);
- goto skip;
+ if (c_iflags & CALLOUT_TRYLOCK) {
+ if (__predict_false(class->lc_trylock(c_lock,
+ lock_status) == 0)) {
+ cc_exec_curr(cc, direct) = NULL;
+ callout_cc_add(c, cc,
+ cc->cc_lastscan + c->c_precision / 2,
+ qmax(c->c_precision / 2, 1), c_func, c_arg,
+ (direct) ? C_DIRECT_EXEC : 0);
+ return;
+ }
+ CC_UNLOCK(cc);
+ } else {
+ CC_UNLOCK(cc);
+ class->lc_lock(c_lock, lock_status);
+ /*
+ * The callout may have been cancelled
+ * while we switched locks.
+ */
+ if (cc_exec_cancel(cc, direct)) {
+ class->lc_unlock(c_lock);
+ goto skip;
+ }
}
/* The callout cannot be stopped now. */
cc_exec_cancel(cc, direct) = true;
@@ -698,6 +712,7 @@ softclock_call_cc(struct callout *c, struct callout_cpu *cc,
c, c_func, c_arg);
}
} else {
+ CC_UNLOCK(cc);
#ifdef CALLOUT_PROFILING
(*mpcalls)++;
#endif
@@ -1332,10 +1347,15 @@ void
_callout_init_lock(struct callout *c, struct lock_object *lock, int flags)
{
KASSERT(lock != NULL, ("%s: no lock", __func__));
- KASSERT((flags & ~(CALLOUT_RETURNUNLOCKED | CALLOUT_SHAREDLOCK)) == 0,
+ KASSERT((flags & ~(CALLOUT_RETURNUNLOCKED | CALLOUT_SHAREDLOCK |
+ CALLOUT_TRYLOCK)) == 0,
("%s: bad flags %d", __func__, flags));
KASSERT(!(LOCK_CLASS(lock)->lc_flags & LC_SLEEPABLE),
("%s: callout %p has sleepable lock", __func__, c));
+ KASSERT(!(flags & CALLOUT_TRYLOCK) ||
+ (LOCK_CLASS(lock)->lc_trylock != NULL),
+ ("%s: CALLOUT_TRYLOCK requested for %s",
+ __func__, LOCK_CLASS(lock)->lc_name));
*c = (struct callout ){
.c_lock = lock,
diff --git a/sys/sys/callout.h b/sys/sys/callout.h
index 5d0d896b5d3b..c3350b324296 100644
--- a/sys/sys/callout.h
+++ b/sys/sys/callout.h
@@ -39,7 +39,7 @@
#include <sys/_callout.h>
-#define CALLOUT_LOCAL_ALLOC 0x0001 /* was allocated from callfree */
+#define CALLOUT_TRYLOCK 0x0001 /* try semantic in softclock_call_cc */
#define CALLOUT_ACTIVE 0x0002 /* callout is currently active */
#define CALLOUT_PENDING 0x0004 /* callout is waiting for timeout */
#define CALLOUT_MPSAFE 0x0008 /* deprecated */