Bug about sched_4bsd?

Attilio Rao attilio at freebsd.org
Mon Jan 18 07:06:52 UTC 2010


2010/1/18 Kohji Okuno <okuno.kohji at jp.panasonic.com>:
> Hello,
>
> Thank you, Attilio.
> I checked your patch. I think that your patch is better.
> I tested the patch quickly, and I think it's OK.
> # This probrem does not occur easily :-<
>
>
> What do you think about maybe_resched()?
> I have never experienced about maybe_resched(), but I think that the
> race condition may occur.
>
> <<Back Trace>>
> sched_4bsd.c:     maybe_resched()
> sched_4bsd.c:     resetpriority_thread()
> sched_4bsd.c:     sched_nice()                  get thread_lock(td)
> kern_resource.c:  donice()
> kern_resource.c:  setpriority()                 get PROC_LOCK()
>
> static void
> maybe_resched(struct thread *td)
> {
>        THREAD_LOCK_ASSERT(td, MA_OWNED);
>        if (td->td_priority < curthread->td_priority)
>                curthread->td_flags |= TDF_NEEDRESCHED;
> }
>
> I think, when td->td_lock is not &sched_lock, curthread->td_lock is
> not locked in maybe_resched().

I didn't look closely to the maybe_resched() callers but I think it is
ok. The thread_lock() function works in a way that the callers don't
need to know which container lock is present in a particular moment,
there is always a guarantee that the contenders will spin if the lock
on the struct can't be held.
In the case you outlined something very particular was happening.
Basically, we get &sched_lock but sched_lock was not the lock present
on td_lock. That means all the other paths willing to access to
td_lock for that thread (via thread_lock()) were allowed to do that
even if we wanted to keep the critical path closed.

Attilio


-- 
Peace can only be achieved by understanding - A. Einstein


More information about the freebsd-current mailing list