Fix umtx locking, for libthr, in the kernel.

1. There was a race condition between a thread unlocking
   a umtx and the thread contesting it. If the unlocking
   thread won the race it may try to wakeup a thread that
   was not yet in msleep(). The contesting thread would then
   go to sleep to await a wakeup that would never come. It's
   not possible to close the race by using a lock because
   calls to casuptr() may have to fault a page in from swap.
   Instead, the race was closed by introducing a flag that
   the unlocking thread will set when waking up a thread.
   The contesting thread will check for this flag before
   going to sleep. For now the flag is kept in td_flags,
   but it may be better to use some other member or create
   a new one because of the possible performance/contention
   issues of having to own sched_lock. Thanks to jhb for
   pointing me in the right direction on this one.

2. Once a umtx was contested all future locks and unlocks
   were happening in the kernel, regardless of whether it
   was contested or not. To prevent this from happening,
   when a thread locks a umtx it checks the queue for that
   umtx and unsets the contested bit if there are no other
   threads waiting on it. Again, this is slightly more
   complicated than it needs to be because we can't hold
   a lock across casuptr(). So, the thread has to check
   the queue again after unseting the bit, and reset the
   contested bit if it finds that another thread has put
   itself on the queue in the mean time.

3. Remove the if... block for unlocking an uncontested
   umtx, and replace it with a KASSERT. The _only_ time
   a thread should be unlocking a umtx in the kernel is
   if it is contested.
This commit is contained in:
Mike Makonnen 2003-07-17 11:06:40 +00:00
parent 7c615cf053
commit 994599d782
Notes: svn2git 2020-12-20 02:59:44 +00:00
svn path=/head/; revision=117685
2 changed files with 48 additions and 24 deletions

View File

@ -181,7 +181,7 @@ _umtx_lock(struct thread *td, struct _umtx_lock_args *uap)
return (EFAULT); return (EFAULT);
if (owner == UMTX_CONTESTED) if (owner == UMTX_CONTESTED)
return (0); goto out;
/* If this failed the lock has changed, restart. */ /* If this failed the lock has changed, restart. */
continue; continue;
@ -211,14 +211,20 @@ _umtx_lock(struct thread *td, struct _umtx_lock_args *uap)
/* /*
* We set the contested bit, sleep. Otherwise the lock changed * We set the contested bit, sleep. Otherwise the lock changed
* and we need to retry. * and we need to retry or we lost a race to the thread
* unlocking the umtx.
*/ */
UMTX_LOCK(); UMTX_LOCK();
if (old == owner) mtx_lock_spin(&sched_lock);
if (old == owner && (td->td_flags & TDF_UMTXWAKEUP) == 0) {
mtx_unlock_spin(&sched_lock);
error = msleep(td, &umtx_lock, error = msleep(td, &umtx_lock,
td->td_priority | PCATCH, "umtx", 0); td->td_priority | PCATCH, "umtx", 0);
else mtx_lock_spin(&sched_lock);
} else
error = 0; error = 0;
td->td_flags &= ~TDF_UMTXWAKEUP;
mtx_unlock_spin(&sched_lock);
umtx_remove(uq, td); umtx_remove(uq, td);
UMTX_UNLOCK(); UMTX_UNLOCK();
@ -230,8 +236,37 @@ _umtx_lock(struct thread *td, struct _umtx_lock_args *uap)
if (error) if (error)
return (error); return (error);
} }
out:
return (0); /*
* We reach here only if we just acquired a contested umtx.
*
* If there are no other threads on this umtx's queue
* clear the contested bit. However, we cannot hold
* a lock across casuptr(). So after we unset it we
* have to recheck, and set it again if another thread has
* put itself on the queue in the mean time.
*/
error = 0;
UMTX_LOCK();
uq = umtx_lookup(td, umtx);
UMTX_UNLOCK();
if (uq == NULL)
old = casuptr((intptr_t *)&umtx->u_owner,
((intptr_t)td | UMTX_CONTESTED), (intptr_t)td);
if (uq == NULL && old == ((intptr_t)td | UMTX_CONTESTED)) {
UMTX_LOCK();
uq = umtx_lookup(td, umtx);
UMTX_UNLOCK();
if (uq != NULL) {
old = casuptr((intptr_t *)&umtx->u_owner,
(intptr_t)td, ((intptr_t)td | UMTX_CONTESTED));
if (old == -1)
error = EFAULT;
else if (old != (intptr_t)td)
error = EINVAL;
}
}
return (error);
} }
int int
@ -257,25 +292,9 @@ _umtx_unlock(struct thread *td, struct _umtx_unlock_args *uap)
if ((struct thread *)(owner & ~UMTX_CONTESTED) != td) if ((struct thread *)(owner & ~UMTX_CONTESTED) != td)
return (EPERM); return (EPERM);
/*
* If we own it but it isn't contested then we can just release and
* return.
*/
if ((owner & UMTX_CONTESTED) == 0) {
owner = casuptr((intptr_t *)&umtx->u_owner,
(intptr_t)td, UMTX_UNOWNED);
if (owner == -1) /* We should only ever be in here for contested locks */
return (EFAULT); KASSERT((owner & UMTX_CONTESTED) != 0, ("contested umtx is not."));
/*
* If this failed someone modified the memory without going
* through this api.
*/
if (owner != (intptr_t)td)
return (EINVAL);
return (0);
}
old = casuptr((intptr_t *)&umtx->u_owner, owner, UMTX_CONTESTED); old = casuptr((intptr_t *)&umtx->u_owner, owner, UMTX_CONTESTED);
@ -296,6 +315,10 @@ _umtx_unlock(struct thread *td, struct _umtx_unlock_args *uap)
uq = umtx_lookup(td, umtx); uq = umtx_lookup(td, umtx);
if (uq != NULL) { if (uq != NULL) {
blocked = TAILQ_FIRST(&uq->uq_tdq); blocked = TAILQ_FIRST(&uq->uq_tdq);
KASSERT(blocked != NULL, ("umtx_q with no waiting threads."));
mtx_lock_spin(&sched_lock);
blocked->td_flags |= TDF_UMTXWAKEUP;
mtx_unlock_spin(&sched_lock);
wakeup(blocked); wakeup(blocked);
} }

View File

@ -350,6 +350,7 @@ struct thread {
#define TDF_NEEDRESCHED 0x010000 /* Thread needs to yield. */ #define TDF_NEEDRESCHED 0x010000 /* Thread needs to yield. */
#define TDF_NEEDSIGCHK 0x020000 /* Thread may need signal delivery. */ #define TDF_NEEDSIGCHK 0x020000 /* Thread may need signal delivery. */
#define TDF_SA 0x040000 /* A scheduler activation based thread */ #define TDF_SA 0x040000 /* A scheduler activation based thread */
#define TDF_UMTXWAKEUP 0x080000 /* Libthr thread must not sleep on a umtx. */
#define TDF_DEADLKTREAT 0x800000 /* Lock aquisition - deadlock treatment. */ #define TDF_DEADLKTREAT 0x800000 /* Lock aquisition - deadlock treatment. */
/* "private" flags kept in td_pflags */ /* "private" flags kept in td_pflags */