7de1ecef2d
locked and unlocked completely in userland. by locking and unlocking mutex in userland, it reduces the total time a mutex is locked by a thread, in some application code, a mutex only protects a small piece of code, the code's execution time is less than a simple system call, if a lock contention happens, however in current implemenation, the lock holder has to extend its locking time and enter kernel to unlock it, the change avoids this disadvantage, it first sets mutex to free state and then enters kernel and wake one waiter up. This improves performance dramatically in some sysbench mutex tests. Tested by: kris Sounds great: jeff
217 lines
5.7 KiB
C
217 lines
5.7 KiB
C
/*
|
|
* Copyright (c) 2005 David Xu <davidxu@freebsd.org>
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice unmodified, this list of conditions, and the following
|
|
* disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
|
|
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
|
|
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
|
|
* IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
|
|
* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
|
|
* NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
|
|
* THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*
|
|
* $FreeBSD$
|
|
*
|
|
*/
|
|
|
|
#include "thr_private.h"
|
|
#include "thr_umtx.h"
|
|
|
|
#ifndef HAS__UMTX_OP_ERR
|
|
int _umtx_op_err(void *obj, int op, u_long val, void *uaddr, void *uaddr2)
|
|
{
|
|
if (_umtx_op(obj, op, val, uaddr, uaddr2) == -1)
|
|
return (errno);
|
|
return (0);
|
|
}
|
|
#endif
|
|
|
|
void
|
|
_thr_umutex_init(struct umutex *mtx)
|
|
{
|
|
static struct umutex default_mtx = DEFAULT_UMUTEX;
|
|
|
|
*mtx = default_mtx;
|
|
}
|
|
|
|
int
|
|
__thr_umutex_lock(struct umutex *mtx, uint32_t id)
|
|
{
|
|
uint32_t owner;
|
|
|
|
if ((mtx->m_flags & (UMUTEX_PRIO_PROTECT | UMUTEX_PRIO_INHERIT)) == 0) {
|
|
for (;;) {
|
|
/* wait in kernel */
|
|
_umtx_op_err(mtx, UMTX_OP_MUTEX_WAIT, 0, 0, 0);
|
|
|
|
owner = mtx->m_owner;
|
|
if ((owner & ~UMUTEX_CONTESTED) == 0 &&
|
|
atomic_cmpset_acq_32(&mtx->m_owner, owner, id|owner))
|
|
return (0);
|
|
}
|
|
}
|
|
|
|
return _umtx_op_err(mtx, UMTX_OP_MUTEX_LOCK, 0, 0, 0);
|
|
}
|
|
|
|
int
|
|
__thr_umutex_timedlock(struct umutex *mtx, uint32_t id,
|
|
const struct timespec *ets)
|
|
{
|
|
struct timespec timo, cts;
|
|
uint32_t owner;
|
|
int ret;
|
|
|
|
clock_gettime(CLOCK_REALTIME, &cts);
|
|
TIMESPEC_SUB(&timo, ets, &cts);
|
|
|
|
if (timo.tv_sec < 0)
|
|
return (ETIMEDOUT);
|
|
|
|
for (;;) {
|
|
if ((mtx->m_flags & (UMUTEX_PRIO_PROTECT | UMUTEX_PRIO_INHERIT)) == 0) {
|
|
|
|
/* wait in kernel */
|
|
ret = _umtx_op_err(mtx, UMTX_OP_MUTEX_WAIT, 0, 0, &timo);
|
|
|
|
/* now try to lock it */
|
|
owner = mtx->m_owner;
|
|
if ((owner & ~UMUTEX_CONTESTED) == 0 &&
|
|
atomic_cmpset_acq_32(&mtx->m_owner, owner, id|owner))
|
|
return (0);
|
|
} else {
|
|
ret = _umtx_op_err(mtx, UMTX_OP_MUTEX_LOCK, 0, 0, &timo);
|
|
if (ret == 0)
|
|
break;
|
|
}
|
|
if (ret == ETIMEDOUT)
|
|
break;
|
|
clock_gettime(CLOCK_REALTIME, &cts);
|
|
TIMESPEC_SUB(&timo, ets, &cts);
|
|
if (timo.tv_sec < 0 || (timo.tv_sec == 0 && timo.tv_nsec == 0)) {
|
|
ret = ETIMEDOUT;
|
|
break;
|
|
}
|
|
}
|
|
return (ret);
|
|
}
|
|
|
|
int
|
|
__thr_umutex_unlock(struct umutex *mtx, uint32_t id)
|
|
{
|
|
if ((mtx->m_flags & (UMUTEX_PRIO_PROTECT | UMUTEX_PRIO_INHERIT)) == 0) {
|
|
atomic_cmpset_rel_32(&mtx->m_owner, id | UMUTEX_CONTESTED, UMUTEX_CONTESTED);
|
|
return _umtx_op_err(mtx, UMTX_OP_MUTEX_WAKE, 0, 0, 0);
|
|
}
|
|
return _umtx_op_err(mtx, UMTX_OP_MUTEX_UNLOCK, 0, 0, 0);
|
|
}
|
|
|
|
int
|
|
__thr_umutex_trylock(struct umutex *mtx)
|
|
{
|
|
return _umtx_op_err(mtx, UMTX_OP_MUTEX_TRYLOCK, 0, 0, 0);
|
|
}
|
|
|
|
int
|
|
__thr_umutex_set_ceiling(struct umutex *mtx, uint32_t ceiling,
|
|
uint32_t *oldceiling)
|
|
{
|
|
return _umtx_op_err(mtx, UMTX_OP_SET_CEILING, ceiling, oldceiling, 0);
|
|
}
|
|
|
|
int
|
|
_thr_umtx_wait(volatile long *mtx, long id, const struct timespec *timeout)
|
|
{
|
|
if (timeout && (timeout->tv_sec < 0 || (timeout->tv_sec == 0 &&
|
|
timeout->tv_nsec <= 0)))
|
|
return (ETIMEDOUT);
|
|
return _umtx_op_err(__DEVOLATILE(void *, mtx), UMTX_OP_WAIT, id, 0,
|
|
__DECONST(void*, timeout));
|
|
}
|
|
|
|
int
|
|
_thr_umtx_wait_uint(volatile u_int *mtx, u_int id, const struct timespec *timeout, int shared)
|
|
{
|
|
if (timeout && (timeout->tv_sec < 0 || (timeout->tv_sec == 0 &&
|
|
timeout->tv_nsec <= 0)))
|
|
return (ETIMEDOUT);
|
|
return _umtx_op_err(__DEVOLATILE(void *, mtx),
|
|
shared ? UMTX_OP_WAIT_UINT : UMTX_OP_WAIT_UINT_PRIVATE, id, 0,
|
|
__DECONST(void*, timeout));
|
|
}
|
|
|
|
int
|
|
_thr_umtx_wake(volatile void *mtx, int nr_wakeup, int shared)
|
|
{
|
|
return _umtx_op_err(__DEVOLATILE(void *, mtx), shared ? UMTX_OP_WAKE : UMTX_OP_WAKE_PRIVATE,
|
|
nr_wakeup, 0, 0);
|
|
}
|
|
|
|
void
|
|
_thr_ucond_init(struct ucond *cv)
|
|
{
|
|
bzero(cv, sizeof(struct ucond));
|
|
}
|
|
|
|
int
|
|
_thr_ucond_wait(struct ucond *cv, struct umutex *m,
|
|
const struct timespec *timeout, int check_unparking)
|
|
{
|
|
if (timeout && (timeout->tv_sec < 0 || (timeout->tv_sec == 0 &&
|
|
timeout->tv_nsec <= 0))) {
|
|
struct pthread *curthread = _get_curthread();
|
|
_thr_umutex_unlock(m, TID(curthread));
|
|
return (ETIMEDOUT);
|
|
}
|
|
return _umtx_op_err(cv, UMTX_OP_CV_WAIT,
|
|
check_unparking ? UMTX_CHECK_UNPARKING : 0,
|
|
m, __DECONST(void*, timeout));
|
|
}
|
|
|
|
int
|
|
_thr_ucond_signal(struct ucond *cv)
|
|
{
|
|
if (!cv->c_has_waiters)
|
|
return (0);
|
|
return _umtx_op_err(cv, UMTX_OP_CV_SIGNAL, 0, NULL, NULL);
|
|
}
|
|
|
|
int
|
|
_thr_ucond_broadcast(struct ucond *cv)
|
|
{
|
|
if (!cv->c_has_waiters)
|
|
return (0);
|
|
return _umtx_op_err(cv, UMTX_OP_CV_BROADCAST, 0, NULL, NULL);
|
|
}
|
|
|
|
int
|
|
__thr_rwlock_rdlock(struct urwlock *rwlock, int flags, struct timespec *tsp)
|
|
{
|
|
return _umtx_op_err(rwlock, UMTX_OP_RW_RDLOCK, flags, NULL, tsp);
|
|
}
|
|
|
|
int
|
|
__thr_rwlock_wrlock(struct urwlock *rwlock, struct timespec *tsp)
|
|
{
|
|
return _umtx_op_err(rwlock, UMTX_OP_RW_WRLOCK, 0, NULL, tsp);
|
|
}
|
|
|
|
int
|
|
__thr_rwlock_unlock(struct urwlock *rwlock)
|
|
{
|
|
return _umtx_op_err(rwlock, UMTX_OP_RW_UNLOCK, 0, NULL, NULL);
|
|
}
|