28362a5c80
low-level lock used by the libpthread implementation. In the future, we'll eliminate spinlocks from libc but that will wait until after 5.1-release. Don't call an application signal handler if the handler is the same as the library-installed handler. This seems to be possible after a fork and is the cause of konsole hangs. Approved by: re@ (jhb)
153 lines
4.6 KiB
C
153 lines
4.6 KiB
C
/*
|
|
* Copyright (c) 1997 John Birrell <jb@cimlogic.com.au>.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
* must display the following acknowledgement:
|
|
* This product includes software developed by John Birrell.
|
|
* 4. Neither the name of the author nor the names of any co-contributors
|
|
* may be used to endorse or promote products derived from this software
|
|
* without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY JOHN BIRRELL AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* $FreeBSD$
|
|
*
|
|
*/
|
|
|
|
#include <sys/types.h>
|
|
#include <machine/atomic.h>
|
|
|
|
#include <libc_private.h>
|
|
#include "spinlock.h"
|
|
#include "thr_private.h"
|
|
|
|
#define MAX_SPINLOCKS 5
|
|
|
|
struct spinlock_extra {
|
|
struct lock lock;
|
|
kse_critical_t crit;
|
|
};
|
|
|
|
static void init_spinlock(spinlock_t *lck);
|
|
|
|
static struct lock spinlock_static_lock;
|
|
static struct spinlock_extra extra[MAX_SPINLOCKS];
|
|
static int spinlock_count = 0;
|
|
static int initialized = 0;
|
|
|
|
/*
|
|
* These are for compatability only. Spinlocks of this type
|
|
* are deprecated.
|
|
*/
|
|
|
|
void
|
|
_spinunlock(spinlock_t *lck)
|
|
{
|
|
struct spinlock_extra *extra;
|
|
kse_critical_t crit;
|
|
|
|
extra = (struct spinlock_extra *)lck->fname;
|
|
crit = extra->crit;
|
|
KSE_LOCK_RELEASE(_get_curkse(), &extra->lock);
|
|
_kse_critical_leave(crit);
|
|
}
|
|
|
|
|
|
/*
|
|
* Lock a location for the running thread. Yield to allow other
|
|
* threads to run if this thread is blocked because the lock is
|
|
* not available. Note that this function does not sleep. It
|
|
* assumes that the lock will be available very soon.
|
|
*/
|
|
void
|
|
_spinlock(spinlock_t *lck)
|
|
{
|
|
struct spinlock_extra *extra;
|
|
kse_critical_t crit;
|
|
|
|
THR_ASSERT(__isthreaded != 0, "Spinlock called when not threaded.");
|
|
THR_ASSERT(initialized != 0, "Spinlocks not initialized.");
|
|
/*
|
|
* Try to grab the lock and loop if another thread grabs
|
|
* it before we do.
|
|
*/
|
|
crit = _kse_critical_enter();
|
|
if (lck->fname == NULL)
|
|
init_spinlock(lck);
|
|
extra = (struct spinlock_extra *)lck->fname;
|
|
KSE_LOCK_ACQUIRE(_get_curkse(), &extra->lock);
|
|
extra->crit = crit;
|
|
}
|
|
|
|
/*
|
|
* Lock a location for the running thread. Yield to allow other
|
|
* threads to run if this thread is blocked because the lock is
|
|
* not available. Note that this function does not sleep. It
|
|
* assumes that the lock will be available very soon.
|
|
*
|
|
* This function checks if the running thread has already locked the
|
|
* location, warns if this occurs and creates a thread dump before
|
|
* returning.
|
|
*/
|
|
void
|
|
_spinlock_debug(spinlock_t *lck, char *fname, int lineno)
|
|
{
|
|
_spinlock(lck);
|
|
}
|
|
|
|
static void
|
|
init_spinlock(spinlock_t *lck)
|
|
{
|
|
struct kse *curkse = _get_curkse();
|
|
|
|
KSE_LOCK_ACQUIRE(curkse, &spinlock_static_lock);
|
|
if ((lck->fname == NULL) && (spinlock_count < MAX_SPINLOCKS)) {
|
|
lck->fname = (char *)&extra[spinlock_count];
|
|
spinlock_count++;
|
|
}
|
|
KSE_LOCK_RELEASE(curkse, &spinlock_static_lock);
|
|
THR_ASSERT(lck->fname != NULL, "Exceeded max spinlocks");
|
|
}
|
|
|
|
void
|
|
_thr_spinlock_init(void)
|
|
{
|
|
int i;
|
|
|
|
if (initialized != 0) {
|
|
_lock_destroy(&spinlock_static_lock);
|
|
for (i = 0; i < MAX_SPINLOCKS; i++) {
|
|
_lock_destroy(&extra[i].lock);
|
|
}
|
|
}
|
|
|
|
if (_lock_init(&spinlock_static_lock, LCK_ADAPTIVE,
|
|
_kse_lock_wait, _kse_lock_wakeup) != 0)
|
|
PANIC("Cannot initialize spinlock_static_lock");
|
|
for (i = 0; i < MAX_SPINLOCKS; i++) {
|
|
if (_lock_init(&extra[i].lock, LCK_ADAPTIVE,
|
|
_kse_lock_wait, _kse_lock_wakeup) != 0)
|
|
PANIC("Cannot initialize spinlock extra");
|
|
}
|
|
initialized = 1;
|
|
}
|