1fb62fb074
CK is a toolkit providing different lockfree algorithms/data structures. More information can be found here : www.concurrencykit.org
280 lines
6.5 KiB
C
280 lines
6.5 KiB
C
/*
|
|
* Copyright 2011-2015 Samy Al Bahra.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*/
|
|
|
|
#ifndef CK_BRLOCK_H
|
|
#define CK_BRLOCK_H
|
|
|
|
/*
|
|
* Big reader spinlocks provide cache-local contention-free read
|
|
* lock acquisition in the absence of writers. This comes at the
|
|
* cost of O(n) write lock acquisition. They were first implemented
|
|
* in the Linux kernel by Ingo Molnar and David S. Miller around the
|
|
* year 2000.
|
|
*
|
|
* This implementation is thread-agnostic which comes at the cost
|
|
* of larger reader objects due to necessary linkage overhead. In
|
|
* order to cut down on TLB pressure, it is recommended to allocate
|
|
* these objects on the same page.
|
|
*/
|
|
|
|
#include <ck_pr.h>
|
|
#include <ck_stdbool.h>
|
|
#include <ck_stddef.h>
|
|
|
|
struct ck_brlock_reader {
|
|
unsigned int n_readers;
|
|
struct ck_brlock_reader *previous;
|
|
struct ck_brlock_reader *next;
|
|
};
|
|
typedef struct ck_brlock_reader ck_brlock_reader_t;
|
|
|
|
#define CK_BRLOCK_READER_INITIALIZER {0}
|
|
|
|
struct ck_brlock {
|
|
struct ck_brlock_reader *readers;
|
|
unsigned int writer;
|
|
};
|
|
typedef struct ck_brlock ck_brlock_t;
|
|
|
|
#define CK_BRLOCK_INITIALIZER {NULL, false}
|
|
|
|
CK_CC_INLINE static void
|
|
ck_brlock_init(struct ck_brlock *br)
|
|
{
|
|
|
|
br->readers = NULL;
|
|
br->writer = false;
|
|
ck_pr_barrier();
|
|
return;
|
|
}
|
|
|
|
CK_CC_INLINE static void
|
|
ck_brlock_write_lock(struct ck_brlock *br)
|
|
{
|
|
struct ck_brlock_reader *cursor;
|
|
|
|
/*
|
|
* As the frequency of write acquisitions should be low,
|
|
* there is no point to more advanced contention avoidance.
|
|
*/
|
|
while (ck_pr_fas_uint(&br->writer, true) == true)
|
|
ck_pr_stall();
|
|
|
|
ck_pr_fence_atomic_load();
|
|
|
|
/* The reader list is protected under the writer br. */
|
|
for (cursor = br->readers; cursor != NULL; cursor = cursor->next) {
|
|
while (ck_pr_load_uint(&cursor->n_readers) != 0)
|
|
ck_pr_stall();
|
|
}
|
|
|
|
ck_pr_fence_lock();
|
|
return;
|
|
}
|
|
|
|
CK_CC_INLINE static void
|
|
ck_brlock_write_unlock(struct ck_brlock *br)
|
|
{
|
|
|
|
ck_pr_fence_unlock();
|
|
ck_pr_store_uint(&br->writer, false);
|
|
return;
|
|
}
|
|
|
|
CK_CC_INLINE static bool
|
|
ck_brlock_write_trylock(struct ck_brlock *br, unsigned int factor)
|
|
{
|
|
struct ck_brlock_reader *cursor;
|
|
unsigned int steps = 0;
|
|
|
|
while (ck_pr_fas_uint(&br->writer, true) == true) {
|
|
if (++steps >= factor)
|
|
return false;
|
|
|
|
ck_pr_stall();
|
|
}
|
|
|
|
/*
|
|
* We do not require a strict fence here as atomic RMW operations
|
|
* are serializing.
|
|
*/
|
|
ck_pr_fence_atomic_load();
|
|
|
|
for (cursor = br->readers; cursor != NULL; cursor = cursor->next) {
|
|
while (ck_pr_load_uint(&cursor->n_readers) != 0) {
|
|
if (++steps >= factor) {
|
|
ck_brlock_write_unlock(br);
|
|
return false;
|
|
}
|
|
|
|
ck_pr_stall();
|
|
}
|
|
}
|
|
|
|
ck_pr_fence_lock();
|
|
return true;
|
|
}
|
|
|
|
CK_CC_INLINE static void
|
|
ck_brlock_read_register(struct ck_brlock *br, struct ck_brlock_reader *reader)
|
|
{
|
|
|
|
reader->n_readers = 0;
|
|
reader->previous = NULL;
|
|
|
|
/* Implicit compiler barrier. */
|
|
ck_brlock_write_lock(br);
|
|
|
|
reader->next = ck_pr_load_ptr(&br->readers);
|
|
if (reader->next != NULL)
|
|
reader->next->previous = reader;
|
|
ck_pr_store_ptr(&br->readers, reader);
|
|
|
|
ck_brlock_write_unlock(br);
|
|
return;
|
|
}
|
|
|
|
CK_CC_INLINE static void
|
|
ck_brlock_read_unregister(struct ck_brlock *br, struct ck_brlock_reader *reader)
|
|
{
|
|
|
|
ck_brlock_write_lock(br);
|
|
|
|
if (reader->next != NULL)
|
|
reader->next->previous = reader->previous;
|
|
|
|
if (reader->previous != NULL)
|
|
reader->previous->next = reader->next;
|
|
else
|
|
br->readers = reader->next;
|
|
|
|
ck_brlock_write_unlock(br);
|
|
return;
|
|
}
|
|
|
|
CK_CC_INLINE static void
|
|
ck_brlock_read_lock(struct ck_brlock *br, struct ck_brlock_reader *reader)
|
|
{
|
|
|
|
if (reader->n_readers >= 1) {
|
|
ck_pr_store_uint(&reader->n_readers, reader->n_readers + 1);
|
|
return;
|
|
}
|
|
|
|
for (;;) {
|
|
while (ck_pr_load_uint(&br->writer) == true)
|
|
ck_pr_stall();
|
|
|
|
#if defined(__x86__) || defined(__x86_64__)
|
|
ck_pr_fas_uint(&reader->n_readers, 1);
|
|
|
|
/*
|
|
* Serialize reader counter update with respect to load of
|
|
* writer.
|
|
*/
|
|
ck_pr_fence_atomic_load();
|
|
#else
|
|
ck_pr_store_uint(&reader->n_readers, 1);
|
|
|
|
/*
|
|
* Serialize reader counter update with respect to load of
|
|
* writer.
|
|
*/
|
|
ck_pr_fence_store_load();
|
|
#endif
|
|
|
|
if (ck_pr_load_uint(&br->writer) == false)
|
|
break;
|
|
|
|
ck_pr_store_uint(&reader->n_readers, 0);
|
|
}
|
|
|
|
ck_pr_fence_lock();
|
|
return;
|
|
}
|
|
|
|
CK_CC_INLINE static bool
|
|
ck_brlock_read_trylock(struct ck_brlock *br,
|
|
struct ck_brlock_reader *reader,
|
|
unsigned int factor)
|
|
{
|
|
unsigned int steps = 0;
|
|
|
|
if (reader->n_readers >= 1) {
|
|
ck_pr_store_uint(&reader->n_readers, reader->n_readers + 1);
|
|
return true;
|
|
}
|
|
|
|
for (;;) {
|
|
while (ck_pr_load_uint(&br->writer) == true) {
|
|
if (++steps >= factor)
|
|
return false;
|
|
|
|
ck_pr_stall();
|
|
}
|
|
|
|
#if defined(__x86__) || defined(__x86_64__)
|
|
ck_pr_fas_uint(&reader->n_readers, 1);
|
|
|
|
/*
|
|
* Serialize reader counter update with respect to load of
|
|
* writer.
|
|
*/
|
|
ck_pr_fence_atomic_load();
|
|
#else
|
|
ck_pr_store_uint(&reader->n_readers, 1);
|
|
|
|
/*
|
|
* Serialize reader counter update with respect to load of
|
|
* writer.
|
|
*/
|
|
ck_pr_fence_store_load();
|
|
#endif
|
|
|
|
if (ck_pr_load_uint(&br->writer) == false)
|
|
break;
|
|
|
|
ck_pr_store_uint(&reader->n_readers, 0);
|
|
|
|
if (++steps >= factor)
|
|
return false;
|
|
}
|
|
|
|
ck_pr_fence_lock();
|
|
return true;
|
|
}
|
|
|
|
CK_CC_INLINE static void
|
|
ck_brlock_read_unlock(struct ck_brlock_reader *reader)
|
|
{
|
|
|
|
ck_pr_fence_unlock();
|
|
ck_pr_store_uint(&reader->n_readers, reader->n_readers - 1);
|
|
return;
|
|
}
|
|
|
|
#endif /* CK_BRLOCK_H */
|