a5d547add3
2) Fix all "magic numbers" to be constants. 3) A collision case that would generate two associations to the same peer due to a missing lock is fixed. 4) Added tracking of where timers are stopped. Approved by: gnn
388 lines
13 KiB
C
388 lines
13 KiB
C
#ifndef __sctp_lock_bsd_h__
|
|
#define __sctp_lock_bsd_h__
|
|
/*-
|
|
* Copyright (c) 2001-2006, Cisco Systems, Inc. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are met:
|
|
*
|
|
* a) Redistributions of source code must retain the above copyright notice,
|
|
* this list of conditions and the following disclaimer.
|
|
*
|
|
* b) Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in
|
|
* the documentation and/or other materials provided with the distribution.
|
|
*
|
|
* c) Neither the name of Cisco Systems, Inc. nor the names of its
|
|
* contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
|
|
* THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
|
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
|
|
* THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
/*
|
|
* General locking concepts: The goal of our locking is to of course provide
|
|
* consistency and yet minimize overhead. We will attempt to use
|
|
* non-recursive locks which are supposed to be quite inexpensive. Now in
|
|
* order to do this the goal is that most functions are not aware of locking.
|
|
* Once we have a TCB we lock it and unlock when we are through. This means
|
|
* that the TCB lock is kind-of a "global" lock when working on an
|
|
* association. Caution must be used when asserting a TCB_LOCK since if we
|
|
* recurse we deadlock.
|
|
*
|
|
* Most other locks (INP and INFO) attempt to localize the locking i.e. we try
|
|
* to contain the lock and unlock within the function that needs to lock it.
|
|
* This sometimes mean we do extra locks and unlocks and lose a bit of
|
|
* efficency, but if the performance statements about non-recursive locks are
|
|
* true this should not be a problem. One issue that arises with this only
|
|
* lock when needed is that if an implicit association setup is done we have
|
|
* a problem. If at the time I lookup an association I have NULL in the tcb
|
|
* return, by the time I call to create the association some other processor
|
|
* could have created it. This is what the CREATE lock on the endpoint.
|
|
* Places where we will be implicitly creating the association OR just
|
|
* creating an association (the connect call) will assert the CREATE_INP
|
|
* lock. This will assure us that during all the lookup of INP and INFO if
|
|
* another creator is also locking/looking up we can gate the two to
|
|
* synchronize. So the CREATE_INP lock is also another one we must use
|
|
* extreme caution in locking to make sure we don't hit a re-entrancy issue.
|
|
*
|
|
* For non FreeBSD 5.x we provide a bunch of EMPTY lock macros so we can
|
|
* blatantly put locks everywhere and they reduce to nothing on
|
|
* NetBSD/OpenBSD and FreeBSD 4.x
|
|
*
|
|
*/
|
|
|
|
/*
|
|
* When working with the global SCTP lists we lock and unlock the INP_INFO
|
|
* lock. So when we go to lookup an association we will want to do a
|
|
* SCTP_INP_INFO_RLOCK() and then when we want to add a new association to
|
|
* the sctppcbinfo list's we will do a SCTP_INP_INFO_WLOCK().
|
|
*/
|
|
#include <sys/cdefs.h>
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
|
|
extern struct sctp_foo_stuff sctp_logoff[];
|
|
extern int sctp_logoff_stuff;
|
|
|
|
#define SCTP_IPI_COUNT_INIT()
|
|
|
|
#define SCTP_STATLOG_INIT_LOCK()
|
|
#define SCTP_STATLOG_LOCK()
|
|
#define SCTP_STATLOG_UNLOCK()
|
|
#define SCTP_STATLOG_DESTROY()
|
|
|
|
#define SCTP_STATLOG_GETREF(x) { \
|
|
x = atomic_fetchadd_int(&global_sctp_cwnd_log_at, 1); \
|
|
if(x == SCTP_STAT_LOG_SIZE) { \
|
|
global_sctp_cwnd_log_at = 1; \
|
|
x = 0; \
|
|
global_sctp_cwnd_log_rolled = 1; \
|
|
} \
|
|
}
|
|
|
|
|
|
#define SCTP_INP_INFO_LOCK_INIT() \
|
|
mtx_init(&sctppcbinfo.ipi_ep_mtx, "sctp-info", "inp_info", MTX_DEF)
|
|
|
|
|
|
#define SCTP_INP_INFO_RLOCK() do { \
|
|
mtx_lock(&sctppcbinfo.ipi_ep_mtx); \
|
|
} while (0)
|
|
|
|
|
|
#define SCTP_INP_INFO_WLOCK() do { \
|
|
mtx_lock(&sctppcbinfo.ipi_ep_mtx); \
|
|
} while (0)
|
|
|
|
|
|
|
|
#define SCTP_IPI_ADDR_INIT() \
|
|
mtx_init(&sctppcbinfo.ipi_addr_mtx, "sctp-addr-wq", "sctp_addr_wq", MTX_DEF)
|
|
|
|
#define SCTP_IPI_ADDR_DESTROY() \
|
|
mtx_destroy(&sctppcbinfo.ipi_addr_mtx)
|
|
|
|
#define SCTP_IPI_ADDR_LOCK() do { \
|
|
mtx_lock(&sctppcbinfo.ipi_addr_mtx); \
|
|
} while (0)
|
|
|
|
#define SCTP_IPI_ADDR_UNLOCK() mtx_unlock(&sctppcbinfo.ipi_addr_mtx)
|
|
|
|
#define SCTP_INP_INFO_RUNLOCK() mtx_unlock(&sctppcbinfo.ipi_ep_mtx)
|
|
#define SCTP_INP_INFO_WUNLOCK() mtx_unlock(&sctppcbinfo.ipi_ep_mtx)
|
|
|
|
/*
|
|
* The INP locks we will use for locking an SCTP endpoint, so for example if
|
|
* we want to change something at the endpoint level for example random_store
|
|
* or cookie secrets we lock the INP level.
|
|
*/
|
|
|
|
#define SCTP_INP_READ_INIT(_inp) \
|
|
mtx_init(&(_inp)->inp_rdata_mtx, "sctp-read", "inpr", MTX_DEF | MTX_DUPOK)
|
|
|
|
#define SCTP_INP_READ_DESTROY(_inp) \
|
|
mtx_destroy(&(_inp)->inp_rdata_mtx)
|
|
|
|
#define SCTP_INP_READ_LOCK(_inp) do { \
|
|
mtx_lock(&(_inp)->inp_rdata_mtx); \
|
|
} while (0)
|
|
|
|
|
|
#define SCTP_INP_READ_UNLOCK(_inp) mtx_unlock(&(_inp)->inp_rdata_mtx)
|
|
|
|
|
|
#define SCTP_INP_LOCK_INIT(_inp) \
|
|
mtx_init(&(_inp)->inp_mtx, "sctp-inp", "inp", MTX_DEF | MTX_DUPOK)
|
|
#define SCTP_ASOC_CREATE_LOCK_INIT(_inp) \
|
|
mtx_init(&(_inp)->inp_create_mtx, "sctp-create", "inp_create", \
|
|
MTX_DEF | MTX_DUPOK)
|
|
|
|
#define SCTP_INP_LOCK_DESTROY(_inp) \
|
|
mtx_destroy(&(_inp)->inp_mtx)
|
|
|
|
#define SCTP_ASOC_CREATE_LOCK_DESTROY(_inp) \
|
|
mtx_destroy(&(_inp)->inp_create_mtx)
|
|
|
|
|
|
#ifdef SCTP_LOCK_LOGGING
|
|
#define SCTP_INP_RLOCK(_inp) do { \
|
|
sctp_log_lock(_inp, (struct sctp_tcb *)NULL, SCTP_LOG_LOCK_INP);\
|
|
mtx_lock(&(_inp)->inp_mtx); \
|
|
} while (0)
|
|
|
|
#define SCTP_INP_WLOCK(_inp) do { \
|
|
sctp_log_lock(_inp, (struct sctp_tcb *)NULL, SCTP_LOG_LOCK_INP);\
|
|
mtx_lock(&(_inp)->inp_mtx); \
|
|
} while (0)
|
|
|
|
#else
|
|
|
|
#define SCTP_INP_RLOCK(_inp) do { \
|
|
mtx_lock(&(_inp)->inp_mtx); \
|
|
} while (0)
|
|
|
|
#define SCTP_INP_WLOCK(_inp) do { \
|
|
mtx_lock(&(_inp)->inp_mtx); \
|
|
} while (0)
|
|
|
|
#endif
|
|
|
|
|
|
#define SCTP_TCB_SEND_LOCK_INIT(_tcb) \
|
|
mtx_init(&(_tcb)->tcb_send_mtx, "sctp-send-tcb", "tcbs", MTX_DEF | MTX_DUPOK)
|
|
|
|
#define SCTP_TCB_SEND_LOCK_DESTROY(_tcb) mtx_destroy(&(_tcb)->tcb_send_mtx)
|
|
|
|
#define SCTP_TCB_SEND_LOCK(_tcb) do { \
|
|
mtx_lock(&(_tcb)->tcb_send_mtx); \
|
|
} while (0)
|
|
|
|
#define SCTP_TCB_SEND_UNLOCK(_tcb) mtx_unlock(&(_tcb)->tcb_send_mtx)
|
|
|
|
#ifdef INVARIANTS
|
|
|
|
#define SCTP_INP_INCR_REF(_inp) { int x; \
|
|
atomic_add_int(&((_inp)->refcount), 1); \
|
|
x = atomic_fetchadd_int(&sctp_logoff_stuff, 1); \
|
|
if(x == 30000) \
|
|
sctp_logoff_stuff = x = 0; \
|
|
sctp_logoff[x].inp = _inp; \
|
|
sctp_logoff[x].ticks = ticks; \
|
|
sctp_logoff[x].lineno = __LINE__; \
|
|
sctp_logoff[x].updown = 1; \
|
|
}
|
|
|
|
#define SCTP_INP_DECR_REF(_inp) { int x; \
|
|
if (atomic_fetchadd_int(&((_inp)->refcount), -1) == 0 ) panic("refcount goes negative"); \
|
|
x = atomic_fetchadd_int(&sctp_logoff_stuff, 1); \
|
|
if(x == 30000) \
|
|
sctp_logoff_stuff = x = 0; \
|
|
sctp_logoff[x].inp = _inp; \
|
|
sctp_logoff[x].ticks = ticks; \
|
|
sctp_logoff[x].lineno = __LINE__; \
|
|
sctp_logoff[x].updown = 0; \
|
|
}
|
|
|
|
#else
|
|
|
|
#define SCTP_INP_INCR_REF(_inp) atomic_add_int(&((_inp)->refcount), 1)
|
|
#define SCTP_INP_DECR_REF(_inp) atomic_add_int(&((_inp)->refcount), -1)
|
|
|
|
#endif
|
|
|
|
#ifdef SCTP_LOCK_LOGGING
|
|
#define SCTP_ASOC_CREATE_LOCK(_inp) \
|
|
do { \
|
|
sctp_log_lock(_inp, (struct sctp_tcb *)NULL, SCTP_LOG_LOCK_CREATE); \
|
|
mtx_lock(&(_inp)->inp_create_mtx); \
|
|
} while (0)
|
|
#else
|
|
|
|
#define SCTP_ASOC_CREATE_LOCK(_inp) \
|
|
do { \
|
|
mtx_lock(&(_inp)->inp_create_mtx); \
|
|
} while (0)
|
|
#endif
|
|
|
|
#define SCTP_INP_RUNLOCK(_inp) mtx_unlock(&(_inp)->inp_mtx)
|
|
#define SCTP_INP_WUNLOCK(_inp) mtx_unlock(&(_inp)->inp_mtx)
|
|
#define SCTP_ASOC_CREATE_UNLOCK(_inp) mtx_unlock(&(_inp)->inp_create_mtx)
|
|
|
|
/*
|
|
* For the majority of things (once we have found the association) we will
|
|
* lock the actual association mutex. This will protect all the assoiciation
|
|
* level queues and streams and such. We will need to lock the socket layer
|
|
* when we stuff data up into the receiving sb_mb. I.e. we will need to do an
|
|
* extra SOCKBUF_LOCK(&so->so_rcv) even though the association is locked.
|
|
*/
|
|
|
|
#define SCTP_TCB_LOCK_INIT(_tcb) \
|
|
mtx_init(&(_tcb)->tcb_mtx, "sctp-tcb", "tcb", MTX_DEF | MTX_DUPOK)
|
|
|
|
#define SCTP_TCB_LOCK_DESTROY(_tcb) mtx_destroy(&(_tcb)->tcb_mtx)
|
|
|
|
#ifdef SCTP_LOCK_LOGGING
|
|
#define SCTP_TCB_LOCK(_tcb) do { \
|
|
sctp_log_lock(_tcb->sctp_ep, _tcb, SCTP_LOG_LOCK_TCB); \
|
|
mtx_lock(&(_tcb)->tcb_mtx); \
|
|
} while (0)
|
|
|
|
#else
|
|
#define SCTP_TCB_LOCK(_tcb) do { \
|
|
mtx_lock(&(_tcb)->tcb_mtx); \
|
|
} while (0)
|
|
|
|
#endif
|
|
|
|
|
|
#define SCTP_TCB_TRYLOCK(_tcb) mtx_trylock(&(_tcb)->tcb_mtx)
|
|
|
|
#define SCTP_TCB_UNLOCK(_tcb) mtx_unlock(&(_tcb)->tcb_mtx)
|
|
|
|
#define SCTP_TCB_UNLOCK_IFOWNED(_tcb) do { \
|
|
if (mtx_owned(&(_tcb)->tcb_mtx)) \
|
|
mtx_unlock(&(_tcb)->tcb_mtx); \
|
|
} while (0)
|
|
|
|
|
|
|
|
#ifdef INVARIANTS
|
|
#define SCTP_TCB_LOCK_ASSERT(_tcb) do { \
|
|
if (mtx_owned(&(_tcb)->tcb_mtx) == 0) \
|
|
panic("Don't own TCB lock"); \
|
|
} while (0)
|
|
#else
|
|
#define SCTP_TCB_LOCK_ASSERT(_tcb)
|
|
#endif
|
|
|
|
#define SCTP_ITERATOR_LOCK_INIT() \
|
|
mtx_init(&sctppcbinfo.it_mtx, "sctp-it", "iterator", MTX_DEF)
|
|
|
|
#ifdef INVARIANTS
|
|
#define SCTP_ITERATOR_LOCK() \
|
|
do { \
|
|
if (mtx_owned(&sctppcbinfo.it_mtx)) \
|
|
panic("Iterator Lock"); \
|
|
mtx_lock(&sctppcbinfo.it_mtx); \
|
|
} while (0)
|
|
#else
|
|
#define SCTP_ITERATOR_LOCK() \
|
|
do { \
|
|
mtx_lock(&sctppcbinfo.it_mtx); \
|
|
} while (0)
|
|
|
|
#endif
|
|
|
|
#define SCTP_ITERATOR_UNLOCK() mtx_unlock(&sctppcbinfo.it_mtx)
|
|
#define SCTP_ITERATOR_LOCK_DESTROY() mtx_destroy(&sctppcbinfo.it_mtx)
|
|
|
|
|
|
#define SCTP_INCR_EP_COUNT() \
|
|
do { \
|
|
atomic_add_int(&sctppcbinfo.ipi_count_ep, 1); \
|
|
} while (0)
|
|
|
|
#define SCTP_DECR_EP_COUNT() \
|
|
do { \
|
|
atomic_add_int(&sctppcbinfo.ipi_count_ep,-1); \
|
|
} while (0)
|
|
|
|
#define SCTP_INCR_ASOC_COUNT() \
|
|
do { \
|
|
atomic_add_int(&sctppcbinfo.ipi_count_asoc, 1); \
|
|
} while (0)
|
|
|
|
#define SCTP_DECR_ASOC_COUNT() \
|
|
do { \
|
|
atomic_add_int(&sctppcbinfo.ipi_count_asoc, -1); \
|
|
} while (0)
|
|
|
|
#define SCTP_INCR_LADDR_COUNT() \
|
|
do { \
|
|
atomic_add_int(&sctppcbinfo.ipi_count_laddr, 1); \
|
|
} while (0)
|
|
|
|
#define SCTP_DECR_LADDR_COUNT() \
|
|
do { \
|
|
atomic_add_int(&sctppcbinfo.ipi_count_laddr, -1); \
|
|
} while (0)
|
|
|
|
#define SCTP_INCR_RADDR_COUNT() \
|
|
do { \
|
|
atomic_add_int(&sctppcbinfo.ipi_count_raddr,1); \
|
|
} while (0)
|
|
|
|
#define SCTP_DECR_RADDR_COUNT() \
|
|
do { \
|
|
atomic_add_int(&sctppcbinfo.ipi_count_raddr,-1); \
|
|
} while (0)
|
|
|
|
#define SCTP_INCR_CHK_COUNT() \
|
|
do { \
|
|
atomic_add_int(&sctppcbinfo.ipi_count_chunk, 1); \
|
|
} while (0)
|
|
|
|
#define SCTP_DECR_CHK_COUNT() \
|
|
do { \
|
|
if(sctppcbinfo.ipi_count_chunk == 0) \
|
|
panic("chunk count to 0?"); \
|
|
atomic_add_int(&sctppcbinfo.ipi_count_chunk,-1); \
|
|
} while (0)
|
|
|
|
#define SCTP_INCR_READQ_COUNT() \
|
|
do { \
|
|
atomic_add_int(&sctppcbinfo.ipi_count_readq,1); \
|
|
} while (0)
|
|
|
|
#define SCTP_DECR_READQ_COUNT() \
|
|
do { \
|
|
atomic_add_int(&sctppcbinfo.ipi_count_readq, -1); \
|
|
} while (0)
|
|
|
|
#define SCTP_INCR_STRMOQ_COUNT() \
|
|
do { \
|
|
atomic_add_int(&sctppcbinfo.ipi_count_strmoq, 1); \
|
|
} while (0)
|
|
|
|
#define SCTP_DECR_STRMOQ_COUNT() \
|
|
do { \
|
|
atomic_add_int(&sctppcbinfo.ipi_count_strmoq,-1); \
|
|
} while (0)
|
|
|
|
|
|
|
|
|
|
|
|
#endif
|