5a50eb6585
Reviewed by: tuexen, kib MFC after: 1 week Sponsored by: Netflix Differential Revision: https://reviews.freebsd.org/D29206
530 lines
14 KiB
C
530 lines
14 KiB
C
/*-
|
|
* SPDX-License-Identifier: BSD-3-Clause
|
|
*
|
|
* Copyright (c) 2001-2007, by Cisco Systems, Inc. All rights reserved.
|
|
* Copyright (c) 2008-2012, by Randall Stewart. All rights reserved.
|
|
* Copyright (c) 2008-2012, by Michael Tuexen. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are met:
|
|
*
|
|
* a) Redistributions of source code must retain the above copyright notice,
|
|
* this list of conditions and the following disclaimer.
|
|
*
|
|
* b) Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in
|
|
* the documentation and/or other materials provided with the distribution.
|
|
*
|
|
* c) Neither the name of Cisco Systems, Inc. nor the names of its
|
|
* contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
|
|
* THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
|
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
|
|
* THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
#include <netinet/sctp_os.h>
|
|
#include <netinet/sctp_var.h>
|
|
#include <netinet/sctp_pcb.h>
|
|
#include <netinet/sctp_header.h>
|
|
#include <netinet/sctputil.h>
|
|
#include <netinet/sctp_output.h>
|
|
#include <netinet/sctp_bsd_addr.h>
|
|
#include <netinet/sctp_uio.h>
|
|
#include <netinet/sctputil.h>
|
|
#include <netinet/sctp_timer.h>
|
|
#include <netinet/sctp_asconf.h>
|
|
#include <netinet/sctp_sysctl.h>
|
|
#include <netinet/sctp_indata.h>
|
|
#include <sys/unistd.h>
|
|
|
|
/* Declare all of our malloc named types */
|
|
MALLOC_DEFINE(SCTP_M_MAP, "sctp_map", "sctp asoc map descriptor");
|
|
MALLOC_DEFINE(SCTP_M_STRMI, "sctp_stri", "sctp stream in array");
|
|
MALLOC_DEFINE(SCTP_M_STRMO, "sctp_stro", "sctp stream out array");
|
|
MALLOC_DEFINE(SCTP_M_ASC_ADDR, "sctp_aadr", "sctp asconf address");
|
|
MALLOC_DEFINE(SCTP_M_ASC_IT, "sctp_a_it", "sctp asconf iterator");
|
|
MALLOC_DEFINE(SCTP_M_AUTH_CL, "sctp_atcl", "sctp auth chunklist");
|
|
MALLOC_DEFINE(SCTP_M_AUTH_KY, "sctp_atky", "sctp auth key");
|
|
MALLOC_DEFINE(SCTP_M_AUTH_HL, "sctp_athm", "sctp auth hmac list");
|
|
MALLOC_DEFINE(SCTP_M_AUTH_IF, "sctp_athi", "sctp auth info");
|
|
MALLOC_DEFINE(SCTP_M_STRESET, "sctp_stre", "sctp stream reset");
|
|
MALLOC_DEFINE(SCTP_M_CMSG, "sctp_cmsg", "sctp CMSG buffer");
|
|
MALLOC_DEFINE(SCTP_M_COPYAL, "sctp_cpal", "sctp copy all");
|
|
MALLOC_DEFINE(SCTP_M_VRF, "sctp_vrf", "sctp vrf struct");
|
|
MALLOC_DEFINE(SCTP_M_IFA, "sctp_ifa", "sctp ifa struct");
|
|
MALLOC_DEFINE(SCTP_M_IFN, "sctp_ifn", "sctp ifn struct");
|
|
MALLOC_DEFINE(SCTP_M_TIMW, "sctp_timw", "sctp time block");
|
|
MALLOC_DEFINE(SCTP_M_MVRF, "sctp_mvrf", "sctp mvrf pcb list");
|
|
MALLOC_DEFINE(SCTP_M_ITER, "sctp_iter", "sctp iterator control");
|
|
MALLOC_DEFINE(SCTP_M_SOCKOPT, "sctp_socko", "sctp socket option");
|
|
MALLOC_DEFINE(SCTP_M_MCORE, "sctp_mcore", "sctp mcore queue");
|
|
|
|
/* Global NON-VNET structure that controls the iterator */
|
|
struct iterator_control sctp_it_ctl;
|
|
|
|
void
|
|
sctp_wakeup_iterator(void)
|
|
{
|
|
wakeup(&sctp_it_ctl.iterator_running);
|
|
}
|
|
|
|
static void
|
|
sctp_iterator_thread(void *v SCTP_UNUSED)
|
|
{
|
|
SCTP_IPI_ITERATOR_WQ_LOCK();
|
|
/* In FreeBSD this thread never terminates. */
|
|
for (;;) {
|
|
msleep(&sctp_it_ctl.iterator_running,
|
|
&sctp_it_ctl.ipi_iterator_wq_mtx,
|
|
0, "waiting_for_work", 0);
|
|
sctp_iterator_worker();
|
|
}
|
|
}
|
|
|
|
void
|
|
sctp_startup_iterator(void)
|
|
{
|
|
if (sctp_it_ctl.thread_proc) {
|
|
/* You only get one */
|
|
return;
|
|
}
|
|
/* Initialize global locks here, thus only once. */
|
|
SCTP_ITERATOR_LOCK_INIT();
|
|
SCTP_IPI_ITERATOR_WQ_INIT();
|
|
TAILQ_INIT(&sctp_it_ctl.iteratorhead);
|
|
kproc_create(sctp_iterator_thread,
|
|
(void *)NULL,
|
|
&sctp_it_ctl.thread_proc,
|
|
0,
|
|
SCTP_KTHREAD_PAGES,
|
|
SCTP_KTRHEAD_NAME);
|
|
}
|
|
|
|
#ifdef INET6
|
|
|
|
void
|
|
sctp_gather_internal_ifa_flags(struct sctp_ifa *ifa)
|
|
{
|
|
struct in6_ifaddr *ifa6;
|
|
|
|
ifa6 = (struct in6_ifaddr *)ifa->ifa;
|
|
ifa->flags = ifa6->ia6_flags;
|
|
if (!MODULE_GLOBAL(ip6_use_deprecated)) {
|
|
if (ifa->flags &
|
|
IN6_IFF_DEPRECATED) {
|
|
ifa->localifa_flags |= SCTP_ADDR_IFA_UNUSEABLE;
|
|
} else {
|
|
ifa->localifa_flags &= ~SCTP_ADDR_IFA_UNUSEABLE;
|
|
}
|
|
} else {
|
|
ifa->localifa_flags &= ~SCTP_ADDR_IFA_UNUSEABLE;
|
|
}
|
|
if (ifa->flags &
|
|
(IN6_IFF_DETACHED |
|
|
IN6_IFF_ANYCAST |
|
|
IN6_IFF_NOTREADY)) {
|
|
ifa->localifa_flags |= SCTP_ADDR_IFA_UNUSEABLE;
|
|
} else {
|
|
ifa->localifa_flags &= ~SCTP_ADDR_IFA_UNUSEABLE;
|
|
}
|
|
}
|
|
#endif /* INET6 */
|
|
|
|
static uint32_t
|
|
sctp_is_desired_interface_type(struct ifnet *ifn)
|
|
{
|
|
int result;
|
|
|
|
/* check the interface type to see if it's one we care about */
|
|
switch (ifn->if_type) {
|
|
case IFT_ETHER:
|
|
case IFT_ISO88023:
|
|
case IFT_ISO88024:
|
|
case IFT_ISO88025:
|
|
case IFT_ISO88026:
|
|
case IFT_STARLAN:
|
|
case IFT_P10:
|
|
case IFT_P80:
|
|
case IFT_HY:
|
|
case IFT_FDDI:
|
|
case IFT_XETHER:
|
|
case IFT_ISDNBASIC:
|
|
case IFT_ISDNPRIMARY:
|
|
case IFT_PTPSERIAL:
|
|
case IFT_OTHER:
|
|
case IFT_PPP:
|
|
case IFT_LOOP:
|
|
case IFT_SLIP:
|
|
case IFT_GIF:
|
|
case IFT_L2VLAN:
|
|
case IFT_STF:
|
|
case IFT_IP:
|
|
case IFT_IPOVERCDLC:
|
|
case IFT_IPOVERCLAW:
|
|
case IFT_PROPVIRTUAL: /* NetGraph Virtual too */
|
|
case IFT_VIRTUALIPADDRESS:
|
|
result = 1;
|
|
break;
|
|
default:
|
|
result = 0;
|
|
}
|
|
|
|
return (result);
|
|
}
|
|
|
|
static void
|
|
sctp_init_ifns_for_vrf(int vrfid)
|
|
{
|
|
/*
|
|
* Here we must apply ANY locks needed by the IFN we access and also
|
|
* make sure we lock any IFA that exists as we float through the
|
|
* list of IFA's
|
|
*/
|
|
struct epoch_tracker et;
|
|
struct ifnet *ifn;
|
|
struct ifaddr *ifa;
|
|
struct sctp_ifa *sctp_ifa;
|
|
uint32_t ifa_flags;
|
|
#ifdef INET6
|
|
struct in6_ifaddr *ifa6;
|
|
#endif
|
|
|
|
IFNET_RLOCK();
|
|
NET_EPOCH_ENTER(et);
|
|
CK_STAILQ_FOREACH(ifn, &MODULE_GLOBAL(ifnet), if_link) {
|
|
if (sctp_is_desired_interface_type(ifn) == 0) {
|
|
/* non desired type */
|
|
continue;
|
|
}
|
|
CK_STAILQ_FOREACH(ifa, &ifn->if_addrhead, ifa_link) {
|
|
if (ifa->ifa_addr == NULL) {
|
|
continue;
|
|
}
|
|
switch (ifa->ifa_addr->sa_family) {
|
|
#ifdef INET
|
|
case AF_INET:
|
|
if (((struct sockaddr_in *)ifa->ifa_addr)->sin_addr.s_addr == 0) {
|
|
continue;
|
|
}
|
|
break;
|
|
#endif
|
|
#ifdef INET6
|
|
case AF_INET6:
|
|
if (IN6_IS_ADDR_UNSPECIFIED(&((struct sockaddr_in6 *)ifa->ifa_addr)->sin6_addr)) {
|
|
/* skip unspecifed addresses */
|
|
continue;
|
|
}
|
|
break;
|
|
#endif
|
|
default:
|
|
continue;
|
|
}
|
|
switch (ifa->ifa_addr->sa_family) {
|
|
#ifdef INET
|
|
case AF_INET:
|
|
ifa_flags = 0;
|
|
break;
|
|
#endif
|
|
#ifdef INET6
|
|
case AF_INET6:
|
|
ifa6 = (struct in6_ifaddr *)ifa;
|
|
ifa_flags = ifa6->ia6_flags;
|
|
break;
|
|
#endif
|
|
default:
|
|
ifa_flags = 0;
|
|
break;
|
|
}
|
|
sctp_ifa = sctp_add_addr_to_vrf(vrfid,
|
|
(void *)ifn,
|
|
ifn->if_index,
|
|
ifn->if_type,
|
|
ifn->if_xname,
|
|
(void *)ifa,
|
|
ifa->ifa_addr,
|
|
ifa_flags,
|
|
0);
|
|
if (sctp_ifa) {
|
|
sctp_ifa->localifa_flags &= ~SCTP_ADDR_DEFER_USE;
|
|
}
|
|
}
|
|
}
|
|
NET_EPOCH_EXIT(et);
|
|
IFNET_RUNLOCK();
|
|
}
|
|
|
|
void
|
|
sctp_init_vrf_list(int vrfid)
|
|
{
|
|
if (vrfid > SCTP_MAX_VRF_ID)
|
|
/* can't do that */
|
|
return;
|
|
|
|
/* Don't care about return here */
|
|
(void)sctp_allocate_vrf(vrfid);
|
|
|
|
/*
|
|
* Now we need to build all the ifn's for this vrf and there
|
|
* addresses
|
|
*/
|
|
sctp_init_ifns_for_vrf(vrfid);
|
|
}
|
|
|
|
void
|
|
sctp_addr_change(struct ifaddr *ifa, int cmd)
|
|
{
|
|
uint32_t ifa_flags = 0;
|
|
|
|
if (SCTP_BASE_VAR(sctp_pcb_initialized) == 0) {
|
|
return;
|
|
}
|
|
/*
|
|
* BSD only has one VRF, if this changes we will need to hook in the
|
|
* right things here to get the id to pass to the address management
|
|
* routine.
|
|
*/
|
|
if (SCTP_BASE_VAR(first_time) == 0) {
|
|
/* Special test to see if my ::1 will showup with this */
|
|
SCTP_BASE_VAR(first_time) = 1;
|
|
sctp_init_ifns_for_vrf(SCTP_DEFAULT_VRFID);
|
|
}
|
|
|
|
if ((cmd != RTM_ADD) && (cmd != RTM_DELETE)) {
|
|
/* don't know what to do with this */
|
|
return;
|
|
}
|
|
|
|
if (ifa->ifa_addr == NULL) {
|
|
return;
|
|
}
|
|
if (sctp_is_desired_interface_type(ifa->ifa_ifp) == 0) {
|
|
/* non desired type */
|
|
return;
|
|
}
|
|
switch (ifa->ifa_addr->sa_family) {
|
|
#ifdef INET
|
|
case AF_INET:
|
|
if (((struct sockaddr_in *)ifa->ifa_addr)->sin_addr.s_addr == 0) {
|
|
return;
|
|
}
|
|
break;
|
|
#endif
|
|
#ifdef INET6
|
|
case AF_INET6:
|
|
ifa_flags = ((struct in6_ifaddr *)ifa)->ia6_flags;
|
|
if (IN6_IS_ADDR_UNSPECIFIED(&((struct sockaddr_in6 *)ifa->ifa_addr)->sin6_addr)) {
|
|
/* skip unspecifed addresses */
|
|
return;
|
|
}
|
|
break;
|
|
#endif
|
|
default:
|
|
/* non inet/inet6 skip */
|
|
return;
|
|
}
|
|
if (cmd == RTM_ADD) {
|
|
(void)sctp_add_addr_to_vrf(SCTP_DEFAULT_VRFID, (void *)ifa->ifa_ifp,
|
|
ifa->ifa_ifp->if_index, ifa->ifa_ifp->if_type, ifa->ifa_ifp->if_xname,
|
|
(void *)ifa, ifa->ifa_addr, ifa_flags, 1);
|
|
} else {
|
|
sctp_del_addr_from_vrf(SCTP_DEFAULT_VRFID, ifa->ifa_addr,
|
|
ifa->ifa_ifp->if_index,
|
|
ifa->ifa_ifp->if_xname);
|
|
|
|
/*
|
|
* We don't bump refcount here so when it completes the
|
|
* final delete will happen.
|
|
*/
|
|
}
|
|
}
|
|
|
|
void
|
|
sctp_addr_change_event_handler(void *arg __unused, struct ifaddr *ifa, int cmd)
|
|
{
|
|
sctp_addr_change(ifa, cmd);
|
|
}
|
|
|
|
struct mbuf *
|
|
sctp_get_mbuf_for_msg(unsigned int space_needed, int want_header,
|
|
int how, int allonebuf, int type)
|
|
{
|
|
struct mbuf *m = NULL;
|
|
|
|
m = m_getm2(NULL, space_needed, how, type, want_header ? M_PKTHDR : 0);
|
|
if (m == NULL) {
|
|
/* bad, no memory */
|
|
return (m);
|
|
}
|
|
if (allonebuf) {
|
|
if (SCTP_BUF_SIZE(m) < space_needed) {
|
|
m_freem(m);
|
|
return (NULL);
|
|
}
|
|
KASSERT(SCTP_BUF_NEXT(m) == NULL, ("%s: no chain allowed", __func__));
|
|
}
|
|
#ifdef SCTP_MBUF_LOGGING
|
|
if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_MBUF_LOGGING_ENABLE) {
|
|
sctp_log_mb(m, SCTP_MBUF_IALLOC);
|
|
}
|
|
#endif
|
|
return (m);
|
|
}
|
|
|
|
#ifdef SCTP_PACKET_LOGGING
|
|
void
|
|
sctp_packet_log(struct mbuf *m)
|
|
{
|
|
int *lenat, thisone;
|
|
void *copyto;
|
|
uint32_t *tick_tock;
|
|
int length;
|
|
int total_len;
|
|
int grabbed_lock = 0;
|
|
int value, newval, thisend, thisbegin;
|
|
|
|
/*
|
|
* Buffer layout. -sizeof this entry (total_len) -previous end
|
|
* (value) -ticks of log (ticks) o -ip packet o -as logged -
|
|
* where this started (thisbegin) x <--end points here
|
|
*/
|
|
length = SCTP_HEADER_LEN(m);
|
|
total_len = SCTP_SIZE32((length + (4 * sizeof(int))));
|
|
/* Log a packet to the buffer. */
|
|
if (total_len > SCTP_PACKET_LOG_SIZE) {
|
|
/* Can't log this packet I have not a buffer big enough */
|
|
return;
|
|
}
|
|
if (length < (int)(SCTP_MIN_V4_OVERHEAD + sizeof(struct sctp_cookie_ack_chunk))) {
|
|
return;
|
|
}
|
|
atomic_add_int(&SCTP_BASE_VAR(packet_log_writers), 1);
|
|
try_again:
|
|
if (SCTP_BASE_VAR(packet_log_writers) > SCTP_PKTLOG_WRITERS_NEED_LOCK) {
|
|
SCTP_IP_PKTLOG_LOCK();
|
|
grabbed_lock = 1;
|
|
again_locked:
|
|
value = SCTP_BASE_VAR(packet_log_end);
|
|
newval = SCTP_BASE_VAR(packet_log_end) + total_len;
|
|
if (newval >= SCTP_PACKET_LOG_SIZE) {
|
|
/* we wrapped */
|
|
thisbegin = 0;
|
|
thisend = total_len;
|
|
} else {
|
|
thisbegin = SCTP_BASE_VAR(packet_log_end);
|
|
thisend = newval;
|
|
}
|
|
if (!(atomic_cmpset_int(&SCTP_BASE_VAR(packet_log_end), value, thisend))) {
|
|
goto again_locked;
|
|
}
|
|
} else {
|
|
value = SCTP_BASE_VAR(packet_log_end);
|
|
newval = SCTP_BASE_VAR(packet_log_end) + total_len;
|
|
if (newval >= SCTP_PACKET_LOG_SIZE) {
|
|
/* we wrapped */
|
|
thisbegin = 0;
|
|
thisend = total_len;
|
|
} else {
|
|
thisbegin = SCTP_BASE_VAR(packet_log_end);
|
|
thisend = newval;
|
|
}
|
|
if (!(atomic_cmpset_int(&SCTP_BASE_VAR(packet_log_end), value, thisend))) {
|
|
goto try_again;
|
|
}
|
|
}
|
|
/* Sanity check */
|
|
if (thisend >= SCTP_PACKET_LOG_SIZE) {
|
|
SCTP_PRINTF("Insanity stops a log thisbegin:%d thisend:%d writers:%d lock:%d end:%d\n",
|
|
thisbegin,
|
|
thisend,
|
|
SCTP_BASE_VAR(packet_log_writers),
|
|
grabbed_lock,
|
|
SCTP_BASE_VAR(packet_log_end));
|
|
SCTP_BASE_VAR(packet_log_end) = 0;
|
|
goto no_log;
|
|
}
|
|
lenat = (int *)&SCTP_BASE_VAR(packet_log_buffer)[thisbegin];
|
|
*lenat = total_len;
|
|
lenat++;
|
|
*lenat = value;
|
|
lenat++;
|
|
tick_tock = (uint32_t *)lenat;
|
|
lenat++;
|
|
*tick_tock = sctp_get_tick_count();
|
|
copyto = (void *)lenat;
|
|
thisone = thisend - sizeof(int);
|
|
lenat = (int *)&SCTP_BASE_VAR(packet_log_buffer)[thisone];
|
|
*lenat = thisbegin;
|
|
if (grabbed_lock) {
|
|
SCTP_IP_PKTLOG_UNLOCK();
|
|
grabbed_lock = 0;
|
|
}
|
|
m_copydata(m, 0, length, (caddr_t)copyto);
|
|
no_log:
|
|
if (grabbed_lock) {
|
|
SCTP_IP_PKTLOG_UNLOCK();
|
|
}
|
|
atomic_subtract_int(&SCTP_BASE_VAR(packet_log_writers), 1);
|
|
}
|
|
|
|
int
|
|
sctp_copy_out_packet_log(uint8_t *target, int length)
|
|
{
|
|
/*
|
|
* We wind through the packet log starting at start copying up to
|
|
* length bytes out. We return the number of bytes copied.
|
|
*/
|
|
int tocopy, this_copy;
|
|
int *lenat;
|
|
int did_delay = 0;
|
|
|
|
tocopy = length;
|
|
if (length < (int)(2 * sizeof(int))) {
|
|
/* not enough room */
|
|
return (0);
|
|
}
|
|
if (SCTP_PKTLOG_WRITERS_NEED_LOCK) {
|
|
atomic_add_int(&SCTP_BASE_VAR(packet_log_writers), SCTP_PKTLOG_WRITERS_NEED_LOCK);
|
|
again:
|
|
if ((did_delay == 0) && (SCTP_BASE_VAR(packet_log_writers) != SCTP_PKTLOG_WRITERS_NEED_LOCK)) {
|
|
/*
|
|
* we delay here for just a moment hoping the
|
|
* writer(s) that were present when we entered will
|
|
* have left and we only have locking ones that will
|
|
* contend with us for the lock. This does not
|
|
* assure 100% access, but its good enough for a
|
|
* logging facility like this.
|
|
*/
|
|
did_delay = 1;
|
|
DELAY(10);
|
|
goto again;
|
|
}
|
|
}
|
|
SCTP_IP_PKTLOG_LOCK();
|
|
lenat = (int *)target;
|
|
*lenat = SCTP_BASE_VAR(packet_log_end);
|
|
lenat++;
|
|
this_copy = min((length - sizeof(int)), SCTP_PACKET_LOG_SIZE);
|
|
memcpy((void *)lenat, (void *)SCTP_BASE_VAR(packet_log_buffer), this_copy);
|
|
if (SCTP_PKTLOG_WRITERS_NEED_LOCK) {
|
|
atomic_subtract_int(&SCTP_BASE_VAR(packet_log_writers),
|
|
SCTP_PKTLOG_WRITERS_NEED_LOCK);
|
|
}
|
|
SCTP_IP_PKTLOG_UNLOCK();
|
|
return (this_copy + sizeof(int));
|
|
}
|
|
|
|
#endif
|