1676 lines
40 KiB
C
1676 lines
40 KiB
C
/**************************************************************************
|
|
|
|
Copyright (c) 2008-2010, BitGravity Inc.
|
|
All rights reserved.
|
|
|
|
Redistribution and use in source and binary forms, with or without
|
|
modification, are permitted provided that the following conditions are met:
|
|
|
|
1. Redistributions of source code must retain the above copyright notice,
|
|
this list of conditions and the following disclaimer.
|
|
|
|
2. Neither the name of the BitGravity Corporation nor the names of its
|
|
contributors may be used to endorse or promote products derived from
|
|
this software without specific prior written permission.
|
|
|
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
|
|
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
***************************************************************************/
|
|
|
|
#include "opt_route.h"
|
|
#include "opt_mpath.h"
|
|
#include "opt_ddb.h"
|
|
#include "opt_inet.h"
|
|
#include "opt_inet6.h"
|
|
|
|
#include <sys/cdefs.h>
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/types.h>
|
|
#include <sys/bitstring.h>
|
|
#include <sys/condvar.h>
|
|
#include <sys/callout.h>
|
|
#include <sys/hash.h>
|
|
#include <sys/kernel.h>
|
|
#include <sys/kthread.h>
|
|
#include <sys/limits.h>
|
|
#include <sys/malloc.h>
|
|
#include <sys/mbuf.h>
|
|
#include <sys/proc.h>
|
|
#include <sys/sbuf.h>
|
|
#include <sys/sched.h>
|
|
#include <sys/smp.h>
|
|
#include <sys/socket.h>
|
|
#include <sys/syslog.h>
|
|
#include <sys/sysctl.h>
|
|
|
|
#include <net/if.h>
|
|
#include <net/if_llatbl.h>
|
|
#include <net/if_var.h>
|
|
#include <net/route.h>
|
|
#include <net/flowtable.h>
|
|
#include <net/vnet.h>
|
|
|
|
#include <netinet/in.h>
|
|
#include <netinet/in_systm.h>
|
|
#include <netinet/in_var.h>
|
|
#include <netinet/if_ether.h>
|
|
#include <netinet/ip.h>
|
|
#ifdef INET6
|
|
#include <netinet/ip6.h>
|
|
#endif
|
|
#include <netinet/tcp.h>
|
|
#include <netinet/udp.h>
|
|
#include <netinet/sctp.h>
|
|
|
|
#include <ddb/ddb.h>
|
|
|
|
struct ipv4_tuple {
|
|
uint16_t ip_sport; /* source port */
|
|
uint16_t ip_dport; /* destination port */
|
|
in_addr_t ip_saddr; /* source address */
|
|
in_addr_t ip_daddr; /* destination address */
|
|
};
|
|
|
|
union ipv4_flow {
|
|
struct ipv4_tuple ipf_ipt;
|
|
uint32_t ipf_key[3];
|
|
};
|
|
|
|
struct ipv6_tuple {
|
|
uint16_t ip_sport; /* source port */
|
|
uint16_t ip_dport; /* destination port */
|
|
struct in6_addr ip_saddr; /* source address */
|
|
struct in6_addr ip_daddr; /* destination address */
|
|
};
|
|
|
|
union ipv6_flow {
|
|
struct ipv6_tuple ipf_ipt;
|
|
uint32_t ipf_key[9];
|
|
};
|
|
|
|
struct flentry {
|
|
volatile uint32_t f_fhash; /* hash flowing forward */
|
|
uint16_t f_flags; /* flow flags */
|
|
uint8_t f_pad;
|
|
uint8_t f_proto; /* protocol */
|
|
uint32_t f_fibnum; /* fib index */
|
|
uint32_t f_uptime; /* uptime at last access */
|
|
struct flentry *f_next; /* pointer to collision entry */
|
|
volatile struct rtentry *f_rt; /* rtentry for flow */
|
|
volatile struct llentry *f_lle; /* llentry for flow */
|
|
};
|
|
|
|
struct flentry_v4 {
|
|
struct flentry fl_entry;
|
|
union ipv4_flow fl_flow;
|
|
};
|
|
|
|
struct flentry_v6 {
|
|
struct flentry fl_entry;
|
|
union ipv6_flow fl_flow;
|
|
};
|
|
|
|
#define SECS_PER_HOUR 3600
|
|
#define SECS_PER_DAY (24*SECS_PER_HOUR)
|
|
|
|
#define SYN_IDLE 300
|
|
#define UDP_IDLE 300
|
|
#define FIN_WAIT_IDLE 600
|
|
#define TCP_IDLE SECS_PER_DAY
|
|
|
|
|
|
typedef void fl_lock_t(struct flowtable *, uint32_t);
|
|
typedef void fl_rtalloc_t(struct route *, uint32_t, u_int);
|
|
|
|
union flentryp {
|
|
struct flentry **global;
|
|
struct flentry **pcpu[MAXCPU];
|
|
};
|
|
|
|
struct flowtable {
|
|
counter_u64_t *ft_stat;
|
|
uma_zone_t ft_zone;
|
|
int ft_size;
|
|
int ft_lock_count;
|
|
uint32_t ft_flags;
|
|
uint32_t ft_max_depth;
|
|
fl_lock_t *ft_lock;
|
|
fl_lock_t *ft_unlock;
|
|
fl_rtalloc_t *ft_rtalloc;
|
|
/*
|
|
* XXX need to pad out
|
|
*/
|
|
struct mtx *ft_locks;
|
|
union flentryp ft_table;
|
|
bitstr_t *ft_masks[MAXCPU];
|
|
bitstr_t *ft_tmpmask;
|
|
|
|
uint32_t ft_udp_idle __aligned(CACHE_LINE_SIZE);
|
|
uint32_t ft_fin_wait_idle;
|
|
uint32_t ft_syn_idle;
|
|
uint32_t ft_tcp_idle;
|
|
boolean_t ft_full;
|
|
} __aligned(CACHE_LINE_SIZE);
|
|
|
|
#define FLOWSTAT_ADD(ft, name, v) \
|
|
counter_u64_add((ft)->ft_stat[offsetof(struct flowtable_stat, name) / sizeof(uint64_t)], (v))
|
|
#define FLOWSTAT_INC(ft, name) FLOWSTAT_ADD(ft, name, 1)
|
|
|
|
static struct proc *flowcleanerproc;
|
|
static uint32_t flow_hashjitter;
|
|
|
|
static struct cv flowclean_f_cv;
|
|
static struct cv flowclean_c_cv;
|
|
static struct mtx flowclean_lock;
|
|
static uint32_t flowclean_cycles;
|
|
static uint32_t flowclean_freq;
|
|
|
|
/*
|
|
* TODO:
|
|
* - add sysctls to resize && flush flow tables
|
|
* - Add per flowtable sysctls for statistics and configuring timeouts
|
|
* - add saturation counter to rtentry to support per-packet load-balancing
|
|
* add flag to indicate round-robin flow, add list lookup from head
|
|
for flows
|
|
* - add sysctl / device node / syscall to support exporting and importing
|
|
* of flows with flag to indicate that a flow was imported so should
|
|
* not be considered for auto-cleaning
|
|
* - support explicit connection state (currently only ad-hoc for DSR)
|
|
* - idetach() cleanup for options VIMAGE builds.
|
|
*/
|
|
#ifdef INET
|
|
static VNET_DEFINE(struct flowtable, ip4_ft);
|
|
#define V_ip4_ft VNET(ip4_ft)
|
|
static uma_zone_t flow_ipv4_zone;
|
|
#endif
|
|
#ifdef INET6
|
|
static VNET_DEFINE(struct flowtable, ip6_ft);
|
|
#define V_ip6_ft VNET(ip6_ft)
|
|
static uma_zone_t flow_ipv6_zone;
|
|
#endif
|
|
|
|
static VNET_DEFINE(int, flowtable_enable) = 1;
|
|
static VNET_DEFINE(int, flowtable_syn_expire) = SYN_IDLE;
|
|
static VNET_DEFINE(int, flowtable_udp_expire) = UDP_IDLE;
|
|
static VNET_DEFINE(int, flowtable_fin_wait_expire) = FIN_WAIT_IDLE;
|
|
static VNET_DEFINE(int, flowtable_tcp_expire) = TCP_IDLE;
|
|
|
|
#define V_flowtable_enable VNET(flowtable_enable)
|
|
#define V_flowtable_syn_expire VNET(flowtable_syn_expire)
|
|
#define V_flowtable_udp_expire VNET(flowtable_udp_expire)
|
|
#define V_flowtable_fin_wait_expire VNET(flowtable_fin_wait_expire)
|
|
#define V_flowtable_tcp_expire VNET(flowtable_tcp_expire)
|
|
|
|
static SYSCTL_NODE(_net, OID_AUTO, flowtable, CTLFLAG_RD, NULL,
|
|
"flowtable");
|
|
SYSCTL_VNET_INT(_net_flowtable, OID_AUTO, enable, CTLFLAG_RW,
|
|
&VNET_NAME(flowtable_enable), 0, "enable flowtable caching.");
|
|
|
|
/*
|
|
* XXX This does not end up updating timeouts at runtime
|
|
* and only reflects the value for the last table added :-/
|
|
*/
|
|
SYSCTL_VNET_INT(_net_flowtable, OID_AUTO, syn_expire, CTLFLAG_RW,
|
|
&VNET_NAME(flowtable_syn_expire), 0,
|
|
"seconds after which to remove syn allocated flow.");
|
|
SYSCTL_VNET_INT(_net_flowtable, OID_AUTO, udp_expire, CTLFLAG_RW,
|
|
&VNET_NAME(flowtable_udp_expire), 0,
|
|
"seconds after which to remove flow allocated to UDP.");
|
|
SYSCTL_VNET_INT(_net_flowtable, OID_AUTO, fin_wait_expire, CTLFLAG_RW,
|
|
&VNET_NAME(flowtable_fin_wait_expire), 0,
|
|
"seconds after which to remove a flow in FIN_WAIT.");
|
|
SYSCTL_VNET_INT(_net_flowtable, OID_AUTO, tcp_expire, CTLFLAG_RW,
|
|
&VNET_NAME(flowtable_tcp_expire), 0,
|
|
"seconds after which to remove flow allocated to a TCP connection.");
|
|
|
|
#ifndef RADIX_MPATH
|
|
static void
|
|
rtalloc_ign_wrapper(struct route *ro, uint32_t hash, u_int fibnum)
|
|
{
|
|
|
|
rtalloc_ign_fib(ro, 0, fibnum);
|
|
}
|
|
#endif
|
|
|
|
static void
|
|
flowtable_global_lock(struct flowtable *table, uint32_t hash)
|
|
{
|
|
int lock_index = (hash)&(table->ft_lock_count - 1);
|
|
|
|
mtx_lock(&table->ft_locks[lock_index]);
|
|
}
|
|
|
|
static void
|
|
flowtable_global_unlock(struct flowtable *table, uint32_t hash)
|
|
{
|
|
int lock_index = (hash)&(table->ft_lock_count - 1);
|
|
|
|
mtx_unlock(&table->ft_locks[lock_index]);
|
|
}
|
|
|
|
static void
|
|
flowtable_pcpu_lock(struct flowtable *table, uint32_t hash)
|
|
{
|
|
|
|
critical_enter();
|
|
}
|
|
|
|
static void
|
|
flowtable_pcpu_unlock(struct flowtable *table, uint32_t hash)
|
|
{
|
|
|
|
critical_exit();
|
|
}
|
|
|
|
#define FL_ENTRY_INDEX(table, hash)((hash) % (table)->ft_size)
|
|
#define FL_ENTRY(table, hash) *flowtable_entry((table), (hash))
|
|
#define FL_ENTRY_LOCK(table, hash) (table)->ft_lock((table), (hash))
|
|
#define FL_ENTRY_UNLOCK(table, hash) (table)->ft_unlock((table), (hash))
|
|
|
|
#define FL_STALE (1<<8)
|
|
|
|
static struct flentry *flowtable_lookup_common(struct flowtable *,
|
|
struct sockaddr_storage *, struct sockaddr_storage *, struct mbuf *, int);
|
|
|
|
static __inline int
|
|
proto_to_flags(uint8_t proto)
|
|
{
|
|
int flag;
|
|
|
|
switch (proto) {
|
|
case IPPROTO_TCP:
|
|
flag = FL_TCP;
|
|
break;
|
|
case IPPROTO_SCTP:
|
|
flag = FL_SCTP;
|
|
break;
|
|
case IPPROTO_UDP:
|
|
flag = FL_UDP;
|
|
break;
|
|
default:
|
|
flag = 0;
|
|
break;
|
|
}
|
|
|
|
return (flag);
|
|
}
|
|
|
|
static __inline int
|
|
flags_to_proto(int flags)
|
|
{
|
|
int proto, protoflags;
|
|
|
|
protoflags = flags & (FL_TCP|FL_SCTP|FL_UDP);
|
|
switch (protoflags) {
|
|
case FL_TCP:
|
|
proto = IPPROTO_TCP;
|
|
break;
|
|
case FL_SCTP:
|
|
proto = IPPROTO_SCTP;
|
|
break;
|
|
case FL_UDP:
|
|
proto = IPPROTO_UDP;
|
|
break;
|
|
default:
|
|
proto = 0;
|
|
break;
|
|
}
|
|
return (proto);
|
|
}
|
|
|
|
#ifdef INET
|
|
#ifdef FLOWTABLE_DEBUG
|
|
static void
|
|
ipv4_flow_print_tuple(int flags, int proto, struct sockaddr_in *ssin,
|
|
struct sockaddr_in *dsin)
|
|
{
|
|
char saddr[4*sizeof "123"], daddr[4*sizeof "123"];
|
|
|
|
if (flags & FL_HASH_ALL) {
|
|
inet_ntoa_r(ssin->sin_addr, saddr);
|
|
inet_ntoa_r(dsin->sin_addr, daddr);
|
|
printf("proto=%d %s:%d->%s:%d\n",
|
|
proto, saddr, ntohs(ssin->sin_port), daddr,
|
|
ntohs(dsin->sin_port));
|
|
} else {
|
|
inet_ntoa_r(*(struct in_addr *) &dsin->sin_addr, daddr);
|
|
printf("proto=%d %s\n", proto, daddr);
|
|
}
|
|
|
|
}
|
|
#endif
|
|
|
|
static int
|
|
ipv4_mbuf_demarshal(struct mbuf *m, struct sockaddr_in *ssin,
|
|
struct sockaddr_in *dsin, uint16_t *flags)
|
|
{
|
|
struct ip *ip;
|
|
uint8_t proto;
|
|
int iphlen;
|
|
struct tcphdr *th;
|
|
struct udphdr *uh;
|
|
struct sctphdr *sh;
|
|
uint16_t sport, dport;
|
|
|
|
proto = sport = dport = 0;
|
|
ip = mtod(m, struct ip *);
|
|
dsin->sin_family = AF_INET;
|
|
dsin->sin_len = sizeof(*dsin);
|
|
dsin->sin_addr = ip->ip_dst;
|
|
ssin->sin_family = AF_INET;
|
|
ssin->sin_len = sizeof(*ssin);
|
|
ssin->sin_addr = ip->ip_src;
|
|
|
|
proto = ip->ip_p;
|
|
if ((*flags & FL_HASH_ALL) == 0)
|
|
goto skipports;
|
|
|
|
iphlen = ip->ip_hl << 2; /* XXX options? */
|
|
|
|
switch (proto) {
|
|
case IPPROTO_TCP:
|
|
th = (struct tcphdr *)((caddr_t)ip + iphlen);
|
|
sport = th->th_sport;
|
|
dport = th->th_dport;
|
|
if ((*flags & FL_HASH_ALL) &&
|
|
(th->th_flags & (TH_RST|TH_FIN)))
|
|
*flags |= FL_STALE;
|
|
break;
|
|
case IPPROTO_UDP:
|
|
uh = (struct udphdr *)((caddr_t)ip + iphlen);
|
|
sport = uh->uh_sport;
|
|
dport = uh->uh_dport;
|
|
break;
|
|
case IPPROTO_SCTP:
|
|
sh = (struct sctphdr *)((caddr_t)ip + iphlen);
|
|
sport = sh->src_port;
|
|
dport = sh->dest_port;
|
|
break;
|
|
default:
|
|
return (ENOTSUP);
|
|
/* no port - hence not a protocol we care about */
|
|
break;
|
|
|
|
}
|
|
|
|
skipports:
|
|
*flags |= proto_to_flags(proto);
|
|
ssin->sin_port = sport;
|
|
dsin->sin_port = dport;
|
|
return (0);
|
|
}
|
|
|
|
static uint32_t
|
|
ipv4_flow_lookup_hash(
|
|
struct sockaddr_in *ssin, struct sockaddr_in *dsin,
|
|
uint32_t *key, uint16_t flags)
|
|
{
|
|
uint16_t sport, dport;
|
|
uint8_t proto;
|
|
int offset = 0;
|
|
|
|
proto = flags_to_proto(flags);
|
|
sport = dport = key[2] = key[1] = key[0] = 0;
|
|
if ((ssin != NULL) && (flags & FL_HASH_ALL)) {
|
|
key[1] = ssin->sin_addr.s_addr;
|
|
sport = ssin->sin_port;
|
|
}
|
|
if (dsin != NULL) {
|
|
key[2] = dsin->sin_addr.s_addr;
|
|
dport = dsin->sin_port;
|
|
}
|
|
if (flags & FL_HASH_ALL) {
|
|
((uint16_t *)key)[0] = sport;
|
|
((uint16_t *)key)[1] = dport;
|
|
} else
|
|
offset = flow_hashjitter + proto;
|
|
|
|
return (jenkins_hash32(key, 3, offset));
|
|
}
|
|
|
|
static struct flentry *
|
|
flowtable_lookup_ipv4(struct mbuf *m)
|
|
{
|
|
struct sockaddr_storage ssa, dsa;
|
|
uint16_t flags;
|
|
struct sockaddr_in *dsin, *ssin;
|
|
|
|
dsin = (struct sockaddr_in *)&dsa;
|
|
ssin = (struct sockaddr_in *)&ssa;
|
|
bzero(dsin, sizeof(*dsin));
|
|
bzero(ssin, sizeof(*ssin));
|
|
flags = V_ip4_ft.ft_flags;
|
|
if (ipv4_mbuf_demarshal(m, ssin, dsin, &flags) != 0)
|
|
return (NULL);
|
|
|
|
return (flowtable_lookup_common(&V_ip4_ft, &ssa, &dsa, m, flags));
|
|
}
|
|
|
|
void
|
|
flow_to_route(struct flentry *fle, struct route *ro)
|
|
{
|
|
uint32_t *hashkey = NULL;
|
|
struct sockaddr_in *sin;
|
|
|
|
sin = (struct sockaddr_in *)&ro->ro_dst;
|
|
sin->sin_family = AF_INET;
|
|
sin->sin_len = sizeof(*sin);
|
|
hashkey = ((struct flentry_v4 *)fle)->fl_flow.ipf_key;
|
|
sin->sin_addr.s_addr = hashkey[2];
|
|
ro->ro_rt = __DEVOLATILE(struct rtentry *, fle->f_rt);
|
|
ro->ro_lle = __DEVOLATILE(struct llentry *, fle->f_lle);
|
|
ro->ro_flags |= RT_NORTREF;
|
|
}
|
|
#endif /* INET */
|
|
|
|
#ifdef INET6
|
|
/*
|
|
* PULLUP_TO(len, p, T) makes sure that len + sizeof(T) is contiguous,
|
|
* then it sets p to point at the offset "len" in the mbuf. WARNING: the
|
|
* pointer might become stale after other pullups (but we never use it
|
|
* this way).
|
|
*/
|
|
#define PULLUP_TO(_len, p, T) \
|
|
do { \
|
|
int x = (_len) + sizeof(T); \
|
|
if ((m)->m_len < x) { \
|
|
goto receive_failed; \
|
|
} \
|
|
p = (mtod(m, char *) + (_len)); \
|
|
} while (0)
|
|
|
|
#define TCP(p) ((struct tcphdr *)(p))
|
|
#define SCTP(p) ((struct sctphdr *)(p))
|
|
#define UDP(p) ((struct udphdr *)(p))
|
|
|
|
static int
|
|
ipv6_mbuf_demarshal(struct mbuf *m, struct sockaddr_in6 *ssin6,
|
|
struct sockaddr_in6 *dsin6, uint16_t *flags)
|
|
{
|
|
struct ip6_hdr *ip6;
|
|
uint8_t proto;
|
|
int hlen;
|
|
uint16_t src_port, dst_port;
|
|
u_short offset;
|
|
void *ulp;
|
|
|
|
offset = hlen = src_port = dst_port = 0;
|
|
ulp = NULL;
|
|
ip6 = mtod(m, struct ip6_hdr *);
|
|
hlen = sizeof(struct ip6_hdr);
|
|
proto = ip6->ip6_nxt;
|
|
|
|
if ((*flags & FL_HASH_ALL) == 0)
|
|
goto skipports;
|
|
|
|
while (ulp == NULL) {
|
|
switch (proto) {
|
|
case IPPROTO_ICMPV6:
|
|
case IPPROTO_OSPFIGP:
|
|
case IPPROTO_PIM:
|
|
case IPPROTO_CARP:
|
|
case IPPROTO_ESP:
|
|
case IPPROTO_NONE:
|
|
ulp = ip6;
|
|
break;
|
|
case IPPROTO_TCP:
|
|
PULLUP_TO(hlen, ulp, struct tcphdr);
|
|
dst_port = TCP(ulp)->th_dport;
|
|
src_port = TCP(ulp)->th_sport;
|
|
if ((*flags & FL_HASH_ALL) &&
|
|
(TCP(ulp)->th_flags & (TH_RST|TH_FIN)))
|
|
*flags |= FL_STALE;
|
|
break;
|
|
case IPPROTO_SCTP:
|
|
PULLUP_TO(hlen, ulp, struct sctphdr);
|
|
src_port = SCTP(ulp)->src_port;
|
|
dst_port = SCTP(ulp)->dest_port;
|
|
break;
|
|
case IPPROTO_UDP:
|
|
PULLUP_TO(hlen, ulp, struct udphdr);
|
|
dst_port = UDP(ulp)->uh_dport;
|
|
src_port = UDP(ulp)->uh_sport;
|
|
break;
|
|
case IPPROTO_HOPOPTS: /* RFC 2460 */
|
|
PULLUP_TO(hlen, ulp, struct ip6_hbh);
|
|
hlen += (((struct ip6_hbh *)ulp)->ip6h_len + 1) << 3;
|
|
proto = ((struct ip6_hbh *)ulp)->ip6h_nxt;
|
|
ulp = NULL;
|
|
break;
|
|
case IPPROTO_ROUTING: /* RFC 2460 */
|
|
PULLUP_TO(hlen, ulp, struct ip6_rthdr);
|
|
hlen += (((struct ip6_rthdr *)ulp)->ip6r_len + 1) << 3;
|
|
proto = ((struct ip6_rthdr *)ulp)->ip6r_nxt;
|
|
ulp = NULL;
|
|
break;
|
|
case IPPROTO_FRAGMENT: /* RFC 2460 */
|
|
PULLUP_TO(hlen, ulp, struct ip6_frag);
|
|
hlen += sizeof (struct ip6_frag);
|
|
proto = ((struct ip6_frag *)ulp)->ip6f_nxt;
|
|
offset = ((struct ip6_frag *)ulp)->ip6f_offlg &
|
|
IP6F_OFF_MASK;
|
|
ulp = NULL;
|
|
break;
|
|
case IPPROTO_DSTOPTS: /* RFC 2460 */
|
|
PULLUP_TO(hlen, ulp, struct ip6_hbh);
|
|
hlen += (((struct ip6_hbh *)ulp)->ip6h_len + 1) << 3;
|
|
proto = ((struct ip6_hbh *)ulp)->ip6h_nxt;
|
|
ulp = NULL;
|
|
break;
|
|
case IPPROTO_AH: /* RFC 2402 */
|
|
PULLUP_TO(hlen, ulp, struct ip6_ext);
|
|
hlen += (((struct ip6_ext *)ulp)->ip6e_len + 2) << 2;
|
|
proto = ((struct ip6_ext *)ulp)->ip6e_nxt;
|
|
ulp = NULL;
|
|
break;
|
|
default:
|
|
PULLUP_TO(hlen, ulp, struct ip6_ext);
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (src_port == 0) {
|
|
receive_failed:
|
|
return (ENOTSUP);
|
|
}
|
|
|
|
skipports:
|
|
dsin6->sin6_family = AF_INET6;
|
|
dsin6->sin6_len = sizeof(*dsin6);
|
|
dsin6->sin6_port = dst_port;
|
|
memcpy(&dsin6->sin6_addr, &ip6->ip6_dst, sizeof(struct in6_addr));
|
|
|
|
ssin6->sin6_family = AF_INET6;
|
|
ssin6->sin6_len = sizeof(*ssin6);
|
|
ssin6->sin6_port = src_port;
|
|
memcpy(&ssin6->sin6_addr, &ip6->ip6_src, sizeof(struct in6_addr));
|
|
*flags |= proto_to_flags(proto);
|
|
|
|
return (0);
|
|
}
|
|
|
|
#define zero_key(key) \
|
|
do { \
|
|
key[0] = 0; \
|
|
key[1] = 0; \
|
|
key[2] = 0; \
|
|
key[3] = 0; \
|
|
key[4] = 0; \
|
|
key[5] = 0; \
|
|
key[6] = 0; \
|
|
key[7] = 0; \
|
|
key[8] = 0; \
|
|
} while (0)
|
|
|
|
static uint32_t
|
|
ipv6_flow_lookup_hash(
|
|
struct sockaddr_in6 *ssin6, struct sockaddr_in6 *dsin6,
|
|
uint32_t *key, uint16_t flags)
|
|
{
|
|
uint16_t sport, dport;
|
|
uint8_t proto;
|
|
int offset = 0;
|
|
|
|
proto = flags_to_proto(flags);
|
|
zero_key(key);
|
|
sport = dport = 0;
|
|
if (dsin6 != NULL) {
|
|
memcpy(&key[1], &dsin6->sin6_addr, sizeof(struct in6_addr));
|
|
dport = dsin6->sin6_port;
|
|
}
|
|
if ((ssin6 != NULL) && (flags & FL_HASH_ALL)) {
|
|
memcpy(&key[5], &ssin6->sin6_addr, sizeof(struct in6_addr));
|
|
sport = ssin6->sin6_port;
|
|
}
|
|
if (flags & FL_HASH_ALL) {
|
|
((uint16_t *)key)[0] = sport;
|
|
((uint16_t *)key)[1] = dport;
|
|
} else
|
|
offset = flow_hashjitter + proto;
|
|
|
|
return (jenkins_hash32(key, 9, offset));
|
|
}
|
|
|
|
static struct flentry *
|
|
flowtable_lookup_ipv6(struct mbuf *m)
|
|
{
|
|
struct sockaddr_storage ssa, dsa;
|
|
struct sockaddr_in6 *dsin6, *ssin6;
|
|
uint16_t flags;
|
|
|
|
dsin6 = (struct sockaddr_in6 *)&dsa;
|
|
ssin6 = (struct sockaddr_in6 *)&ssa;
|
|
bzero(dsin6, sizeof(*dsin6));
|
|
bzero(ssin6, sizeof(*ssin6));
|
|
flags = V_ip6_ft.ft_flags;
|
|
|
|
if (ipv6_mbuf_demarshal(m, ssin6, dsin6, &flags) != 0)
|
|
return (NULL);
|
|
|
|
return (flowtable_lookup_common(&V_ip6_ft, &ssa, &dsa, m, flags));
|
|
}
|
|
|
|
void
|
|
flow_to_route_in6(struct flentry *fle, struct route_in6 *ro)
|
|
{
|
|
uint32_t *hashkey = NULL;
|
|
struct sockaddr_in6 *sin6;
|
|
|
|
sin6 = (struct sockaddr_in6 *)&ro->ro_dst;
|
|
|
|
sin6->sin6_family = AF_INET6;
|
|
sin6->sin6_len = sizeof(*sin6);
|
|
hashkey = ((struct flentry_v6 *)fle)->fl_flow.ipf_key;
|
|
memcpy(&sin6->sin6_addr, &hashkey[5], sizeof (struct in6_addr));
|
|
ro->ro_rt = __DEVOLATILE(struct rtentry *, fle->f_rt);
|
|
ro->ro_lle = __DEVOLATILE(struct llentry *, fle->f_lle);
|
|
ro->ro_flags |= RT_NORTREF;
|
|
}
|
|
#endif /* INET6 */
|
|
|
|
static bitstr_t *
|
|
flowtable_mask(struct flowtable *ft)
|
|
{
|
|
bitstr_t *mask;
|
|
|
|
if (ft->ft_flags & FL_PCPU)
|
|
mask = ft->ft_masks[curcpu];
|
|
else
|
|
mask = ft->ft_masks[0];
|
|
|
|
return (mask);
|
|
}
|
|
|
|
static struct flentry **
|
|
flowtable_entry(struct flowtable *ft, uint32_t hash)
|
|
{
|
|
struct flentry **fle;
|
|
int index = (hash % ft->ft_size);
|
|
|
|
if (ft->ft_flags & FL_PCPU) {
|
|
KASSERT(&ft->ft_table.pcpu[curcpu][0] != NULL, ("pcpu not set"));
|
|
fle = &ft->ft_table.pcpu[curcpu][index];
|
|
} else {
|
|
KASSERT(&ft->ft_table.global[0] != NULL, ("global not set"));
|
|
fle = &ft->ft_table.global[index];
|
|
}
|
|
|
|
return (fle);
|
|
}
|
|
|
|
static int
|
|
flow_stale(struct flowtable *ft, struct flentry *fle)
|
|
{
|
|
time_t idle_time;
|
|
|
|
if ((fle->f_fhash == 0)
|
|
|| ((fle->f_rt->rt_flags & RTF_HOST) &&
|
|
((fle->f_rt->rt_flags & (RTF_UP))
|
|
!= (RTF_UP)))
|
|
|| (fle->f_rt->rt_ifp == NULL)
|
|
|| !RT_LINK_IS_UP(fle->f_rt->rt_ifp))
|
|
return (1);
|
|
|
|
idle_time = time_uptime - fle->f_uptime;
|
|
|
|
if ((fle->f_flags & FL_STALE) ||
|
|
((fle->f_flags & (TH_SYN|TH_ACK|TH_FIN)) == 0
|
|
&& (idle_time > ft->ft_udp_idle)) ||
|
|
((fle->f_flags & TH_FIN)
|
|
&& (idle_time > ft->ft_fin_wait_idle)) ||
|
|
((fle->f_flags & (TH_SYN|TH_ACK)) == TH_SYN
|
|
&& (idle_time > ft->ft_syn_idle)) ||
|
|
((fle->f_flags & (TH_SYN|TH_ACK)) == (TH_SYN|TH_ACK)
|
|
&& (idle_time > ft->ft_tcp_idle)) ||
|
|
((fle->f_rt->rt_flags & RTF_UP) == 0 ||
|
|
(fle->f_rt->rt_ifp == NULL)))
|
|
return (1);
|
|
|
|
return (0);
|
|
}
|
|
|
|
static void
|
|
flowtable_set_hashkey(struct flentry *fle, uint32_t *key)
|
|
{
|
|
uint32_t *hashkey;
|
|
int i, nwords;
|
|
|
|
if (fle->f_flags & FL_IPV6) {
|
|
nwords = 9;
|
|
hashkey = ((struct flentry_v4 *)fle)->fl_flow.ipf_key;
|
|
} else {
|
|
nwords = 3;
|
|
hashkey = ((struct flentry_v6 *)fle)->fl_flow.ipf_key;
|
|
}
|
|
|
|
for (i = 0; i < nwords; i++)
|
|
hashkey[i] = key[i];
|
|
}
|
|
|
|
static int
|
|
flow_full(struct flowtable *ft)
|
|
{
|
|
boolean_t full;
|
|
int count, max;
|
|
|
|
full = ft->ft_full;
|
|
count = uma_zone_get_cur(ft->ft_zone);
|
|
max = uma_zone_get_max(ft->ft_zone);
|
|
|
|
if (full && (count < (max - (max >> 3))))
|
|
ft->ft_full = FALSE;
|
|
else if (!full && (count > (max - (max >> 5))))
|
|
ft->ft_full = TRUE;
|
|
|
|
if (full && !ft->ft_full) {
|
|
flowclean_freq = 4*hz;
|
|
if ((ft->ft_flags & FL_HASH_ALL) == 0)
|
|
ft->ft_udp_idle = ft->ft_fin_wait_idle =
|
|
ft->ft_syn_idle = ft->ft_tcp_idle = 5;
|
|
cv_broadcast(&flowclean_c_cv);
|
|
} else if (!full && ft->ft_full) {
|
|
flowclean_freq = 20*hz;
|
|
if ((ft->ft_flags & FL_HASH_ALL) == 0)
|
|
ft->ft_udp_idle = ft->ft_fin_wait_idle =
|
|
ft->ft_syn_idle = ft->ft_tcp_idle = 30;
|
|
}
|
|
|
|
return (ft->ft_full);
|
|
}
|
|
|
|
static int
|
|
flowtable_insert(struct flowtable *ft, uint32_t hash, uint32_t *key,
|
|
uint32_t fibnum, struct route *ro, uint16_t flags)
|
|
{
|
|
struct flentry *fle, *fletail, *newfle, **flep;
|
|
int depth;
|
|
bitstr_t *mask;
|
|
uint8_t proto;
|
|
|
|
newfle = uma_zalloc(ft->ft_zone, M_NOWAIT | M_ZERO);
|
|
if (newfle == NULL)
|
|
return (ENOMEM);
|
|
|
|
newfle->f_flags |= (flags & FL_IPV6);
|
|
proto = flags_to_proto(flags);
|
|
|
|
FL_ENTRY_LOCK(ft, hash);
|
|
mask = flowtable_mask(ft);
|
|
flep = flowtable_entry(ft, hash);
|
|
fletail = fle = *flep;
|
|
|
|
if (fle == NULL) {
|
|
bit_set(mask, FL_ENTRY_INDEX(ft, hash));
|
|
*flep = fle = newfle;
|
|
goto skip;
|
|
}
|
|
|
|
depth = 0;
|
|
FLOWSTAT_INC(ft, ft_collisions);
|
|
/*
|
|
* find end of list and make sure that we were not
|
|
* preempted by another thread handling this flow
|
|
*/
|
|
while (fle != NULL) {
|
|
if (fle->f_fhash == hash && !flow_stale(ft, fle)) {
|
|
/*
|
|
* there was either a hash collision
|
|
* or we lost a race to insert
|
|
*/
|
|
FL_ENTRY_UNLOCK(ft, hash);
|
|
uma_zfree(ft->ft_zone, newfle);
|
|
|
|
return (EEXIST);
|
|
}
|
|
/*
|
|
* re-visit this double condition XXX
|
|
*/
|
|
if (fletail->f_next != NULL)
|
|
fletail = fle->f_next;
|
|
|
|
depth++;
|
|
fle = fle->f_next;
|
|
}
|
|
|
|
if (depth > ft->ft_max_depth)
|
|
ft->ft_max_depth = depth;
|
|
fletail->f_next = newfle;
|
|
fle = newfle;
|
|
skip:
|
|
flowtable_set_hashkey(fle, key);
|
|
|
|
fle->f_proto = proto;
|
|
fle->f_rt = ro->ro_rt;
|
|
fle->f_lle = ro->ro_lle;
|
|
fle->f_fhash = hash;
|
|
fle->f_fibnum = fibnum;
|
|
fle->f_uptime = time_uptime;
|
|
FL_ENTRY_UNLOCK(ft, hash);
|
|
return (0);
|
|
}
|
|
|
|
static int
|
|
flowtable_key_equal(struct flentry *fle, uint32_t *key)
|
|
{
|
|
uint32_t *hashkey;
|
|
int i, nwords;
|
|
|
|
if (fle->f_flags & FL_IPV6) {
|
|
nwords = 9;
|
|
hashkey = ((struct flentry_v4 *)fle)->fl_flow.ipf_key;
|
|
} else {
|
|
nwords = 3;
|
|
hashkey = ((struct flentry_v6 *)fle)->fl_flow.ipf_key;
|
|
}
|
|
|
|
for (i = 0; i < nwords; i++)
|
|
if (hashkey[i] != key[i])
|
|
return (0);
|
|
|
|
return (1);
|
|
}
|
|
|
|
struct flentry *
|
|
flowtable_lookup(sa_family_t sa, struct mbuf *m)
|
|
{
|
|
|
|
switch (sa) {
|
|
#ifdef INET
|
|
case AF_INET:
|
|
return (flowtable_lookup_ipv4(m));
|
|
#endif
|
|
#ifdef INET6
|
|
case AF_INET6:
|
|
return (flowtable_lookup_ipv6(m));
|
|
#endif
|
|
default:
|
|
panic("%s: sa %d", __func__, sa);
|
|
}
|
|
}
|
|
|
|
static struct flentry *
|
|
flowtable_lookup_common(struct flowtable *ft, struct sockaddr_storage *ssa,
|
|
struct sockaddr_storage *dsa, struct mbuf *m, int flags)
|
|
{
|
|
struct route_in6 sro6;
|
|
struct route sro, *ro;
|
|
struct flentry *fle;
|
|
struct rtentry *rt;
|
|
struct llentry *lle;
|
|
struct sockaddr_storage *l3addr;
|
|
struct ifnet *ifp;
|
|
uint32_t key[9], hash, fibnum;
|
|
uint8_t proto;
|
|
|
|
if (V_flowtable_enable == 0)
|
|
return (NULL);
|
|
|
|
sro.ro_rt = sro6.ro_rt = NULL;
|
|
sro.ro_lle = sro6.ro_lle = NULL;
|
|
flags |= ft->ft_flags;
|
|
proto = flags_to_proto(flags);
|
|
fibnum = M_GETFIB(m);
|
|
|
|
switch (ssa->ss_family) {
|
|
#ifdef INET
|
|
case AF_INET: {
|
|
struct sockaddr_in *ssin, *dsin;
|
|
|
|
KASSERT(dsa->ss_family == AF_INET,
|
|
("%s: dsa family %d\n", __func__, dsa->ss_family));
|
|
|
|
ro = &sro;
|
|
memcpy(&ro->ro_dst, dsa, sizeof(struct sockaddr_in));
|
|
/*
|
|
* The harvested source and destination addresses
|
|
* may contain port information if the packet is
|
|
* from a transport protocol (e.g. TCP/UDP). The
|
|
* port field must be cleared before performing
|
|
* a route lookup.
|
|
*/
|
|
((struct sockaddr_in *)&ro->ro_dst)->sin_port = 0;
|
|
dsin = (struct sockaddr_in *)dsa;
|
|
ssin = (struct sockaddr_in *)ssa;
|
|
if ((dsin->sin_addr.s_addr == ssin->sin_addr.s_addr) ||
|
|
(ntohl(dsin->sin_addr.s_addr) >> IN_CLASSA_NSHIFT) == IN_LOOPBACKNET ||
|
|
(ntohl(ssin->sin_addr.s_addr) >> IN_CLASSA_NSHIFT) == IN_LOOPBACKNET)
|
|
return (NULL);
|
|
|
|
hash = ipv4_flow_lookup_hash(ssin, dsin, key, flags);
|
|
break;
|
|
}
|
|
#endif
|
|
#ifdef INET6
|
|
case AF_INET6: {
|
|
struct sockaddr_in6 *ssin6, *dsin6;
|
|
|
|
KASSERT(dsa->ss_family == AF_INET6,
|
|
("%s: dsa family %d\n", __func__, dsa->ss_family));
|
|
|
|
ro = (struct route *)&sro6;
|
|
memcpy(&sro6.ro_dst, dsa,
|
|
sizeof(struct sockaddr_in6));
|
|
((struct sockaddr_in6 *)&ro->ro_dst)->sin6_port = 0;
|
|
dsin6 = (struct sockaddr_in6 *)dsa;
|
|
ssin6 = (struct sockaddr_in6 *)ssa;
|
|
|
|
flags |= FL_IPV6;
|
|
hash = ipv6_flow_lookup_hash(ssin6, dsin6, key, flags);
|
|
break;
|
|
}
|
|
#endif
|
|
default:
|
|
panic("%s: ssa family %d", __func__, ssa->ss_family);
|
|
}
|
|
|
|
/*
|
|
* Ports are zero and this isn't a transmit cache
|
|
* - thus not a protocol for which we need to keep
|
|
* state
|
|
* FL_HASH_ALL => key[0] != 0 for TCP || UDP || SCTP
|
|
*/
|
|
if (key[0] == 0 && (ft->ft_flags & FL_HASH_ALL))
|
|
return (NULL);
|
|
|
|
FLOWSTAT_INC(ft, ft_lookups);
|
|
FL_ENTRY_LOCK(ft, hash);
|
|
if ((fle = FL_ENTRY(ft, hash)) == NULL) {
|
|
FL_ENTRY_UNLOCK(ft, hash);
|
|
goto uncached;
|
|
}
|
|
keycheck:
|
|
rt = __DEVOLATILE(struct rtentry *, fle->f_rt);
|
|
lle = __DEVOLATILE(struct llentry *, fle->f_lle);
|
|
if ((rt != NULL)
|
|
&& lle != NULL
|
|
&& fle->f_fhash == hash
|
|
&& flowtable_key_equal(fle, key)
|
|
&& (proto == fle->f_proto)
|
|
&& (fibnum == fle->f_fibnum)
|
|
&& (rt->rt_flags & RTF_UP)
|
|
&& (rt->rt_ifp != NULL)
|
|
&& (lle->la_flags & LLE_VALID)) {
|
|
FLOWSTAT_INC(ft, ft_hits);
|
|
fle->f_uptime = time_uptime;
|
|
fle->f_flags |= flags;
|
|
FL_ENTRY_UNLOCK(ft, hash);
|
|
goto success;
|
|
} else if (fle->f_next != NULL) {
|
|
fle = fle->f_next;
|
|
goto keycheck;
|
|
}
|
|
FL_ENTRY_UNLOCK(ft, hash);
|
|
uncached:
|
|
if (flags & FL_NOAUTO || flow_full(ft))
|
|
return (NULL);
|
|
|
|
FLOWSTAT_INC(ft, ft_misses);
|
|
/*
|
|
* This bit of code ends up locking the
|
|
* same route 3 times (just like ip_output + ether_output)
|
|
* - at lookup
|
|
* - in rt_check when called by arpresolve
|
|
* - dropping the refcount for the rtentry
|
|
*
|
|
* This could be consolidated to one if we wrote a variant
|
|
* of arpresolve with an rt_check variant that expected to
|
|
* receive the route locked
|
|
*/
|
|
|
|
ft->ft_rtalloc(ro, hash, fibnum);
|
|
if (ro->ro_rt == NULL)
|
|
return (NULL);
|
|
|
|
rt = ro->ro_rt;
|
|
ifp = rt->rt_ifp;
|
|
|
|
if (ifp->if_flags & (IFF_POINTOPOINT | IFF_LOOPBACK)) {
|
|
RTFREE(rt);
|
|
return (NULL);
|
|
}
|
|
|
|
switch (ssa->ss_family) {
|
|
#ifdef INET
|
|
case AF_INET:
|
|
if (rt->rt_flags & RTF_GATEWAY)
|
|
l3addr = (struct sockaddr_storage *)rt->rt_gateway;
|
|
else
|
|
l3addr = (struct sockaddr_storage *)&ro->ro_dst;
|
|
lle = llentry_alloc(ifp, LLTABLE(ifp), l3addr);
|
|
break;
|
|
#endif
|
|
#ifdef INET6
|
|
case AF_INET6: {
|
|
struct sockaddr_in6 *dsin6;
|
|
|
|
dsin6 = (struct sockaddr_in6 *)dsa;
|
|
if (in6_localaddr(&dsin6->sin6_addr)) {
|
|
RTFREE(rt);
|
|
return (NULL);
|
|
}
|
|
|
|
if (rt->rt_flags & RTF_GATEWAY)
|
|
l3addr = (struct sockaddr_storage *)rt->rt_gateway;
|
|
else
|
|
l3addr = (struct sockaddr_storage *)&ro->ro_dst;
|
|
lle = llentry_alloc(ifp, LLTABLE6(ifp), l3addr);
|
|
break;
|
|
}
|
|
#endif
|
|
}
|
|
|
|
if (lle == NULL) {
|
|
RTFREE(rt);
|
|
return (NULL);
|
|
}
|
|
ro->ro_lle = lle;
|
|
|
|
if (flowtable_insert(ft, hash, key, fibnum, ro, flags) != 0) {
|
|
RTFREE(rt);
|
|
LLE_FREE(lle);
|
|
return (NULL);
|
|
}
|
|
|
|
success:
|
|
if (fle != NULL && (m->m_flags & M_FLOWID) == 0) {
|
|
m->m_flags |= M_FLOWID;
|
|
m->m_pkthdr.flowid = fle->f_fhash;
|
|
}
|
|
return (fle);
|
|
}
|
|
|
|
/*
|
|
* used by the bit_alloc macro
|
|
*/
|
|
#define calloc(count, size) malloc((count)*(size), M_DEVBUF, M_WAITOK|M_ZERO)
|
|
|
|
static void
|
|
flowtable_alloc(struct flowtable *ft)
|
|
{
|
|
|
|
#ifdef RADIX_MPATH
|
|
ft->ft_rtalloc = rtalloc_mpath_fib;
|
|
#else
|
|
ft->ft_rtalloc = rtalloc_ign_wrapper;
|
|
#endif
|
|
if (ft->ft_flags & FL_PCPU) {
|
|
ft->ft_lock = flowtable_pcpu_lock;
|
|
ft->ft_unlock = flowtable_pcpu_unlock;
|
|
|
|
for (int i = 0; i <= mp_maxid; i++) {
|
|
ft->ft_table.pcpu[i] =
|
|
malloc(ft->ft_size * sizeof(struct flentry *),
|
|
M_RTABLE, M_WAITOK | M_ZERO);
|
|
ft->ft_masks[i] = bit_alloc(ft->ft_size);
|
|
}
|
|
} else {
|
|
ft->ft_lock_count = 2*(powerof2(mp_maxid + 1) ? (mp_maxid + 1):
|
|
(fls(mp_maxid + 1) << 1));
|
|
|
|
ft->ft_lock = flowtable_global_lock;
|
|
ft->ft_unlock = flowtable_global_unlock;
|
|
ft->ft_table.global =
|
|
malloc(ft->ft_size * sizeof(struct flentry *),
|
|
M_RTABLE, M_WAITOK | M_ZERO);
|
|
ft->ft_locks = malloc(ft->ft_lock_count*sizeof(struct mtx),
|
|
M_RTABLE, M_WAITOK | M_ZERO);
|
|
for (int i = 0; i < ft->ft_lock_count; i++)
|
|
mtx_init(&ft->ft_locks[i], "flow", NULL,
|
|
MTX_DEF | MTX_DUPOK);
|
|
|
|
ft->ft_masks[0] = bit_alloc(ft->ft_size);
|
|
}
|
|
ft->ft_tmpmask = bit_alloc(ft->ft_size);
|
|
|
|
/*
|
|
* In the local transmit case the table truly is
|
|
* just a cache - so everything is eligible for
|
|
* replacement after 5s of non-use
|
|
*/
|
|
if (ft->ft_flags & FL_HASH_ALL) {
|
|
ft->ft_udp_idle = V_flowtable_udp_expire;
|
|
ft->ft_syn_idle = V_flowtable_syn_expire;
|
|
ft->ft_fin_wait_idle = V_flowtable_fin_wait_expire;
|
|
ft->ft_tcp_idle = V_flowtable_fin_wait_expire;
|
|
} else {
|
|
ft->ft_udp_idle = ft->ft_fin_wait_idle =
|
|
ft->ft_syn_idle = ft->ft_tcp_idle = 30;
|
|
|
|
}
|
|
}
|
|
|
|
/*
|
|
* The rest of the code is devoted to garbage collection of expired entries.
|
|
* It is a new additon made necessary by the switch to dynamically allocating
|
|
* flow tables.
|
|
*
|
|
*/
|
|
static void
|
|
fle_free(struct flentry *fle, struct flowtable *ft)
|
|
{
|
|
struct rtentry *rt;
|
|
struct llentry *lle;
|
|
|
|
rt = __DEVOLATILE(struct rtentry *, fle->f_rt);
|
|
lle = __DEVOLATILE(struct llentry *, fle->f_lle);
|
|
if (rt != NULL)
|
|
RTFREE(rt);
|
|
if (lle != NULL)
|
|
LLE_FREE(lle);
|
|
uma_zfree(ft->ft_zone, fle);
|
|
}
|
|
|
|
static void
|
|
flowtable_free_stale(struct flowtable *ft, struct rtentry *rt)
|
|
{
|
|
int curbit = 0, tmpsize;
|
|
struct flentry *fle, **flehead, *fleprev;
|
|
struct flentry *flefreehead, *flefreetail, *fletmp;
|
|
bitstr_t *mask, *tmpmask;
|
|
|
|
flefreehead = flefreetail = NULL;
|
|
mask = flowtable_mask(ft);
|
|
tmpmask = ft->ft_tmpmask;
|
|
tmpsize = ft->ft_size;
|
|
memcpy(tmpmask, mask, ft->ft_size/8);
|
|
/*
|
|
* XXX Note to self, bit_ffs operates at the byte level
|
|
* and thus adds gratuitous overhead
|
|
*/
|
|
bit_ffs(tmpmask, ft->ft_size, &curbit);
|
|
while (curbit != -1) {
|
|
if (curbit >= ft->ft_size || curbit < -1) {
|
|
log(LOG_ALERT,
|
|
"warning: bad curbit value %d \n",
|
|
curbit);
|
|
break;
|
|
}
|
|
|
|
FL_ENTRY_LOCK(ft, curbit);
|
|
flehead = flowtable_entry(ft, curbit);
|
|
fle = fleprev = *flehead;
|
|
|
|
FLOWSTAT_INC(ft, ft_free_checks);
|
|
#ifdef DIAGNOSTIC
|
|
if (fle == NULL && curbit > 0) {
|
|
log(LOG_ALERT,
|
|
"warning bit=%d set, but no fle found\n",
|
|
curbit);
|
|
}
|
|
#endif
|
|
while (fle != NULL) {
|
|
if (rt != NULL) {
|
|
if (__DEVOLATILE(struct rtentry *, fle->f_rt) != rt) {
|
|
fleprev = fle;
|
|
fle = fle->f_next;
|
|
continue;
|
|
}
|
|
} else if (!flow_stale(ft, fle)) {
|
|
fleprev = fle;
|
|
fle = fle->f_next;
|
|
continue;
|
|
}
|
|
/*
|
|
* delete head of the list
|
|
*/
|
|
if (fleprev == *flehead) {
|
|
fletmp = fleprev;
|
|
if (fle == fleprev) {
|
|
fleprev = *flehead = fle->f_next;
|
|
} else
|
|
fleprev = *flehead = fle;
|
|
fle = fle->f_next;
|
|
} else {
|
|
/*
|
|
* don't advance fleprev
|
|
*/
|
|
fletmp = fle;
|
|
fleprev->f_next = fle->f_next;
|
|
fle = fleprev->f_next;
|
|
}
|
|
|
|
if (flefreehead == NULL)
|
|
flefreehead = flefreetail = fletmp;
|
|
else {
|
|
flefreetail->f_next = fletmp;
|
|
flefreetail = fletmp;
|
|
}
|
|
fletmp->f_next = NULL;
|
|
}
|
|
if (*flehead == NULL)
|
|
bit_clear(mask, curbit);
|
|
FL_ENTRY_UNLOCK(ft, curbit);
|
|
bit_clear(tmpmask, curbit);
|
|
tmpmask += (curbit / 8);
|
|
tmpsize -= (curbit / 8) * 8;
|
|
bit_ffs(tmpmask, tmpsize, &curbit);
|
|
}
|
|
while ((fle = flefreehead) != NULL) {
|
|
flefreehead = fle->f_next;
|
|
FLOWSTAT_INC(ft, ft_frees);
|
|
fle_free(fle, ft);
|
|
}
|
|
}
|
|
|
|
void
|
|
flowtable_route_flush(sa_family_t sa, struct rtentry *rt)
|
|
{
|
|
struct flowtable *ft;
|
|
int i;
|
|
|
|
switch (sa) {
|
|
#ifdef INET
|
|
case AF_INET:
|
|
ft = &V_ip4_ft;
|
|
break;
|
|
#endif
|
|
#ifdef INET6
|
|
case AF_INET6:
|
|
ft = &V_ip6_ft;
|
|
break;
|
|
#endif
|
|
default:
|
|
panic("%s: sa %d", __func__, sa);
|
|
}
|
|
|
|
if (ft->ft_flags & FL_PCPU) {
|
|
CPU_FOREACH(i) {
|
|
if (smp_started == 1) {
|
|
thread_lock(curthread);
|
|
sched_bind(curthread, i);
|
|
thread_unlock(curthread);
|
|
}
|
|
|
|
flowtable_free_stale(ft, rt);
|
|
|
|
if (smp_started == 1) {
|
|
thread_lock(curthread);
|
|
sched_unbind(curthread);
|
|
thread_unlock(curthread);
|
|
}
|
|
}
|
|
} else {
|
|
flowtable_free_stale(ft, rt);
|
|
}
|
|
}
|
|
|
|
static void
|
|
flowtable_clean_vnet(struct flowtable *ft)
|
|
{
|
|
|
|
if (ft->ft_flags & FL_PCPU) {
|
|
int i;
|
|
|
|
CPU_FOREACH(i) {
|
|
if (smp_started == 1) {
|
|
thread_lock(curthread);
|
|
sched_bind(curthread, i);
|
|
thread_unlock(curthread);
|
|
}
|
|
|
|
flowtable_free_stale(ft, NULL);
|
|
|
|
if (smp_started == 1) {
|
|
thread_lock(curthread);
|
|
sched_unbind(curthread);
|
|
thread_unlock(curthread);
|
|
}
|
|
}
|
|
} else
|
|
flowtable_free_stale(ft, NULL);
|
|
}
|
|
|
|
static void
|
|
flowtable_cleaner(void)
|
|
{
|
|
VNET_ITERATOR_DECL(vnet_iter);
|
|
struct thread *td;
|
|
|
|
if (bootverbose)
|
|
log(LOG_INFO, "flowtable cleaner started\n");
|
|
td = curthread;
|
|
while (1) {
|
|
VNET_LIST_RLOCK();
|
|
VNET_FOREACH(vnet_iter) {
|
|
CURVNET_SET(vnet_iter);
|
|
#ifdef INET
|
|
flowtable_clean_vnet(&V_ip4_ft);
|
|
#endif
|
|
#ifdef INET6
|
|
flowtable_clean_vnet(&V_ip6_ft);
|
|
#endif
|
|
CURVNET_RESTORE();
|
|
}
|
|
VNET_LIST_RUNLOCK();
|
|
|
|
/*
|
|
* The 10 second interval between cleaning checks
|
|
* is arbitrary
|
|
*/
|
|
mtx_lock(&flowclean_lock);
|
|
thread_lock(td);
|
|
sched_prio(td, PPAUSE);
|
|
thread_unlock(td);
|
|
flowclean_cycles++;
|
|
cv_broadcast(&flowclean_f_cv);
|
|
cv_timedwait(&flowclean_c_cv, &flowclean_lock, flowclean_freq);
|
|
mtx_unlock(&flowclean_lock);
|
|
}
|
|
}
|
|
|
|
static void
|
|
flowtable_flush(void *unused __unused)
|
|
{
|
|
uint64_t start;
|
|
|
|
mtx_lock(&flowclean_lock);
|
|
start = flowclean_cycles;
|
|
while (start == flowclean_cycles) {
|
|
cv_broadcast(&flowclean_c_cv);
|
|
cv_wait(&flowclean_f_cv, &flowclean_lock);
|
|
}
|
|
mtx_unlock(&flowclean_lock);
|
|
}
|
|
|
|
static struct kproc_desc flow_kp = {
|
|
"flowcleaner",
|
|
flowtable_cleaner,
|
|
&flowcleanerproc
|
|
};
|
|
SYSINIT(flowcleaner, SI_SUB_KTHREAD_IDLE, SI_ORDER_ANY, kproc_start, &flow_kp);
|
|
|
|
static int
|
|
flowtable_get_size(char *name)
|
|
{
|
|
int size;
|
|
|
|
if (TUNABLE_INT_FETCH(name, &size)) {
|
|
if (size < 256)
|
|
size = 256;
|
|
if (!powerof2(size)) {
|
|
printf("%s must be power of 2\n", name);
|
|
size = 2048;
|
|
}
|
|
} else {
|
|
/*
|
|
* round up to the next power of 2
|
|
*/
|
|
size = 1 << fls((1024 + maxusers * 64) - 1);
|
|
}
|
|
|
|
return (size);
|
|
}
|
|
|
|
static void
|
|
flowtable_init(const void *unused __unused)
|
|
{
|
|
|
|
flow_hashjitter = arc4random();
|
|
|
|
#ifdef INET
|
|
flow_ipv4_zone = uma_zcreate("ip4flow", sizeof(struct flentry_v4),
|
|
NULL, NULL, NULL, NULL, UMA_ALIGN_CACHE, UMA_ZONE_MAXBUCKET);
|
|
uma_zone_set_max(flow_ipv4_zone, 1024 + maxusers * 64 * mp_ncpus);
|
|
#endif
|
|
#ifdef INET6
|
|
flow_ipv6_zone = uma_zcreate("ip6flow", sizeof(struct flentry_v6),
|
|
NULL, NULL, NULL, NULL, UMA_ALIGN_CACHE, UMA_ZONE_MAXBUCKET);
|
|
uma_zone_set_max(flow_ipv6_zone, 1024 + maxusers * 64 * mp_ncpus);
|
|
#endif
|
|
|
|
cv_init(&flowclean_c_cv, "c_flowcleanwait");
|
|
cv_init(&flowclean_f_cv, "f_flowcleanwait");
|
|
mtx_init(&flowclean_lock, "flowclean lock", NULL, MTX_DEF);
|
|
EVENTHANDLER_REGISTER(ifnet_departure_event, flowtable_flush, NULL,
|
|
EVENTHANDLER_PRI_ANY);
|
|
flowclean_freq = 20*hz;
|
|
}
|
|
SYSINIT(flowtable_init, SI_SUB_PROTO_BEGIN, SI_ORDER_FIRST,
|
|
flowtable_init, NULL);
|
|
|
|
#ifdef INET
|
|
static SYSCTL_NODE(_net_flowtable, OID_AUTO, ip4, CTLFLAG_RD, NULL,
|
|
"Flowtable for IPv4");
|
|
SYSCTL_UMA_MAX(_net_flowtable_ip4, OID_AUTO, maxflows, CTLFLAG_RW,
|
|
&flow_ipv4_zone, "Maximum number of IPv4 flows allowed");
|
|
|
|
static VNET_PCPUSTAT_DEFINE(struct flowtable_stat, ip4_ftstat);
|
|
VNET_PCPUSTAT_SYSINIT(ip4_ftstat);
|
|
VNET_PCPUSTAT_SYSUNINIT(ip4_ftstat);
|
|
SYSCTL_VNET_PCPUSTAT(_net_flowtable_ip4, OID_AUTO, stat, struct flowtable_stat,
|
|
ip4_ftstat, "Flowtable statistics for IPv4 "
|
|
"(struct flowtable_stat, net/flowtable.h)");
|
|
|
|
static void
|
|
flowtable_init_vnet_v4(const void *unused __unused)
|
|
{
|
|
|
|
V_ip4_ft.ft_zone = flow_ipv4_zone;
|
|
V_ip4_ft.ft_size = flowtable_get_size("net.flowtable.ip4.size");
|
|
V_ip4_ft.ft_flags = FL_PCPU;
|
|
V_ip4_ft.ft_stat = VNET(ip4_ftstat);
|
|
flowtable_alloc(&V_ip4_ft);
|
|
}
|
|
VNET_SYSINIT(ft_vnet_v4, SI_SUB_PROTO_IFATTACHDOMAIN, SI_ORDER_ANY,
|
|
flowtable_init_vnet_v4, NULL);
|
|
#endif /* INET */
|
|
|
|
#ifdef INET6
|
|
static SYSCTL_NODE(_net_flowtable, OID_AUTO, ip6, CTLFLAG_RD, NULL,
|
|
"Flowtable for IPv6");
|
|
SYSCTL_UMA_MAX(_net_flowtable_ip6, OID_AUTO, maxflows, CTLFLAG_RW,
|
|
&flow_ipv6_zone, "Maximum number of IPv6 flows allowed");
|
|
|
|
static VNET_PCPUSTAT_DEFINE(struct flowtable_stat, ip6_ftstat);
|
|
VNET_PCPUSTAT_SYSINIT(ip6_ftstat);
|
|
VNET_PCPUSTAT_SYSUNINIT(ip6_ftstat);
|
|
SYSCTL_VNET_PCPUSTAT(_net_flowtable_ip6, OID_AUTO, stat, struct flowtable_stat,
|
|
ip6_ftstat, "Flowtable statistics for IPv6 "
|
|
"(struct flowtable_stat, net/flowtable.h)");
|
|
|
|
static void
|
|
flowtable_init_vnet_v6(const void *unused __unused)
|
|
{
|
|
|
|
V_ip6_ft.ft_zone = flow_ipv6_zone;
|
|
V_ip6_ft.ft_size = flowtable_get_size("net.flowtable.ip6.size");
|
|
V_ip6_ft.ft_flags = FL_PCPU;
|
|
V_ip6_ft.ft_stat = VNET(ip6_ftstat);
|
|
flowtable_alloc(&V_ip6_ft);
|
|
}
|
|
VNET_SYSINIT(flowtable_init_vnet_v6, SI_SUB_PROTO_IFATTACHDOMAIN, SI_ORDER_ANY,
|
|
flowtable_init_vnet_v6, NULL);
|
|
#endif /* INET6 */
|
|
|
|
#ifdef DDB
|
|
static uint32_t *
|
|
flowtable_get_hashkey(struct flentry *fle)
|
|
{
|
|
uint32_t *hashkey;
|
|
|
|
if (fle->f_flags & FL_IPV6)
|
|
hashkey = ((struct flentry_v4 *)fle)->fl_flow.ipf_key;
|
|
else
|
|
hashkey = ((struct flentry_v6 *)fle)->fl_flow.ipf_key;
|
|
|
|
return (hashkey);
|
|
}
|
|
|
|
static bitstr_t *
|
|
flowtable_mask_pcpu(struct flowtable *ft, int cpuid)
|
|
{
|
|
bitstr_t *mask;
|
|
|
|
if (ft->ft_flags & FL_PCPU)
|
|
mask = ft->ft_masks[cpuid];
|
|
else
|
|
mask = ft->ft_masks[0];
|
|
|
|
return (mask);
|
|
}
|
|
|
|
static struct flentry **
|
|
flowtable_entry_pcpu(struct flowtable *ft, uint32_t hash, int cpuid)
|
|
{
|
|
struct flentry **fle;
|
|
int index = (hash % ft->ft_size);
|
|
|
|
if (ft->ft_flags & FL_PCPU) {
|
|
fle = &ft->ft_table.pcpu[cpuid][index];
|
|
} else {
|
|
fle = &ft->ft_table.global[index];
|
|
}
|
|
|
|
return (fle);
|
|
}
|
|
|
|
static void
|
|
flow_show(struct flowtable *ft, struct flentry *fle)
|
|
{
|
|
int idle_time;
|
|
int rt_valid, ifp_valid;
|
|
uint16_t sport, dport;
|
|
uint32_t *hashkey;
|
|
char saddr[4*sizeof "123"], daddr[4*sizeof "123"];
|
|
volatile struct rtentry *rt;
|
|
struct ifnet *ifp = NULL;
|
|
|
|
idle_time = (int)(time_uptime - fle->f_uptime);
|
|
rt = fle->f_rt;
|
|
rt_valid = rt != NULL;
|
|
if (rt_valid)
|
|
ifp = rt->rt_ifp;
|
|
ifp_valid = ifp != NULL;
|
|
hashkey = flowtable_get_hashkey(fle);
|
|
if (fle->f_flags & FL_IPV6)
|
|
goto skipaddr;
|
|
|
|
inet_ntoa_r(*(struct in_addr *) &hashkey[2], daddr);
|
|
if (ft->ft_flags & FL_HASH_ALL) {
|
|
inet_ntoa_r(*(struct in_addr *) &hashkey[1], saddr);
|
|
sport = ntohs(((uint16_t *)hashkey)[0]);
|
|
dport = ntohs(((uint16_t *)hashkey)[1]);
|
|
db_printf("%s:%d->%s:%d",
|
|
saddr, sport, daddr,
|
|
dport);
|
|
} else
|
|
db_printf("%s ", daddr);
|
|
|
|
skipaddr:
|
|
if (fle->f_flags & FL_STALE)
|
|
db_printf(" FL_STALE ");
|
|
if (fle->f_flags & FL_TCP)
|
|
db_printf(" FL_TCP ");
|
|
if (fle->f_flags & FL_UDP)
|
|
db_printf(" FL_UDP ");
|
|
if (rt_valid) {
|
|
if (rt->rt_flags & RTF_UP)
|
|
db_printf(" RTF_UP ");
|
|
}
|
|
if (ifp_valid) {
|
|
if (ifp->if_flags & IFF_LOOPBACK)
|
|
db_printf(" IFF_LOOPBACK ");
|
|
if (ifp->if_flags & IFF_UP)
|
|
db_printf(" IFF_UP ");
|
|
if (ifp->if_flags & IFF_POINTOPOINT)
|
|
db_printf(" IFF_POINTOPOINT ");
|
|
}
|
|
if (fle->f_flags & FL_IPV6)
|
|
db_printf("\n\tkey=%08x:%08x:%08x%08x:%08x:%08x%08x:%08x:%08x",
|
|
hashkey[0], hashkey[1], hashkey[2],
|
|
hashkey[3], hashkey[4], hashkey[5],
|
|
hashkey[6], hashkey[7], hashkey[8]);
|
|
else
|
|
db_printf("\n\tkey=%08x:%08x:%08x ",
|
|
hashkey[0], hashkey[1], hashkey[2]);
|
|
db_printf("hash=%08x idle_time=%03d"
|
|
"\n\tfibnum=%02d rt=%p",
|
|
fle->f_fhash, idle_time, fle->f_fibnum, fle->f_rt);
|
|
db_printf("\n");
|
|
}
|
|
|
|
static void
|
|
flowtable_show(struct flowtable *ft, int cpuid)
|
|
{
|
|
int curbit = 0;
|
|
struct flentry *fle, **flehead;
|
|
bitstr_t *mask, *tmpmask;
|
|
|
|
if (cpuid != -1)
|
|
db_printf("cpu: %d\n", cpuid);
|
|
mask = flowtable_mask_pcpu(ft, cpuid);
|
|
tmpmask = ft->ft_tmpmask;
|
|
memcpy(tmpmask, mask, ft->ft_size/8);
|
|
/*
|
|
* XXX Note to self, bit_ffs operates at the byte level
|
|
* and thus adds gratuitous overhead
|
|
*/
|
|
bit_ffs(tmpmask, ft->ft_size, &curbit);
|
|
while (curbit != -1) {
|
|
if (curbit >= ft->ft_size || curbit < -1) {
|
|
db_printf("warning: bad curbit value %d \n",
|
|
curbit);
|
|
break;
|
|
}
|
|
|
|
flehead = flowtable_entry_pcpu(ft, curbit, cpuid);
|
|
fle = *flehead;
|
|
|
|
while (fle != NULL) {
|
|
flow_show(ft, fle);
|
|
fle = fle->f_next;
|
|
continue;
|
|
}
|
|
bit_clear(tmpmask, curbit);
|
|
bit_ffs(tmpmask, ft->ft_size, &curbit);
|
|
}
|
|
}
|
|
|
|
static void
|
|
flowtable_show_vnet(struct flowtable *ft)
|
|
{
|
|
|
|
if (ft->ft_flags & FL_PCPU) {
|
|
int i;
|
|
|
|
CPU_FOREACH(i) {
|
|
flowtable_show(ft, i);
|
|
}
|
|
} else
|
|
flowtable_show(ft, -1);
|
|
}
|
|
|
|
DB_SHOW_COMMAND(flowtables, db_show_flowtables)
|
|
{
|
|
VNET_ITERATOR_DECL(vnet_iter);
|
|
|
|
VNET_FOREACH(vnet_iter) {
|
|
CURVNET_SET(vnet_iter);
|
|
#ifdef VIMAGE
|
|
db_printf("vnet %p\n", vnet_iter);
|
|
#endif
|
|
#ifdef INET
|
|
printf("IPv4:\n");
|
|
flowtable_show_vnet(&V_ip4_ft);
|
|
#endif
|
|
#ifdef INET6
|
|
printf("IPv6:\n");
|
|
flowtable_show_vnet(&V_ip6_ft);
|
|
#endif
|
|
CURVNET_RESTORE();
|
|
}
|
|
}
|
|
#endif
|