674d86bf91
context always. Convert nodes to consistently use M_WAITOK flag for memory allocation. Reviewed by: julian
995 lines
27 KiB
C
995 lines
27 KiB
C
/*-
|
|
* Copyright (c) 2004-2010 University of Zagreb
|
|
* Copyright (c) 2007-2008 FreeBSD Foundation
|
|
*
|
|
* This software was developed by the University of Zagreb and the
|
|
* FreeBSD Foundation under sponsorship by the Stichting NLnet and the
|
|
* FreeBSD Foundation.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* $FreeBSD$
|
|
*/
|
|
|
|
/*
|
|
* This node permits simple traffic shaping by emulating bandwidth
|
|
* and delay, as well as random packet losses.
|
|
* The node has two hooks, upper and lower. Traffic flowing from upper to
|
|
* lower hook is referenced as downstream, and vice versa. Parameters for
|
|
* both directions can be set separately, except for delay.
|
|
*/
|
|
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/errno.h>
|
|
#include <sys/systm.h>
|
|
#include <sys/kernel.h>
|
|
#include <sys/malloc.h>
|
|
#include <sys/mbuf.h>
|
|
#include <sys/time.h>
|
|
|
|
#include <vm/uma.h>
|
|
|
|
#include <net/vnet.h>
|
|
|
|
#include <netinet/in.h>
|
|
#include <netinet/in_systm.h>
|
|
#include <netinet/ip.h>
|
|
|
|
#include <netgraph/ng_message.h>
|
|
#include <netgraph/netgraph.h>
|
|
#include <netgraph/ng_parse.h>
|
|
#include <netgraph/ng_pipe.h>
|
|
|
|
static MALLOC_DEFINE(M_NG_PIPE, "ng_pipe", "ng_pipe");
|
|
|
|
/* Packet header struct */
|
|
struct ngp_hdr {
|
|
TAILQ_ENTRY(ngp_hdr) ngp_link; /* next pkt in queue */
|
|
struct timeval when; /* this packet's due time */
|
|
struct mbuf *m; /* ptr to the packet data */
|
|
};
|
|
TAILQ_HEAD(p_head, ngp_hdr);
|
|
|
|
/* FIFO queue struct */
|
|
struct ngp_fifo {
|
|
TAILQ_ENTRY(ngp_fifo) fifo_le; /* list of active queues only */
|
|
struct p_head packet_head; /* FIFO queue head */
|
|
u_int32_t hash; /* flow signature */
|
|
struct timeval vtime; /* virtual time, for WFQ */
|
|
u_int32_t rr_deficit; /* for DRR */
|
|
u_int32_t packets; /* # of packets in this queue */
|
|
};
|
|
|
|
/* Per hook info */
|
|
struct hookinfo {
|
|
hook_p hook;
|
|
int noqueue; /* bypass any processing */
|
|
TAILQ_HEAD(, ngp_fifo) fifo_head; /* FIFO queues */
|
|
TAILQ_HEAD(, ngp_hdr) qout_head; /* delay queue head */
|
|
struct timeval qin_utime;
|
|
struct ng_pipe_hookcfg cfg;
|
|
struct ng_pipe_hookrun run;
|
|
struct ng_pipe_hookstat stats;
|
|
uint64_t *ber_p; /* loss_p(BER,psize) map */
|
|
};
|
|
|
|
/* Per node info */
|
|
struct node_priv {
|
|
u_int64_t delay;
|
|
u_int32_t overhead;
|
|
u_int32_t header_offset;
|
|
struct hookinfo lower;
|
|
struct hookinfo upper;
|
|
struct callout timer;
|
|
int timer_scheduled;
|
|
};
|
|
typedef struct node_priv *priv_p;
|
|
|
|
/* Macro for calculating the virtual time for packet dequeueing in WFQ */
|
|
#define FIFO_VTIME_SORT(plen) \
|
|
if (hinfo->cfg.wfq && hinfo->cfg.bandwidth) { \
|
|
ngp_f->vtime.tv_usec = now->tv_usec + ((uint64_t) (plen) \
|
|
+ priv->overhead ) * hinfo->run.fifo_queues * \
|
|
8000000 / hinfo->cfg.bandwidth; \
|
|
ngp_f->vtime.tv_sec = now->tv_sec + \
|
|
ngp_f->vtime.tv_usec / 1000000; \
|
|
ngp_f->vtime.tv_usec = ngp_f->vtime.tv_usec % 1000000; \
|
|
TAILQ_FOREACH(ngp_f1, &hinfo->fifo_head, fifo_le) \
|
|
if (ngp_f1->vtime.tv_sec > ngp_f->vtime.tv_sec || \
|
|
(ngp_f1->vtime.tv_sec == ngp_f->vtime.tv_sec && \
|
|
ngp_f1->vtime.tv_usec > ngp_f->vtime.tv_usec)) \
|
|
break; \
|
|
if (ngp_f1 == NULL) \
|
|
TAILQ_INSERT_TAIL(&hinfo->fifo_head, ngp_f, fifo_le); \
|
|
else \
|
|
TAILQ_INSERT_BEFORE(ngp_f1, ngp_f, fifo_le); \
|
|
} else \
|
|
TAILQ_INSERT_TAIL(&hinfo->fifo_head, ngp_f, fifo_le); \
|
|
|
|
|
|
static void parse_cfg(struct ng_pipe_hookcfg *, struct ng_pipe_hookcfg *,
|
|
struct hookinfo *, priv_p);
|
|
static void pipe_dequeue(struct hookinfo *, struct timeval *);
|
|
static void ngp_callout(node_p, hook_p, void *, int);
|
|
static int ngp_modevent(module_t, int, void *);
|
|
|
|
/* zone for storing ngp_hdr-s */
|
|
static uma_zone_t ngp_zone;
|
|
|
|
/* Netgraph methods */
|
|
static ng_constructor_t ngp_constructor;
|
|
static ng_rcvmsg_t ngp_rcvmsg;
|
|
static ng_shutdown_t ngp_shutdown;
|
|
static ng_newhook_t ngp_newhook;
|
|
static ng_rcvdata_t ngp_rcvdata;
|
|
static ng_disconnect_t ngp_disconnect;
|
|
|
|
/* Parse type for struct ng_pipe_hookstat */
|
|
static const struct ng_parse_struct_field
|
|
ng_pipe_hookstat_type_fields[] = NG_PIPE_HOOKSTAT_INFO;
|
|
static const struct ng_parse_type ng_pipe_hookstat_type = {
|
|
&ng_parse_struct_type,
|
|
&ng_pipe_hookstat_type_fields
|
|
};
|
|
|
|
/* Parse type for struct ng_pipe_stats */
|
|
static const struct ng_parse_struct_field ng_pipe_stats_type_fields[] =
|
|
NG_PIPE_STATS_INFO(&ng_pipe_hookstat_type);
|
|
static const struct ng_parse_type ng_pipe_stats_type = {
|
|
&ng_parse_struct_type,
|
|
&ng_pipe_stats_type_fields
|
|
};
|
|
|
|
/* Parse type for struct ng_pipe_hookrun */
|
|
static const struct ng_parse_struct_field
|
|
ng_pipe_hookrun_type_fields[] = NG_PIPE_HOOKRUN_INFO;
|
|
static const struct ng_parse_type ng_pipe_hookrun_type = {
|
|
&ng_parse_struct_type,
|
|
&ng_pipe_hookrun_type_fields
|
|
};
|
|
|
|
/* Parse type for struct ng_pipe_run */
|
|
static const struct ng_parse_struct_field
|
|
ng_pipe_run_type_fields[] = NG_PIPE_RUN_INFO(&ng_pipe_hookrun_type);
|
|
static const struct ng_parse_type ng_pipe_run_type = {
|
|
&ng_parse_struct_type,
|
|
&ng_pipe_run_type_fields
|
|
};
|
|
|
|
/* Parse type for struct ng_pipe_hookcfg */
|
|
static const struct ng_parse_struct_field
|
|
ng_pipe_hookcfg_type_fields[] = NG_PIPE_HOOKCFG_INFO;
|
|
static const struct ng_parse_type ng_pipe_hookcfg_type = {
|
|
&ng_parse_struct_type,
|
|
&ng_pipe_hookcfg_type_fields
|
|
};
|
|
|
|
/* Parse type for struct ng_pipe_cfg */
|
|
static const struct ng_parse_struct_field
|
|
ng_pipe_cfg_type_fields[] = NG_PIPE_CFG_INFO(&ng_pipe_hookcfg_type);
|
|
static const struct ng_parse_type ng_pipe_cfg_type = {
|
|
&ng_parse_struct_type,
|
|
&ng_pipe_cfg_type_fields
|
|
};
|
|
|
|
/* List of commands and how to convert arguments to/from ASCII */
|
|
static const struct ng_cmdlist ngp_cmds[] = {
|
|
{
|
|
.cookie = NGM_PIPE_COOKIE,
|
|
.cmd = NGM_PIPE_GET_STATS,
|
|
.name = "getstats",
|
|
.respType = &ng_pipe_stats_type
|
|
},
|
|
{
|
|
.cookie = NGM_PIPE_COOKIE,
|
|
.cmd = NGM_PIPE_CLR_STATS,
|
|
.name = "clrstats"
|
|
},
|
|
{
|
|
.cookie = NGM_PIPE_COOKIE,
|
|
.cmd = NGM_PIPE_GETCLR_STATS,
|
|
.name = "getclrstats",
|
|
.respType = &ng_pipe_stats_type
|
|
},
|
|
{
|
|
.cookie = NGM_PIPE_COOKIE,
|
|
.cmd = NGM_PIPE_GET_RUN,
|
|
.name = "getrun",
|
|
.respType = &ng_pipe_run_type
|
|
},
|
|
{
|
|
.cookie = NGM_PIPE_COOKIE,
|
|
.cmd = NGM_PIPE_GET_CFG,
|
|
.name = "getcfg",
|
|
.respType = &ng_pipe_cfg_type
|
|
},
|
|
{
|
|
.cookie = NGM_PIPE_COOKIE,
|
|
.cmd = NGM_PIPE_SET_CFG,
|
|
.name = "setcfg",
|
|
.mesgType = &ng_pipe_cfg_type,
|
|
},
|
|
{ 0 }
|
|
};
|
|
|
|
/* Netgraph type descriptor */
|
|
static struct ng_type ng_pipe_typestruct = {
|
|
.version = NG_ABI_VERSION,
|
|
.name = NG_PIPE_NODE_TYPE,
|
|
.mod_event = ngp_modevent,
|
|
.constructor = ngp_constructor,
|
|
.shutdown = ngp_shutdown,
|
|
.rcvmsg = ngp_rcvmsg,
|
|
.newhook = ngp_newhook,
|
|
.rcvdata = ngp_rcvdata,
|
|
.disconnect = ngp_disconnect,
|
|
.cmdlist = ngp_cmds
|
|
};
|
|
NETGRAPH_INIT(pipe, &ng_pipe_typestruct);
|
|
|
|
/* Node constructor */
|
|
static int
|
|
ngp_constructor(node_p node)
|
|
{
|
|
priv_p priv;
|
|
|
|
priv = malloc(sizeof(*priv), M_NG_PIPE, M_ZERO | M_WAITOK);
|
|
NG_NODE_SET_PRIVATE(node, priv);
|
|
|
|
/* Mark node as single-threaded */
|
|
NG_NODE_FORCE_WRITER(node);
|
|
|
|
ng_callout_init(&priv->timer);
|
|
|
|
return (0);
|
|
}
|
|
|
|
/* Add a hook */
|
|
static int
|
|
ngp_newhook(node_p node, hook_p hook, const char *name)
|
|
{
|
|
const priv_p priv = NG_NODE_PRIVATE(node);
|
|
struct hookinfo *hinfo;
|
|
|
|
if (strcmp(name, NG_PIPE_HOOK_UPPER) == 0) {
|
|
bzero(&priv->upper, sizeof(priv->upper));
|
|
priv->upper.hook = hook;
|
|
NG_HOOK_SET_PRIVATE(hook, &priv->upper);
|
|
} else if (strcmp(name, NG_PIPE_HOOK_LOWER) == 0) {
|
|
bzero(&priv->lower, sizeof(priv->lower));
|
|
priv->lower.hook = hook;
|
|
NG_HOOK_SET_PRIVATE(hook, &priv->lower);
|
|
} else
|
|
return (EINVAL);
|
|
|
|
/* Load non-zero initial cfg values */
|
|
hinfo = NG_HOOK_PRIVATE(hook);
|
|
hinfo->cfg.qin_size_limit = 50;
|
|
hinfo->cfg.fifo = 1;
|
|
hinfo->cfg.droptail = 1;
|
|
TAILQ_INIT(&hinfo->fifo_head);
|
|
TAILQ_INIT(&hinfo->qout_head);
|
|
return (0);
|
|
}
|
|
|
|
/* Receive a control message */
|
|
static int
|
|
ngp_rcvmsg(node_p node, item_p item, hook_p lasthook)
|
|
{
|
|
const priv_p priv = NG_NODE_PRIVATE(node);
|
|
struct ng_mesg *resp = NULL;
|
|
struct ng_mesg *msg;
|
|
struct ng_pipe_stats *stats;
|
|
struct ng_pipe_run *run;
|
|
struct ng_pipe_cfg *cfg;
|
|
int error = 0;
|
|
|
|
NGI_GET_MSG(item, msg);
|
|
switch (msg->header.typecookie) {
|
|
case NGM_PIPE_COOKIE:
|
|
switch (msg->header.cmd) {
|
|
case NGM_PIPE_GET_STATS:
|
|
case NGM_PIPE_CLR_STATS:
|
|
case NGM_PIPE_GETCLR_STATS:
|
|
if (msg->header.cmd != NGM_PIPE_CLR_STATS) {
|
|
NG_MKRESPONSE(resp, msg,
|
|
sizeof(*stats), M_NOWAIT);
|
|
if (resp == NULL) {
|
|
error = ENOMEM;
|
|
break;
|
|
}
|
|
stats = (struct ng_pipe_stats *) resp->data;
|
|
bcopy(&priv->upper.stats, &stats->downstream,
|
|
sizeof(stats->downstream));
|
|
bcopy(&priv->lower.stats, &stats->upstream,
|
|
sizeof(stats->upstream));
|
|
}
|
|
if (msg->header.cmd != NGM_PIPE_GET_STATS) {
|
|
bzero(&priv->upper.stats,
|
|
sizeof(priv->upper.stats));
|
|
bzero(&priv->lower.stats,
|
|
sizeof(priv->lower.stats));
|
|
}
|
|
break;
|
|
case NGM_PIPE_GET_RUN:
|
|
NG_MKRESPONSE(resp, msg, sizeof(*run), M_NOWAIT);
|
|
if (resp == NULL) {
|
|
error = ENOMEM;
|
|
break;
|
|
}
|
|
run = (struct ng_pipe_run *) resp->data;
|
|
bcopy(&priv->upper.run, &run->downstream,
|
|
sizeof(run->downstream));
|
|
bcopy(&priv->lower.run, &run->upstream,
|
|
sizeof(run->upstream));
|
|
break;
|
|
case NGM_PIPE_GET_CFG:
|
|
NG_MKRESPONSE(resp, msg, sizeof(*cfg), M_NOWAIT);
|
|
if (resp == NULL) {
|
|
error = ENOMEM;
|
|
break;
|
|
}
|
|
cfg = (struct ng_pipe_cfg *) resp->data;
|
|
bcopy(&priv->upper.cfg, &cfg->downstream,
|
|
sizeof(cfg->downstream));
|
|
bcopy(&priv->lower.cfg, &cfg->upstream,
|
|
sizeof(cfg->upstream));
|
|
cfg->delay = priv->delay;
|
|
cfg->overhead = priv->overhead;
|
|
cfg->header_offset = priv->header_offset;
|
|
if (cfg->upstream.bandwidth ==
|
|
cfg->downstream.bandwidth) {
|
|
cfg->bandwidth = cfg->upstream.bandwidth;
|
|
cfg->upstream.bandwidth = 0;
|
|
cfg->downstream.bandwidth = 0;
|
|
} else
|
|
cfg->bandwidth = 0;
|
|
break;
|
|
case NGM_PIPE_SET_CFG:
|
|
cfg = (struct ng_pipe_cfg *) msg->data;
|
|
if (msg->header.arglen != sizeof(*cfg)) {
|
|
error = EINVAL;
|
|
break;
|
|
}
|
|
|
|
if (cfg->delay == -1)
|
|
priv->delay = 0;
|
|
else if (cfg->delay > 0 && cfg->delay < 10000000)
|
|
priv->delay = cfg->delay;
|
|
|
|
if (cfg->bandwidth == -1) {
|
|
priv->upper.cfg.bandwidth = 0;
|
|
priv->lower.cfg.bandwidth = 0;
|
|
priv->overhead = 0;
|
|
} else if (cfg->bandwidth >= 100 &&
|
|
cfg->bandwidth <= 1000000000) {
|
|
priv->upper.cfg.bandwidth = cfg->bandwidth;
|
|
priv->lower.cfg.bandwidth = cfg->bandwidth;
|
|
if (cfg->bandwidth >= 10000000)
|
|
priv->overhead = 8+4+12; /* Ethernet */
|
|
else
|
|
priv->overhead = 10; /* HDLC */
|
|
}
|
|
|
|
if (cfg->overhead == -1)
|
|
priv->overhead = 0;
|
|
else if (cfg->overhead > 0 &&
|
|
cfg->overhead < MAX_OHSIZE)
|
|
priv->overhead = cfg->overhead;
|
|
|
|
if (cfg->header_offset == -1)
|
|
priv->header_offset = 0;
|
|
else if (cfg->header_offset > 0 &&
|
|
cfg->header_offset < 64)
|
|
priv->header_offset = cfg->header_offset;
|
|
|
|
parse_cfg(&priv->upper.cfg, &cfg->downstream,
|
|
&priv->upper, priv);
|
|
parse_cfg(&priv->lower.cfg, &cfg->upstream,
|
|
&priv->lower, priv);
|
|
break;
|
|
default:
|
|
error = EINVAL;
|
|
break;
|
|
}
|
|
break;
|
|
default:
|
|
error = EINVAL;
|
|
break;
|
|
}
|
|
NG_RESPOND_MSG(error, node, item, resp);
|
|
NG_FREE_MSG(msg);
|
|
|
|
return (error);
|
|
}
|
|
|
|
static void
|
|
parse_cfg(struct ng_pipe_hookcfg *current, struct ng_pipe_hookcfg *new,
|
|
struct hookinfo *hinfo, priv_p priv)
|
|
{
|
|
|
|
if (new->ber == -1) {
|
|
current->ber = 0;
|
|
if (hinfo->ber_p) {
|
|
free(hinfo->ber_p, M_NG_PIPE);
|
|
hinfo->ber_p = NULL;
|
|
}
|
|
} else if (new->ber >= 1 && new->ber <= 1000000000000) {
|
|
static const uint64_t one = 0x1000000000000; /* = 2^48 */
|
|
uint64_t p0, p;
|
|
uint32_t fsize, i;
|
|
|
|
if (hinfo->ber_p == NULL)
|
|
hinfo->ber_p =
|
|
malloc((MAX_FSIZE + MAX_OHSIZE) * sizeof(uint64_t),
|
|
M_NG_PIPE, M_NOWAIT);
|
|
current->ber = new->ber;
|
|
|
|
/*
|
|
* For given BER and each frame size N (in bytes) calculate
|
|
* the probability P_OK that the frame is clean:
|
|
*
|
|
* P_OK(BER,N) = (1 - 1/BER)^(N*8)
|
|
*
|
|
* We use a 64-bit fixed-point format with decimal point
|
|
* positioned between bits 47 and 48.
|
|
*/
|
|
p0 = one - one / new->ber;
|
|
p = one;
|
|
for (fsize = 0; fsize < MAX_FSIZE + MAX_OHSIZE; fsize++) {
|
|
hinfo->ber_p[fsize] = p;
|
|
for (i = 0; i < 8; i++)
|
|
p = (p * (p0 & 0xffff) >> 48) +
|
|
(p * ((p0 >> 16) & 0xffff) >> 32) +
|
|
(p * (p0 >> 32) >> 16);
|
|
}
|
|
}
|
|
|
|
if (new->qin_size_limit == -1)
|
|
current->qin_size_limit = 0;
|
|
else if (new->qin_size_limit >= 5)
|
|
current->qin_size_limit = new->qin_size_limit;
|
|
|
|
if (new->qout_size_limit == -1)
|
|
current->qout_size_limit = 0;
|
|
else if (new->qout_size_limit >= 5)
|
|
current->qout_size_limit = new->qout_size_limit;
|
|
|
|
if (new->duplicate == -1)
|
|
current->duplicate = 0;
|
|
else if (new->duplicate > 0 && new->duplicate <= 50)
|
|
current->duplicate = new->duplicate;
|
|
|
|
if (new->fifo) {
|
|
current->fifo = 1;
|
|
current->wfq = 0;
|
|
current->drr = 0;
|
|
}
|
|
|
|
if (new->wfq) {
|
|
current->fifo = 0;
|
|
current->wfq = 1;
|
|
current->drr = 0;
|
|
}
|
|
|
|
if (new->drr) {
|
|
current->fifo = 0;
|
|
current->wfq = 0;
|
|
/* DRR quantum */
|
|
if (new->drr >= 32)
|
|
current->drr = new->drr;
|
|
else
|
|
current->drr = 2048; /* default quantum */
|
|
}
|
|
|
|
if (new->droptail) {
|
|
current->droptail = 1;
|
|
current->drophead = 0;
|
|
}
|
|
|
|
if (new->drophead) {
|
|
current->droptail = 0;
|
|
current->drophead = 1;
|
|
}
|
|
|
|
if (new->bandwidth == -1) {
|
|
current->bandwidth = 0;
|
|
current->fifo = 1;
|
|
current->wfq = 0;
|
|
current->drr = 0;
|
|
} else if (new->bandwidth >= 100 && new->bandwidth <= 1000000000)
|
|
current->bandwidth = new->bandwidth;
|
|
|
|
if (current->bandwidth | priv->delay |
|
|
current->duplicate | current->ber)
|
|
hinfo->noqueue = 0;
|
|
else
|
|
hinfo->noqueue = 1;
|
|
}
|
|
|
|
/*
|
|
* Compute a hash signature for a packet. This function suffers from the
|
|
* NIH sindrome, so probably it would be wise to look around what other
|
|
* folks have found out to be a good and efficient IP hash function...
|
|
*/
|
|
static int
|
|
ip_hash(struct mbuf *m, int offset)
|
|
{
|
|
u_int64_t i;
|
|
struct ip *ip = (struct ip *)(mtod(m, u_char *) + offset);
|
|
|
|
if (m->m_len < sizeof(struct ip) + offset ||
|
|
ip->ip_v != 4 || ip->ip_hl << 2 != sizeof(struct ip))
|
|
return 0;
|
|
|
|
i = ((u_int64_t) ip->ip_src.s_addr ^
|
|
((u_int64_t) ip->ip_src.s_addr << 13) ^
|
|
((u_int64_t) ip->ip_dst.s_addr << 7) ^
|
|
((u_int64_t) ip->ip_dst.s_addr << 19));
|
|
return (i ^ (i >> 32));
|
|
}
|
|
|
|
/*
|
|
* Receive data on a hook - both in upstream and downstream direction.
|
|
* We put the frame on the inbound queue, and try to initiate dequeuing
|
|
* sequence immediately. If inbound queue is full, discard one frame
|
|
* depending on dropping policy (from the head or from the tail of the
|
|
* queue).
|
|
*/
|
|
static int
|
|
ngp_rcvdata(hook_p hook, item_p item)
|
|
{
|
|
struct hookinfo *const hinfo = NG_HOOK_PRIVATE(hook);
|
|
const priv_p priv = NG_NODE_PRIVATE(NG_HOOK_NODE(hook));
|
|
struct timeval uuptime;
|
|
struct timeval *now = &uuptime;
|
|
struct ngp_fifo *ngp_f = NULL, *ngp_f1;
|
|
struct ngp_hdr *ngp_h = NULL;
|
|
struct mbuf *m;
|
|
int hash, plen;
|
|
int error = 0;
|
|
|
|
/*
|
|
* Shortcut from inbound to outbound hook when neither of
|
|
* bandwidth, delay, BER or duplication probability is
|
|
* configured, nor we have queued frames to drain.
|
|
*/
|
|
if (hinfo->run.qin_frames == 0 && hinfo->run.qout_frames == 0 &&
|
|
hinfo->noqueue) {
|
|
struct hookinfo *dest;
|
|
if (hinfo == &priv->lower)
|
|
dest = &priv->upper;
|
|
else
|
|
dest = &priv->lower;
|
|
|
|
/* Send the frame. */
|
|
plen = NGI_M(item)->m_pkthdr.len;
|
|
NG_FWD_ITEM_HOOK(error, item, dest->hook);
|
|
|
|
/* Update stats. */
|
|
if (error) {
|
|
hinfo->stats.out_disc_frames++;
|
|
hinfo->stats.out_disc_octets += plen;
|
|
} else {
|
|
hinfo->stats.fwd_frames++;
|
|
hinfo->stats.fwd_octets += plen;
|
|
}
|
|
|
|
return (error);
|
|
}
|
|
|
|
microuptime(now);
|
|
|
|
/*
|
|
* If this was an empty queue, update service deadline time.
|
|
*/
|
|
if (hinfo->run.qin_frames == 0) {
|
|
struct timeval *when = &hinfo->qin_utime;
|
|
if (when->tv_sec < now->tv_sec || (when->tv_sec == now->tv_sec
|
|
&& when->tv_usec < now->tv_usec)) {
|
|
when->tv_sec = now->tv_sec;
|
|
when->tv_usec = now->tv_usec;
|
|
}
|
|
}
|
|
|
|
/* Populate the packet header */
|
|
ngp_h = uma_zalloc(ngp_zone, M_NOWAIT);
|
|
KASSERT((ngp_h != NULL), ("ngp_h zalloc failed (1)"));
|
|
NGI_GET_M(item, m);
|
|
KASSERT(m != NULL, ("NGI_GET_M failed"));
|
|
ngp_h->m = m;
|
|
NG_FREE_ITEM(item);
|
|
|
|
if (hinfo->cfg.fifo)
|
|
hash = 0; /* all packets go into a single FIFO queue */
|
|
else
|
|
hash = ip_hash(m, priv->header_offset);
|
|
|
|
/* Find the appropriate FIFO queue for the packet and enqueue it*/
|
|
TAILQ_FOREACH(ngp_f, &hinfo->fifo_head, fifo_le)
|
|
if (hash == ngp_f->hash)
|
|
break;
|
|
if (ngp_f == NULL) {
|
|
ngp_f = uma_zalloc(ngp_zone, M_NOWAIT);
|
|
KASSERT(ngp_h != NULL, ("ngp_h zalloc failed (2)"));
|
|
TAILQ_INIT(&ngp_f->packet_head);
|
|
ngp_f->hash = hash;
|
|
ngp_f->packets = 1;
|
|
ngp_f->rr_deficit = hinfo->cfg.drr; /* DRR quantum */
|
|
hinfo->run.fifo_queues++;
|
|
TAILQ_INSERT_TAIL(&ngp_f->packet_head, ngp_h, ngp_link);
|
|
FIFO_VTIME_SORT(m->m_pkthdr.len);
|
|
} else {
|
|
TAILQ_INSERT_TAIL(&ngp_f->packet_head, ngp_h, ngp_link);
|
|
ngp_f->packets++;
|
|
}
|
|
hinfo->run.qin_frames++;
|
|
hinfo->run.qin_octets += m->m_pkthdr.len;
|
|
|
|
/* Discard a frame if inbound queue limit has been reached */
|
|
if (hinfo->run.qin_frames > hinfo->cfg.qin_size_limit) {
|
|
struct mbuf *m1;
|
|
int longest = 0;
|
|
|
|
/* Find the longest queue */
|
|
TAILQ_FOREACH(ngp_f1, &hinfo->fifo_head, fifo_le)
|
|
if (ngp_f1->packets > longest) {
|
|
longest = ngp_f1->packets;
|
|
ngp_f = ngp_f1;
|
|
}
|
|
|
|
/* Drop a frame from the queue head/tail, depending on cfg */
|
|
if (hinfo->cfg.drophead)
|
|
ngp_h = TAILQ_FIRST(&ngp_f->packet_head);
|
|
else
|
|
ngp_h = TAILQ_LAST(&ngp_f->packet_head, p_head);
|
|
TAILQ_REMOVE(&ngp_f->packet_head, ngp_h, ngp_link);
|
|
m1 = ngp_h->m;
|
|
uma_zfree(ngp_zone, ngp_h);
|
|
hinfo->run.qin_octets -= m1->m_pkthdr.len;
|
|
hinfo->stats.in_disc_octets += m1->m_pkthdr.len;
|
|
m_freem(m1);
|
|
if (--(ngp_f->packets) == 0) {
|
|
TAILQ_REMOVE(&hinfo->fifo_head, ngp_f, fifo_le);
|
|
uma_zfree(ngp_zone, ngp_f);
|
|
hinfo->run.fifo_queues--;
|
|
}
|
|
hinfo->run.qin_frames--;
|
|
hinfo->stats.in_disc_frames++;
|
|
} else if (hinfo->run.qin_frames > hinfo->cfg.qin_size_limit) {
|
|
struct mbuf *m1;
|
|
int longest = 0;
|
|
|
|
/* Find the longest queue */
|
|
TAILQ_FOREACH(ngp_f1, &hinfo->fifo_head, fifo_le)
|
|
if (ngp_f1->packets > longest) {
|
|
longest = ngp_f1->packets;
|
|
ngp_f = ngp_f1;
|
|
}
|
|
|
|
/* Drop a frame from the queue head/tail, depending on cfg */
|
|
if (hinfo->cfg.drophead)
|
|
ngp_h = TAILQ_FIRST(&ngp_f->packet_head);
|
|
else
|
|
ngp_h = TAILQ_LAST(&ngp_f->packet_head, p_head);
|
|
TAILQ_REMOVE(&ngp_f->packet_head, ngp_h, ngp_link);
|
|
m1 = ngp_h->m;
|
|
uma_zfree(ngp_zone, ngp_h);
|
|
hinfo->run.qin_octets -= m1->m_pkthdr.len;
|
|
hinfo->stats.in_disc_octets += m1->m_pkthdr.len;
|
|
m_freem(m1);
|
|
if (--(ngp_f->packets) == 0) {
|
|
TAILQ_REMOVE(&hinfo->fifo_head, ngp_f, fifo_le);
|
|
uma_zfree(ngp_zone, ngp_f);
|
|
hinfo->run.fifo_queues--;
|
|
}
|
|
hinfo->run.qin_frames--;
|
|
hinfo->stats.in_disc_frames++;
|
|
}
|
|
|
|
/*
|
|
* Try to start the dequeuing process immediately.
|
|
*/
|
|
pipe_dequeue(hinfo, now);
|
|
|
|
return (0);
|
|
}
|
|
|
|
|
|
/*
|
|
* Dequeueing sequence - we basically do the following:
|
|
* 1) Try to extract the frame from the inbound (bandwidth) queue;
|
|
* 2) In accordance to BER specified, discard the frame randomly;
|
|
* 3) If the frame survives BER, prepend it with delay info and move it
|
|
* to outbound (delay) queue;
|
|
* 4) Loop to 2) until bandwidth quota for this timeslice is reached, or
|
|
* inbound queue is flushed completely;
|
|
* 5) Dequeue frames from the outbound queue and send them downstream until
|
|
* outbound queue is flushed completely, or the next frame in the queue
|
|
* is not due to be dequeued yet
|
|
*/
|
|
static void
|
|
pipe_dequeue(struct hookinfo *hinfo, struct timeval *now) {
|
|
static uint64_t rand, oldrand;
|
|
const node_p node = NG_HOOK_NODE(hinfo->hook);
|
|
const priv_p priv = NG_NODE_PRIVATE(node);
|
|
struct hookinfo *dest;
|
|
struct ngp_fifo *ngp_f, *ngp_f1;
|
|
struct ngp_hdr *ngp_h;
|
|
struct timeval *when;
|
|
struct mbuf *m;
|
|
int plen, error = 0;
|
|
|
|
/* Which one is the destination hook? */
|
|
if (hinfo == &priv->lower)
|
|
dest = &priv->upper;
|
|
else
|
|
dest = &priv->lower;
|
|
|
|
/* Bandwidth queue processing */
|
|
while ((ngp_f = TAILQ_FIRST(&hinfo->fifo_head))) {
|
|
when = &hinfo->qin_utime;
|
|
if (when->tv_sec > now->tv_sec || (when->tv_sec == now->tv_sec
|
|
&& when->tv_usec > now->tv_usec))
|
|
break;
|
|
|
|
ngp_h = TAILQ_FIRST(&ngp_f->packet_head);
|
|
m = ngp_h->m;
|
|
|
|
/* Deficit Round Robin (DRR) processing */
|
|
if (hinfo->cfg.drr) {
|
|
if (ngp_f->rr_deficit >= m->m_pkthdr.len) {
|
|
ngp_f->rr_deficit -= m->m_pkthdr.len;
|
|
} else {
|
|
ngp_f->rr_deficit += hinfo->cfg.drr;
|
|
TAILQ_REMOVE(&hinfo->fifo_head, ngp_f, fifo_le);
|
|
TAILQ_INSERT_TAIL(&hinfo->fifo_head,
|
|
ngp_f, fifo_le);
|
|
continue;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Either create a duplicate and pass it on, or dequeue
|
|
* the original packet...
|
|
*/
|
|
if (hinfo->cfg.duplicate &&
|
|
random() % 100 <= hinfo->cfg.duplicate) {
|
|
ngp_h = uma_zalloc(ngp_zone, M_NOWAIT);
|
|
KASSERT(ngp_h != NULL, ("ngp_h zalloc failed (3)"));
|
|
m = m_dup(m, M_NOWAIT);
|
|
KASSERT(m != NULL, ("m_dup failed"));
|
|
ngp_h->m = m;
|
|
} else {
|
|
TAILQ_REMOVE(&ngp_f->packet_head, ngp_h, ngp_link);
|
|
hinfo->run.qin_frames--;
|
|
hinfo->run.qin_octets -= m->m_pkthdr.len;
|
|
ngp_f->packets--;
|
|
}
|
|
|
|
/* Calculate the serialization delay */
|
|
if (hinfo->cfg.bandwidth) {
|
|
hinfo->qin_utime.tv_usec +=
|
|
((uint64_t) m->m_pkthdr.len + priv->overhead ) *
|
|
8000000 / hinfo->cfg.bandwidth;
|
|
hinfo->qin_utime.tv_sec +=
|
|
hinfo->qin_utime.tv_usec / 1000000;
|
|
hinfo->qin_utime.tv_usec =
|
|
hinfo->qin_utime.tv_usec % 1000000;
|
|
}
|
|
when = &ngp_h->when;
|
|
when->tv_sec = hinfo->qin_utime.tv_sec;
|
|
when->tv_usec = hinfo->qin_utime.tv_usec;
|
|
|
|
/* Sort / rearrange inbound queues */
|
|
if (ngp_f->packets) {
|
|
if (hinfo->cfg.wfq) {
|
|
TAILQ_REMOVE(&hinfo->fifo_head, ngp_f, fifo_le);
|
|
FIFO_VTIME_SORT(TAILQ_FIRST(
|
|
&ngp_f->packet_head)->m->m_pkthdr.len)
|
|
}
|
|
} else {
|
|
TAILQ_REMOVE(&hinfo->fifo_head, ngp_f, fifo_le);
|
|
uma_zfree(ngp_zone, ngp_f);
|
|
hinfo->run.fifo_queues--;
|
|
}
|
|
|
|
/* Randomly discard the frame, according to BER setting */
|
|
if (hinfo->cfg.ber) {
|
|
oldrand = rand;
|
|
rand = random();
|
|
if (((oldrand ^ rand) << 17) >=
|
|
hinfo->ber_p[priv->overhead + m->m_pkthdr.len]) {
|
|
hinfo->stats.out_disc_frames++;
|
|
hinfo->stats.out_disc_octets += m->m_pkthdr.len;
|
|
uma_zfree(ngp_zone, ngp_h);
|
|
m_freem(m);
|
|
continue;
|
|
}
|
|
}
|
|
|
|
/* Discard frame if outbound queue size limit exceeded */
|
|
if (hinfo->cfg.qout_size_limit &&
|
|
hinfo->run.qout_frames>=hinfo->cfg.qout_size_limit) {
|
|
hinfo->stats.out_disc_frames++;
|
|
hinfo->stats.out_disc_octets += m->m_pkthdr.len;
|
|
uma_zfree(ngp_zone, ngp_h);
|
|
m_freem(m);
|
|
continue;
|
|
}
|
|
|
|
/* Calculate the propagation delay */
|
|
when->tv_usec += priv->delay;
|
|
when->tv_sec += when->tv_usec / 1000000;
|
|
when->tv_usec = when->tv_usec % 1000000;
|
|
|
|
/* Put the frame into the delay queue */
|
|
TAILQ_INSERT_TAIL(&hinfo->qout_head, ngp_h, ngp_link);
|
|
hinfo->run.qout_frames++;
|
|
hinfo->run.qout_octets += m->m_pkthdr.len;
|
|
}
|
|
|
|
/* Delay queue processing */
|
|
while ((ngp_h = TAILQ_FIRST(&hinfo->qout_head))) {
|
|
when = &ngp_h->when;
|
|
m = ngp_h->m;
|
|
if (when->tv_sec > now->tv_sec ||
|
|
(when->tv_sec == now->tv_sec &&
|
|
when->tv_usec > now->tv_usec))
|
|
break;
|
|
|
|
/* Update outbound queue stats */
|
|
plen = m->m_pkthdr.len;
|
|
hinfo->run.qout_frames--;
|
|
hinfo->run.qout_octets -= plen;
|
|
|
|
/* Dequeue the packet from qout */
|
|
TAILQ_REMOVE(&hinfo->qout_head, ngp_h, ngp_link);
|
|
uma_zfree(ngp_zone, ngp_h);
|
|
|
|
NG_SEND_DATA(error, dest->hook, m, meta);
|
|
if (error) {
|
|
hinfo->stats.out_disc_frames++;
|
|
hinfo->stats.out_disc_octets += plen;
|
|
} else {
|
|
hinfo->stats.fwd_frames++;
|
|
hinfo->stats.fwd_octets += plen;
|
|
}
|
|
}
|
|
|
|
if ((hinfo->run.qin_frames != 0 || hinfo->run.qout_frames != 0) &&
|
|
!priv->timer_scheduled) {
|
|
ng_callout(&priv->timer, node, NULL, 1, ngp_callout, NULL, 0);
|
|
priv->timer_scheduled = 1;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* This routine is called on every clock tick. We poll connected hooks
|
|
* for queued frames by calling pipe_dequeue().
|
|
*/
|
|
static void
|
|
ngp_callout(node_p node, hook_p hook, void *arg1, int arg2)
|
|
{
|
|
const priv_p priv = NG_NODE_PRIVATE(node);
|
|
struct timeval now;
|
|
|
|
priv->timer_scheduled = 0;
|
|
microuptime(&now);
|
|
if (priv->upper.hook != NULL)
|
|
pipe_dequeue(&priv->upper, &now);
|
|
if (priv->lower.hook != NULL)
|
|
pipe_dequeue(&priv->lower, &now);
|
|
}
|
|
|
|
/*
|
|
* Shutdown processing
|
|
*
|
|
* This is tricky. If we have both a lower and upper hook, then we
|
|
* probably want to extricate ourselves and leave the two peers
|
|
* still linked to each other. Otherwise we should just shut down as
|
|
* a normal node would.
|
|
*/
|
|
static int
|
|
ngp_shutdown(node_p node)
|
|
{
|
|
const priv_p priv = NG_NODE_PRIVATE(node);
|
|
|
|
if (priv->timer_scheduled)
|
|
ng_uncallout(&priv->timer, node);
|
|
if (priv->lower.hook && priv->upper.hook)
|
|
ng_bypass(priv->lower.hook, priv->upper.hook);
|
|
else {
|
|
if (priv->upper.hook != NULL)
|
|
ng_rmhook_self(priv->upper.hook);
|
|
if (priv->lower.hook != NULL)
|
|
ng_rmhook_self(priv->lower.hook);
|
|
}
|
|
NG_NODE_UNREF(node);
|
|
free(priv, M_NG_PIPE);
|
|
return (0);
|
|
}
|
|
|
|
|
|
/*
|
|
* Hook disconnection
|
|
*/
|
|
static int
|
|
ngp_disconnect(hook_p hook)
|
|
{
|
|
struct hookinfo *const hinfo = NG_HOOK_PRIVATE(hook);
|
|
struct ngp_fifo *ngp_f;
|
|
struct ngp_hdr *ngp_h;
|
|
|
|
KASSERT(hinfo != NULL, ("%s: null info", __FUNCTION__));
|
|
hinfo->hook = NULL;
|
|
|
|
/* Flush all fifo queues associated with the hook */
|
|
while ((ngp_f = TAILQ_FIRST(&hinfo->fifo_head))) {
|
|
while ((ngp_h = TAILQ_FIRST(&ngp_f->packet_head))) {
|
|
TAILQ_REMOVE(&ngp_f->packet_head, ngp_h, ngp_link);
|
|
m_freem(ngp_h->m);
|
|
uma_zfree(ngp_zone, ngp_h);
|
|
}
|
|
TAILQ_REMOVE(&hinfo->fifo_head, ngp_f, fifo_le);
|
|
uma_zfree(ngp_zone, ngp_f);
|
|
}
|
|
|
|
/* Flush the delay queue */
|
|
while ((ngp_h = TAILQ_FIRST(&hinfo->qout_head))) {
|
|
TAILQ_REMOVE(&hinfo->qout_head, ngp_h, ngp_link);
|
|
m_freem(ngp_h->m);
|
|
uma_zfree(ngp_zone, ngp_h);
|
|
}
|
|
|
|
/* Release the packet loss probability table (BER) */
|
|
if (hinfo->ber_p)
|
|
free(hinfo->ber_p, M_NG_PIPE);
|
|
|
|
return (0);
|
|
}
|
|
|
|
static int
|
|
ngp_modevent(module_t mod, int type, void *unused)
|
|
{
|
|
int error = 0;
|
|
|
|
switch (type) {
|
|
case MOD_LOAD:
|
|
ngp_zone = uma_zcreate("ng_pipe", max(sizeof(struct ngp_hdr),
|
|
sizeof (struct ngp_fifo)), NULL, NULL, NULL, NULL,
|
|
UMA_ALIGN_PTR, 0);
|
|
if (ngp_zone == NULL)
|
|
panic("ng_pipe: couldn't allocate descriptor zone");
|
|
break;
|
|
case MOD_UNLOAD:
|
|
uma_zdestroy(ngp_zone);
|
|
break;
|
|
default:
|
|
error = EOPNOTSUPP;
|
|
break;
|
|
}
|
|
|
|
return (error);
|
|
}
|