freebsd-skq/usr.sbin/mrouted/prune.c
kris aec252afab String buffer safety cleanup. I don't think any of these were exploitable
remotely, but they would be if e.g. it happened to call the logging
function using a DNS hostname.

Also replace random() by arc4random() - only one of these is arguably
required since it's directly used in the protocol, but we might as
well replace both to avoid using two different PRNGs.

Reviewed by:	green, alex
2000-10-09 06:08:00 +00:00

2620 lines
68 KiB
C

/*
* The mrouted program is covered by the license in the accompanying file
* named "LICENSE". Use of the mrouted program represents acceptance of
* the terms and conditions listed in that file.
*
* The mrouted program is COPYRIGHT 1989 by The Board of Trustees of
* Leland Stanford Junior University.
*
*
* prune.c,v 3.8.4.59 1998/03/01 02:06:32 fenner Exp
*/
#ifndef lint
static const char rcsid[] =
"$FreeBSD$";
#endif /* not lint */
#include "defs.h"
extern int cache_lifetime;
extern int prune_lifetime;
extern struct rtentry *routing_table;
extern int phys_vif;
extern int allow_black_holes;
/*
* randomize value to obtain a value between .5x and 1.5x
* in order to prevent synchronization
*/
#ifdef SYSV
#define JITTERED_VALUE(x) ((x)/2 + (lrand48() % (x)))
#else
#define JITTERED_VALUE(x) ((x)/2 + (arc4random() % (x)))
#endif
#define CACHE_LIFETIME(x) JITTERED_VALUE(x) /* XXX */
struct gtable *kernel_table; /* ptr to list of kernel grp entries*/
static struct gtable *kernel_no_route; /* list of grp entries w/o routes */
struct gtable *gtp; /* pointer for kernel rt entries */
unsigned int kroutes; /* current number of cache entries */
/****************************************************************************
Functions that are local to prune.c
****************************************************************************/
static int scoped_addr __P((vifi_t vifi, u_int32 addr));
static void prun_add_ttls __P((struct gtable *gt));
static int pruning_neighbor __P((vifi_t vifi, u_int32 addr));
static int can_mtrace __P((vifi_t vifi, u_int32 addr));
static struct ptable * find_prune_entry __P((u_int32 vr, struct ptable *pt));
static void remove_sources __P((struct gtable *gt));
static void rexmit_prune __P((void *arg));
static void expire_prune __P((vifi_t vifi, struct gtable *gt));
static void send_prune __P((struct gtable *gt));
static void send_graft __P((struct gtable *gt));
static void send_graft_ack __P((u_int32 src, u_int32 dst,
u_int32 origin, u_int32 grp,
vifi_t vifi));
static void update_kernel __P((struct gtable *g));
/*
* Updates the ttl values for each vif.
*/
static void
prun_add_ttls(gt)
struct gtable *gt;
{
struct uvif *v;
vifi_t vifi;
for (vifi = 0, v = uvifs; vifi < numvifs; ++vifi, ++v) {
if (VIFM_ISSET(vifi, gt->gt_grpmems))
gt->gt_ttls[vifi] = v->uv_threshold;
else
gt->gt_ttls[vifi] = 0;
}
}
/*
* checks for scoped multicast addresses
* XXX I want to make the check of allow_black_holes based on ALLOW_BLACK_HOLES
* but macros are not functions.
*/
#define GET_SCOPE(gt) { \
register vifi_t _i; \
VIFM_CLRALL((gt)->gt_scope); \
if (allow_black_holes || \
(ntohl((gt)->gt_mcastgrp) & 0xff000000) == 0xef000000) \
for (_i = 0; _i < numvifs; _i++) \
if (scoped_addr(_i, (gt)->gt_mcastgrp)) \
VIFM_SET(_i, (gt)->gt_scope); \
} \
if ((gt)->gt_route == NULL || ((gt)->gt_route->rt_parent != NO_VIF && \
VIFM_ISSET((gt)->gt_route->rt_parent, (gt)->gt_scope))) \
VIFM_SETALL((gt)->gt_scope);
#define APPLY_SCOPE(gt) VIFM_CLR_MASK((gt)->gt_grpmems, (gt)->gt_scope)
#define GET_MEMBERSHIP(gt, vifi) { \
if ((gt)->gt_route && \
VIFM_ISSET((vifi), (gt)->gt_route->rt_children) && \
(!SUBS_ARE_PRUNED((gt)->gt_route->rt_subordinates, \
uvifs[vifi].uv_nbrmap, (gt)->gt_prunes) || \
grplst_mem((vifi), (gt)->gt_mcastgrp))) \
VIFM_SET((vifi), (gt)->gt_grpmems); \
}
static int
scoped_addr(vifi, addr)
vifi_t vifi;
u_int32 addr;
{
struct vif_acl *acl;
for (acl = uvifs[vifi].uv_acl; acl; acl = acl->acl_next)
if ((addr & acl->acl_mask) == acl->acl_addr)
return 1;
return 0;
}
/*
* Determine the list of outgoing vifs, based upon
* route subordinates, prunes received, and group
* memberships.
*/
void
determine_forwvifs(gt)
struct gtable *gt;
{
vifi_t i;
VIFM_CLRALL(gt->gt_grpmems);
for (i = 0; i < numvifs; i++) {
GET_MEMBERSHIP(gt, i);
}
GET_SCOPE(gt);
APPLY_SCOPE(gt);
}
/*
* Send a prune or a graft if necessary.
*/
void
send_prune_or_graft(gt)
struct gtable *gt;
{
if (VIFM_ISEMPTY(gt->gt_grpmems))
send_prune(gt);
else if (gt->gt_prsent_timer)
send_graft(gt);
}
/*
* Determine if mcastgrp has a listener on vifi
*/
int
grplst_mem(vifi, mcastgrp)
vifi_t vifi;
u_int32 mcastgrp;
{
register struct listaddr *g;
register struct uvif *v;
v = &uvifs[vifi];
for (g = v->uv_groups; g != NULL; g = g->al_next)
if (mcastgrp == g->al_addr)
return 1;
return 0;
}
/*
* Finds the group entry with the specified source and netmask.
* If netmask is 0, it uses the route's netmask.
*
* Returns TRUE if found a match, and the global variable gtp is left
* pointing to entry before the found entry.
* Returns FALSE if no exact match found, gtp is left pointing to before
* the entry in question belongs, or is NULL if the it belongs at the
* head of the list.
*/
int
find_src_grp(src, mask, grp)
u_int32 src;
u_int32 mask;
u_int32 grp;
{
struct gtable *gt;
gtp = NULL;
gt = kernel_table;
while (gt != NULL) {
if (grp == gt->gt_mcastgrp &&
(mask ? (gt->gt_route->rt_origin == src &&
gt->gt_route->rt_originmask == mask) :
((src & gt->gt_route->rt_originmask) ==
gt->gt_route->rt_origin)))
return TRUE;
if (ntohl(grp) > ntohl(gt->gt_mcastgrp) ||
(grp == gt->gt_mcastgrp &&
(ntohl(mask) < ntohl(gt->gt_route->rt_originmask) ||
(mask == gt->gt_route->rt_originmask &&
(ntohl(src) > ntohl(gt->gt_route->rt_origin)))))) {
gtp = gt;
gt = gt->gt_gnext;
}
else break;
}
return FALSE;
}
/*
* Check if the neighbor supports pruning
*/
static int
pruning_neighbor(vifi, addr)
vifi_t vifi;
u_int32 addr;
{
struct listaddr *n = neighbor_info(vifi, addr);
int vers;
if (n == NULL)
return 0;
vers = NBR_VERS(n);
return (vers >= 0x0300 && ((vers & 0xff00) != 0x0a00));
}
/*
* Can the neighbor in question handle multicast traceroute?
*/
static int
can_mtrace(vifi, addr)
vifi_t vifi;
u_int32 addr;
{
struct listaddr *n = neighbor_info(vifi, addr);
int vers;
if (n == NULL)
return 1; /* fail "safe" */
vers = NBR_VERS(n);
return (vers >= 0x0303 && ((vers & 0xff00) != 0x0a00));
}
/*
* Returns the prune entry of the router, or NULL if none exists
*/
static struct ptable *
find_prune_entry(vr, pt)
u_int32 vr;
struct ptable *pt;
{
while (pt) {
if (pt->pt_router == vr)
return pt;
pt = pt->pt_next;
}
return NULL;
}
/*
* Remove all the sources hanging off the group table entry from the kernel
* cache. Remember the packet counts wherever possible, to keep the mtrace
* counters consistent. This prepares for possible prune retransmission,
* either on a multi-access network or when a prune that we sent upstream
* has expired.
*/
static void
remove_sources(gt)
struct gtable *gt;
{
struct stable *st;
struct sioc_sg_req sg_req;
sg_req.grp.s_addr = gt->gt_mcastgrp;
/*
* call k_del_rg() on every one of the gt->gt_srctbl entries
* but first save the packet count so that the mtrace packet
* counters can remain approximately correct. There's a race
* here but it's minor.
*/
for (st = gt->gt_srctbl; st; st = st->st_next) {
if (st->st_ctime == 0)
continue;
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "rexmit_prune deleting (%s %s) (next is %d sec)",
inet_fmt(st->st_origin, s1),
inet_fmt(gt->gt_mcastgrp, s2),
gt->gt_prune_rexmit);
sg_req.src.s_addr = st->st_origin;
if (ioctl(udp_socket, SIOCGETSGCNT, (char *)&sg_req) < 0) {
sg_req.pktcnt = 0;
}
k_del_rg(st->st_origin, gt);
st->st_ctime = 0; /* flag that it's not in the kernel any more */
st->st_savpkt += sg_req.pktcnt;
kroutes--;
}
/*
* Now, add_table_entry will prune when asked to add a cache entry.
*/
}
/*
* Prepare for possible prune retransmission
*/
static void
rexmit_prune(arg)
void *arg;
{
struct gtable *gt = *(struct gtable **)arg;
free(arg);
gt->gt_rexmit_timer = 0;
/* Make sure we're still not forwarding traffic */
if (!VIFM_ISEMPTY(gt->gt_grpmems)) {
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "rexmit_prune (%s %s): gm:%x",
RT_FMT(gt->gt_route, s1), inet_fmt(gt->gt_mcastgrp, s2),
gt->gt_grpmems);
return;
}
remove_sources(gt);
}
/*
* Send a prune message to the dominant router for
* this source.
*
* Record an entry that a prune was sent for this group
*/
static void
send_prune(gt)
struct gtable *gt;
{
struct ptable *pt;
char *p;
int i;
int datalen;
u_int32 dst;
u_int32 tmp;
int rexmitting = 0;
struct uvif *v;
/*
* Can't process a prune if we don't have an associated route
* or if the route points to a local interface.
*/
if (gt->gt_route == NULL || gt->gt_route->rt_parent == NO_VIF ||
gt->gt_route->rt_gateway == 0)
return;
/* Don't send a prune to a non-pruning router */
if (!pruning_neighbor(gt->gt_route->rt_parent, gt->gt_route->rt_gateway))
return;
v = &uvifs[gt->gt_route->rt_parent];
/*
* sends a prune message to the router upstream.
*/
#if 0
dst = v->uv_flags & VIFF_TUNNEL ? dvmrp_group : gt->gt_route->rt_gateway; /*XXX*/
#else
dst = gt->gt_route->rt_gateway;
#endif
p = send_buf + MIN_IP_HEADER_LEN + IGMP_MINLEN;
datalen = 0;
/*
* determine prune lifetime, if this isn't a retransmission.
*
* Use interface-specified lifetime if there is one.
*/
if (gt->gt_prsent_timer == 0) {
int l = prune_lifetime;
if (v->uv_prune_lifetime != 0)
l = v->uv_prune_lifetime;
gt->gt_prsent_timer = JITTERED_VALUE(l);
for (pt = gt->gt_pruntbl; pt; pt = pt->pt_next)
if (pt->pt_timer < gt->gt_prsent_timer)
gt->gt_prsent_timer = pt->pt_timer;
} else if (gt->gt_prsent_timer < 0) {
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "asked to rexmit? (%s,%s)/%d on vif %d to %s with negative time",
RT_FMT(gt->gt_route, s1), inet_fmt(gt->gt_mcastgrp, s2),
gt->gt_prsent_timer, gt->gt_route->rt_parent,
inet_fmt(gt->gt_route->rt_gateway, s3));
return;
} else
rexmitting = 1;
if (rexmitting && !(v->uv_flags & VIFF_REXMIT_PRUNES)) {
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "not rexmitting prune for (%s %s)/%d on vif %d to %s",
RT_FMT(gt->gt_route, s1), inet_fmt(gt->gt_mcastgrp, s2),
gt->gt_prsent_timer, gt->gt_route->rt_parent,
inet_fmt(gt->gt_route->rt_gateway, s3));
return;
}
if (gt->gt_prsent_timer <= MIN_PRUNE_LIFE) {
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "not bothering to send prune for (%s,%s)/%d on vif %d to %s because it's too short",
RT_FMT(gt->gt_route, s1), inet_fmt(gt->gt_mcastgrp, s2),
gt->gt_prsent_timer, gt->gt_route->rt_parent,
inet_fmt(gt->gt_route->rt_gateway, s3));
return;
}
/*
* If we have a graft pending, cancel graft retransmission
*/
gt->gt_grftsnt = 0;
for (i = 0; i < 4; i++)
*p++ = ((char *)&(gt->gt_route->rt_origin))[i];
for (i = 0; i < 4; i++)
*p++ = ((char *)&(gt->gt_mcastgrp))[i];
tmp = htonl(gt->gt_prsent_timer);
for (i = 0; i < 4; i++)
*p++ = ((char *)&(tmp))[i];
datalen += 12;
send_on_vif(v, dst, DVMRP_PRUNE, datalen);
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "%s prune for (%s %s)/%d on vif %d to %s",
rexmitting ? "rexmitted" : "sent",
RT_FMT(gt->gt_route, s1), inet_fmt(gt->gt_mcastgrp, s2),
gt->gt_prsent_timer, gt->gt_route->rt_parent,
inet_fmt(gt->gt_route->rt_gateway, s3));
if ((v->uv_flags & VIFF_REXMIT_PRUNES) &&
gt->gt_rexmit_timer == 0 &&
gt->gt_prsent_timer > gt->gt_prune_rexmit) {
struct gtable **arg =
(struct gtable **)malloc(sizeof (struct gtable **));
*arg = gt;
gt->gt_rexmit_timer = timer_setTimer(
JITTERED_VALUE(gt->gt_prune_rexmit),
rexmit_prune, arg);
gt->gt_prune_rexmit *= 2;
}
}
/*
* a prune was sent upstream
* so, a graft has to be sent to annul the prune
* set up a graft timer so that if an ack is not
* heard within that time, another graft request
* is sent out.
*/
static void
send_graft(gt)
struct gtable *gt;
{
register char *p;
register int i;
int datalen;
u_int32 dst;
/* Can't send a graft without an associated route */
if (gt->gt_route == NULL || gt->gt_route->rt_parent == NO_VIF) {
gt->gt_grftsnt = 0;
return;
}
gt->gt_prsent_timer = 0;
gt->gt_prune_rexmit = PRUNE_REXMIT_VAL;
if (gt->gt_rexmit_timer)
timer_clearTimer(gt->gt_rexmit_timer);
if (gt->gt_grftsnt == 0)
gt->gt_grftsnt = 1;
#if 0
dst = uvifs[gt->gt_route->rt_parent].uv_flags & VIFF_TUNNEL ? dvmrp_group : gt->gt_route->rt_gateway; /*XXX*/
#else
dst = gt->gt_route->rt_gateway;
#endif
p = send_buf + MIN_IP_HEADER_LEN + IGMP_MINLEN;
datalen = 0;
for (i = 0; i < 4; i++)
*p++ = ((char *)&(gt->gt_route->rt_origin))[i];
for (i = 0; i < 4; i++)
*p++ = ((char *)&(gt->gt_mcastgrp))[i];
datalen += 8;
send_on_vif(&uvifs[gt->gt_route->rt_parent], dst, DVMRP_GRAFT, datalen);
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "sent graft for (%s %s) to %s on vif %d",
RT_FMT(gt->gt_route, s1), inet_fmt(gt->gt_mcastgrp, s2),
inet_fmt(gt->gt_route->rt_gateway, s3), gt->gt_route->rt_parent);
}
/*
* Send an ack that a graft was received
*/
static void
send_graft_ack(src, dst, origin, grp, vifi)
u_int32 src;
u_int32 dst;
u_int32 origin;
u_int32 grp;
vifi_t vifi;
{
register char *p;
register int i;
int datalen;
p = send_buf + MIN_IP_HEADER_LEN + IGMP_MINLEN;
datalen = 0;
for (i = 0; i < 4; i++)
*p++ = ((char *)&(origin))[i];
for (i = 0; i < 4; i++)
*p++ = ((char *)&(grp))[i];
datalen += 8;
if (vifi == NO_VIF)
send_igmp(src, dst, IGMP_DVMRP, DVMRP_GRAFT_ACK,
htonl(MROUTED_LEVEL), datalen);
else {
#if 0
if (uvifs[vifi].uv_flags & VIFF_TUNNEL)
dst = dvmrp_group; /* XXX */
#endif
send_on_vif(&uvifs[vifi], dst, DVMRP_GRAFT_ACK, datalen);
}
IF_DEBUG(DEBUG_PRUNE)
if (vifi == NO_VIF)
log(LOG_DEBUG, 0, "sent graft ack for (%s, %s) to %s",
inet_fmt(origin, s1), inet_fmt(grp, s2), inet_fmt(dst, s3));
else
log(LOG_DEBUG, 0, "sent graft ack for (%s, %s) to %s on vif %d",
inet_fmt(origin, s1), inet_fmt(grp, s2), inet_fmt(dst, s3), vifi);
}
/*
* Update the kernel cache with all the routes hanging off the group entry
*/
static void
update_kernel(g)
struct gtable *g;
{
struct stable *st;
for (st = g->gt_srctbl; st; st = st->st_next)
if (st->st_ctime != 0)
k_add_rg(st->st_origin, g);
}
/****************************************************************************
Functions that are used externally
****************************************************************************/
#ifdef SNMP
#include <sys/types.h>
#include "snmp.h"
/*
* Find a specific group entry in the group table
*/
struct gtable *
find_grp(grp)
u_int32 grp;
{
struct gtable *gt;
for (gt = kernel_table; gt; gt = gt->gt_gnext) {
if (ntohl(grp) < ntohl(gt->gt_mcastgrp))
break;
if (gt->gt_mcastgrp == grp)
return gt;
}
return NULL;
}
/*
* Given a group entry and source, find the corresponding source table
* entry
*/
struct stable *
find_grp_src(gt, src)
struct gtable *gt;
u_int32 src;
{
struct stable *st;
u_long grp = gt->gt_mcastgrp;
struct gtable *gtcurr;
for (gtcurr = gt; gtcurr->gt_mcastgrp == grp; gtcurr = gtcurr->gt_gnext) {
for (st = gtcurr->gt_srctbl; st; st = st->st_next)
if (st->st_origin == src)
return st;
}
return NULL;
}
/*
* Find next entry > specification
*/
int
next_grp_src_mask(gtpp, stpp, grp, src, mask)
struct gtable **gtpp; /* ordered by group */
struct stable **stpp; /* ordered by source */
u_int32 grp;
u_int32 src;
u_int32 mask;
{
struct gtable *gt, *gbest = NULL;
struct stable *st, *sbest = NULL;
/* Find first group entry >= grp spec */
(*gtpp) = kernel_table;
while ((*gtpp) && ntohl((*gtpp)->gt_mcastgrp) < ntohl(grp))
(*gtpp)=(*gtpp)->gt_gnext;
if (!(*gtpp))
return 0; /* no more groups */
for (gt = kernel_table; gt; gt=gt->gt_gnext) {
/* Since grps are ordered, we can stop when group changes from gbest */
if (gbest && gbest->gt_mcastgrp != gt->gt_mcastgrp)
break;
for (st = gt->gt_srctbl; st; st=st->st_next) {
/* Among those entries > spec, find "lowest" one */
if (((ntohl(gt->gt_mcastgrp)> ntohl(grp))
|| (ntohl(gt->gt_mcastgrp)==ntohl(grp)
&& ntohl(st->st_origin)> ntohl(src))
|| (ntohl(gt->gt_mcastgrp)==ntohl(grp)
&& ntohl(st->st_origin)==src && 0xFFFFFFFF>ntohl(mask)))
&& (!gbest
|| (ntohl(gt->gt_mcastgrp)< ntohl(gbest->gt_mcastgrp))
|| (ntohl(gt->gt_mcastgrp)==ntohl(gbest->gt_mcastgrp)
&& ntohl(st->st_origin)< ntohl(sbest->st_origin)))) {
gbest = gt;
sbest = st;
}
}
}
(*gtpp) = gbest;
(*stpp) = sbest;
return (*gtpp)!=0;
}
/*
* Ensure that sg contains current information for the given group,source.
* This is fetched from the kernel as a unit so that counts for the entry
* are consistent, i.e. packet and byte counts for the same entry are
* read at the same time.
*/
void
refresh_sg(sg, gt, st)
struct sioc_sg_req *sg;
struct gtable *gt;
struct stable *st;
{
static int lastq = -1;
if (quantum != lastq || sg->src.s_addr!=st->st_origin
|| sg->grp.s_addr!=gt->gt_mcastgrp) {
lastq = quantum;
sg->src.s_addr = st->st_origin;
sg->grp.s_addr = gt->gt_mcastgrp;
ioctl(udp_socket, SIOCGETSGCNT, (char *)sg);
}
}
/*
* Given a routing table entry, and a vifi, find the next entry
* equal to or greater than those
*/
int
next_child(gtpp, stpp, grp, src, mask, vifi)
struct gtable **gtpp;
struct stable **stpp;
u_int32 grp;
u_int32 src;
u_int32 mask;
vifi_t *vifi; /* vif at which to start looking */
{
/* Get (G,S,M) entry */
if (mask!=0xFFFFFFFF
|| !((*gtpp) = find_grp(grp))
|| !((*stpp) = find_grp_src((*gtpp),src)))
if (!next_grp_src_mask(gtpp, stpp, grp, src, mask))
return 0;
/* Continue until we get one with a valid next vif */
do {
for (; (*gtpp)->gt_route->rt_children && *vifi<numvifs; (*vifi)++)
if (VIFM_ISSET(*vifi, (*gtpp)->gt_route->rt_children))
return 1;
*vifi = 0;
} while (next_grp_src_mask(gtpp, stpp, (*gtpp)->gt_mcastgrp,
(*stpp)->st_origin, 0xFFFFFFFF) );
return 0;
}
#endif /* SNMP */
/*
* Initialize the kernel table structure
*/
void
init_ktable()
{
kernel_table = NULL;
kernel_no_route = NULL;
kroutes = 0;
}
/*
* Add a new table entry for (origin, mcastgrp)
*/
void
add_table_entry(origin, mcastgrp)
u_int32 origin;
u_int32 mcastgrp;
{
struct rtentry *r;
struct gtable *gt,**gtnp,*prev_gt;
struct stable *st,**stnp;
/*
* Since we have to enable mrouting to get the version number,
* some cache creation requests can sneak through. Ignore them
* since we're not going to do useful stuff until we've performed
* final initialization.
*/
if (!did_final_init)
return;
#ifdef DEBUG_MFC
md_log(MD_MISS, origin, mcastgrp);
#endif
r = determine_route(origin);
prev_gt = NULL;
if (r == NULL) {
/*
* Look for it on the no_route table; if it is found then
* it will be detected as a duplicate below.
*/
for (gt = kernel_no_route; gt; gt = gt->gt_next)
if (mcastgrp == gt->gt_mcastgrp &&
gt->gt_srctbl && gt->gt_srctbl->st_origin == origin)
break;
gtnp = &kernel_no_route;
} else {
gtnp = &r->rt_groups;
while ((gt = *gtnp) != NULL) {
if (gt->gt_mcastgrp >= mcastgrp)
break;
gtnp = &gt->gt_next;
prev_gt = gt;
}
}
if (gt == NULL || gt->gt_mcastgrp != mcastgrp) {
gt = (struct gtable *)malloc(sizeof(struct gtable));
if (gt == NULL)
log(LOG_ERR, 0, "ran out of memory");
gt->gt_mcastgrp = mcastgrp;
gt->gt_timer = CACHE_LIFETIME(cache_lifetime);
time(&gt->gt_ctime);
gt->gt_prsent_timer = 0;
gt->gt_grftsnt = 0;
gt->gt_srctbl = NULL;
gt->gt_pruntbl = NULL;
gt->gt_route = r;
gt->gt_rexmit_timer = 0;
NBRM_CLRALL(gt->gt_prunes);
gt->gt_prune_rexmit = PRUNE_REXMIT_VAL;
#ifdef RSRR
gt->gt_rsrr_cache = NULL;
#endif
/* Calculate forwarding vifs */
determine_forwvifs(gt);
/* update ttls */
prun_add_ttls(gt);
gt->gt_next = *gtnp;
*gtnp = gt;
if (gt->gt_next)
gt->gt_next->gt_prev = gt;
gt->gt_prev = prev_gt;
if (r) {
if (find_src_grp(r->rt_origin, r->rt_originmask, gt->gt_mcastgrp)) {
struct gtable *g;
g = gtp ? gtp->gt_gnext : kernel_table;
log(LOG_WARNING, 0, "Entry for (%s %s) (rt:%x) exists (rt:%x)",
RT_FMT(r, s1), inet_fmt(g->gt_mcastgrp, s2),
r, g->gt_route);
} else {
if (gtp) {
gt->gt_gnext = gtp->gt_gnext;
gt->gt_gprev = gtp;
gtp->gt_gnext = gt;
} else {
gt->gt_gnext = kernel_table;
gt->gt_gprev = NULL;
kernel_table = gt;
}
if (gt->gt_gnext)
gt->gt_gnext->gt_gprev = gt;
}
} else {
gt->gt_gnext = gt->gt_gprev = NULL;
}
}
stnp = &gt->gt_srctbl;
while ((st = *stnp) != NULL) {
if (ntohl(st->st_origin) >= ntohl(origin))
break;
stnp = &st->st_next;
}
if (st == NULL || st->st_origin != origin) {
st = (struct stable *)malloc(sizeof(struct stable));
if (st == NULL)
log(LOG_ERR, 0, "ran out of memory");
st->st_origin = origin;
st->st_pktcnt = 0;
st->st_savpkt = 0;
time(&st->st_ctime);
st->st_next = *stnp;
*stnp = st;
} else {
if (st->st_ctime == 0) {
/* An old source which we're keeping around for statistics */
time(&st->st_ctime);
} else {
#ifdef DEBUG_MFC
md_log(MD_DUPE, origin, mcastgrp);
#endif
/* Ignore kernel->mrouted retransmissions */
if (time(0) - st->st_ctime > 5)
log(LOG_WARNING, 0, "kernel entry already exists for (%s %s)",
inet_fmt(origin, s1), inet_fmt(mcastgrp, s2));
k_add_rg(origin, gt);
return;
}
}
kroutes++;
k_add_rg(origin, gt);
IF_DEBUG(DEBUG_CACHE)
log(LOG_DEBUG, 0, "add cache entry (%s %s) gm:%x, parent-vif:%d",
inet_fmt(origin, s1),
inet_fmt(mcastgrp, s2),
gt->gt_grpmems, r ? r->rt_parent : -1);
/*
* If there are no downstream routers that want traffic for
* this group, send (or retransmit) a prune upstream.
*/
if (VIFM_ISEMPTY(gt->gt_grpmems))
send_prune(gt);
}
/*
* A router has gone down. Remove prune state pertinent to that router.
*/
void
reset_neighbor_state(vifi, addr)
vifi_t vifi;
u_int32 addr;
{
struct rtentry *r;
struct gtable *g;
struct ptable *pt, **ptnp;
struct stable *st;
for (g = kernel_table; g; g = g->gt_gnext) {
r = g->gt_route;
/*
* If neighbor was the parent, remove the prune sent state
* and all of the source cache info so that prunes get
* regenerated.
*/
if (vifi == r->rt_parent) {
if (addr == r->rt_gateway) {
IF_DEBUG(DEBUG_PEER)
log(LOG_DEBUG, 0, "reset_neighbor_state parent reset (%s %s)",
RT_FMT(r, s1), inet_fmt(g->gt_mcastgrp, s2));
g->gt_prsent_timer = 0;
g->gt_grftsnt = 0;
while ((st = g->gt_srctbl) != NULL) {
g->gt_srctbl = st->st_next;
if (st->st_ctime != 0) {
k_del_rg(st->st_origin, g);
kroutes--;
}
free(st);
}
}
} else {
/*
* Remove any prunes that this router has sent us.
*/
ptnp = &g->gt_pruntbl;
while ((pt = *ptnp) != NULL) {
if (pt->pt_vifi == vifi && pt->pt_router == addr) {
NBRM_CLR(pt->pt_index, g->gt_prunes);
*ptnp = pt->pt_next;
free(pt);
} else
ptnp = &pt->pt_next;
}
/*
* And see if we want to forward again.
*/
if (!VIFM_ISSET(vifi, g->gt_grpmems)) {
GET_MEMBERSHIP(g, vifi);
APPLY_SCOPE(g);
prun_add_ttls(g);
/* Update kernel state */
update_kernel(g);
#ifdef RSRR
/* Send route change notification to reservation protocol. */
rsrr_cache_send(g,1);
#endif /* RSRR */
/*
* If removing this prune causes us to start forwarding
* (e.g. the neighbor rebooted), and we sent a prune upstream,
* send a graft to cancel the prune.
*/
if (!VIFM_ISEMPTY(g->gt_grpmems) && g->gt_prsent_timer)
send_graft(g);
IF_DEBUG(DEBUG_PEER)
log(LOG_DEBUG, 0, "reset neighbor state (%s %s) gm:%x",
RT_FMT(r, s1),
inet_fmt(g->gt_mcastgrp, s2), g->gt_grpmems);
}
}
}
}
/*
* Delete table entry from the kernel
* del_flag determines how many entries to delete
*/
void
del_table_entry(r, mcastgrp, del_flag)
struct rtentry *r;
u_int32 mcastgrp;
u_int del_flag;
{
struct gtable *g, *prev_g;
struct stable *st, *prev_st;
struct ptable *pt, *prev_pt;
if (del_flag == DEL_ALL_ROUTES) {
g = r->rt_groups;
while (g) {
IF_DEBUG(DEBUG_CACHE)
log(LOG_DEBUG, 0, "del_table_entry deleting (%s %s)",
RT_FMT(r, s1), inet_fmt(g->gt_mcastgrp, s2));
st = g->gt_srctbl;
while (st) {
if (st->st_ctime != 0) {
if (k_del_rg(st->st_origin, g) < 0) {
log(LOG_WARNING, errno,
"del_table_entry trying to delete (%s, %s)",
inet_fmt(st->st_origin, s1),
inet_fmt(g->gt_mcastgrp, s2));
}
kroutes--;
}
prev_st = st;
st = st->st_next;
free(prev_st);
}
g->gt_srctbl = NULL;
pt = g->gt_pruntbl;
while (pt) {
prev_pt = pt;
pt = pt->pt_next;
free(prev_pt);
}
g->gt_pruntbl = NULL;
if (g->gt_gnext)
g->gt_gnext->gt_gprev = g->gt_gprev;
if (g->gt_gprev)
g->gt_gprev->gt_gnext = g->gt_gnext;
else
kernel_table = g->gt_gnext;
#ifdef RSRR
/* Send route change notification to reservation protocol. */
rsrr_cache_send(g,0);
rsrr_cache_clean(g);
#endif /* RSRR */
if (g->gt_rexmit_timer)
timer_clearTimer(g->gt_rexmit_timer);
prev_g = g;
g = g->gt_next;
free(prev_g);
}
r->rt_groups = NULL;
}
/*
* Dummy routine - someday this may be needed, so it is just there
*/
if (del_flag == DEL_RTE_GROUP) {
prev_g = (struct gtable *)&r->rt_groups;
for (g = r->rt_groups; g; g = g->gt_next) {
if (g->gt_mcastgrp == mcastgrp) {
IF_DEBUG(DEBUG_CACHE)
log(LOG_DEBUG, 0, "del_table_entry deleting (%s %s)",
RT_FMT(r, s1), inet_fmt(g->gt_mcastgrp, s2));
st = g->gt_srctbl;
while (st) {
if (st->st_ctime != 0) {
if (k_del_rg(st->st_origin, g) < 0) {
log(LOG_WARNING, errno,
"del_table_entry trying to delete (%s, %s)",
inet_fmt(st->st_origin, s1),
inet_fmt(g->gt_mcastgrp, s2));
}
kroutes--;
}
prev_st = st;
st = st->st_next;
free(prev_st);
}
g->gt_srctbl = NULL;
pt = g->gt_pruntbl;
while (pt) {
prev_pt = pt;
pt = pt->pt_next;
free(prev_pt);
}
g->gt_pruntbl = NULL;
if (g->gt_gnext)
g->gt_gnext->gt_gprev = g->gt_gprev;
if (g->gt_gprev)
g->gt_gprev->gt_gnext = g->gt_gnext;
else
kernel_table = g->gt_gnext;
if (prev_g != (struct gtable *)&r->rt_groups)
g->gt_next->gt_prev = prev_g;
else
g->gt_next->gt_prev = NULL;
prev_g->gt_next = g->gt_next;
if (g->gt_rexmit_timer)
timer_clearTimer(g->gt_rexmit_timer);
#ifdef RSRR
/* Send route change notification to reservation protocol. */
rsrr_cache_send(g,0);
rsrr_cache_clean(g);
#endif /* RSRR */
free(g);
g = prev_g;
} else {
prev_g = g;
}
}
}
}
/*
* update kernel table entry when a route entry changes
*/
void
update_table_entry(r, old_parent_gw)
struct rtentry *r;
u_int32 old_parent_gw;
{
struct gtable *g;
struct ptable *pt, **ptnp;
for (g = r->rt_groups; g; g = g->gt_next) {
ptnp = &g->gt_pruntbl;
/*
* Delete prune entries from non-children, or non-subordinates.
*/
while ((pt = *ptnp)) {
if (!VIFM_ISSET(pt->pt_vifi, r->rt_children) ||
!NBRM_ISSET(pt->pt_index, r->rt_subordinates)) {
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "update_table_entry deleting prune for (%s %s) from %s on vif %d -%s%s",
RT_FMT(r, s1), inet_fmt(g->gt_mcastgrp, s2),
inet_fmt(pt->pt_router, s3), pt->pt_vifi,
VIFM_ISSET(pt->pt_vifi, r->rt_children) ? "" : " not a child",
NBRM_ISSET(pt->pt_index, r->rt_subordinates) ? "" : " not a subordinate");
if (!NBRM_ISSET(pt->pt_index, g->gt_prunes)) {
log(LOG_WARNING, 0,
"gt_prunes lost track of (%s %s) from %s on vif %d",
RT_FMT(r, s1), inet_fmt(g->gt_mcastgrp, s2),
inet_fmt(pt->pt_router, s3), pt->pt_vifi);
}
NBRM_CLR(pt->pt_index, g->gt_prunes);
*ptnp = pt->pt_next;
free(pt);
continue;
}
ptnp = &((*ptnp)->pt_next);
}
IF_DEBUG(DEBUG_CACHE)
log(LOG_DEBUG, 0, "updating cache entries (%s %s) old gm:%x",
RT_FMT(r, s1), inet_fmt(g->gt_mcastgrp, s2),
g->gt_grpmems);
/*
* Forget about a prune or graft that we sent previously if we
* have a new parent router (since the new parent router will
* know nothing about what I sent to the previous parent). The
* old parent will forget any prune state it is keeping for us.
*/
if (old_parent_gw != r->rt_gateway) {
g->gt_prsent_timer = 0;
g->gt_grftsnt = 0;
}
/* Recalculate membership */
determine_forwvifs(g);
/* send a prune or graft if needed. */
send_prune_or_graft(g);
IF_DEBUG(DEBUG_CACHE)
log(LOG_DEBUG, 0, "updating cache entries (%s %s) new gm:%x",
RT_FMT(r, s1), inet_fmt(g->gt_mcastgrp, s2),
g->gt_grpmems);
/* update ttls and add entry into kernel */
prun_add_ttls(g);
update_kernel(g);
#ifdef RSRR
/* Send route change notification to reservation protocol. */
rsrr_cache_send(g,1);
#endif /* RSRR */
}
}
/*
* set the forwarding flag for all mcastgrps on this vifi
*/
void
update_lclgrp(vifi, mcastgrp)
vifi_t vifi;
u_int32 mcastgrp;
{
struct rtentry *r;
struct gtable *g;
IF_DEBUG(DEBUG_MEMBER)
log(LOG_DEBUG, 0, "group %s joined on vif %d",
inet_fmt(mcastgrp, s1), vifi);
for (g = kernel_table; g; g = g->gt_gnext) {
if (ntohl(mcastgrp) < ntohl(g->gt_mcastgrp))
break;
r = g->gt_route;
if (g->gt_mcastgrp == mcastgrp &&
VIFM_ISSET(vifi, r->rt_children)) {
VIFM_SET(vifi, g->gt_grpmems);
APPLY_SCOPE(g);
if (VIFM_ISEMPTY(g->gt_grpmems))
continue;
prun_add_ttls(g);
IF_DEBUG(DEBUG_CACHE)
log(LOG_DEBUG, 0, "update lclgrp (%s %s) gm:%x",
RT_FMT(r, s1),
inet_fmt(g->gt_mcastgrp, s2), g->gt_grpmems);
update_kernel(g);
#ifdef RSRR
/* Send route change notification to reservation protocol. */
rsrr_cache_send(g,1);
#endif /* RSRR */
}
}
}
/*
* reset forwarding flag for all mcastgrps on this vifi
*/
void
delete_lclgrp(vifi, mcastgrp)
vifi_t vifi;
u_int32 mcastgrp;
{
struct gtable *g;
IF_DEBUG(DEBUG_MEMBER)
log(LOG_DEBUG, 0, "group %s left on vif %d",
inet_fmt(mcastgrp, s1), vifi);
for (g = kernel_table; g; g = g->gt_gnext) {
if (ntohl(mcastgrp) < ntohl(g->gt_mcastgrp))
break;
if (g->gt_mcastgrp == mcastgrp && VIFM_ISSET(vifi, g->gt_grpmems)) {
if (g->gt_route == NULL ||
SUBS_ARE_PRUNED(g->gt_route->rt_subordinates,
uvifs[vifi].uv_nbrmap, g->gt_prunes)) {
VIFM_CLR(vifi, g->gt_grpmems);
IF_DEBUG(DEBUG_CACHE)
log(LOG_DEBUG, 0, "delete lclgrp (%s %s) gm:%x",
RT_FMT(g->gt_route, s1),
inet_fmt(g->gt_mcastgrp, s2), g->gt_grpmems);
prun_add_ttls(g);
update_kernel(g);
#ifdef RSRR
/* Send route change notification to reservation protocol. */
rsrr_cache_send(g,1);
#endif /* RSRR */
/*
* If there are no more members of this particular group,
* send prune upstream
*/
if (VIFM_ISEMPTY(g->gt_grpmems) && g->gt_route->rt_gateway)
send_prune(g);
}
}
}
}
/*
* Takes the prune message received and then strips it to
* determine the (src, grp) pair to be pruned.
*
* Adds the router to the (src, grp) entry then.
*
* Determines if further packets have to be sent down that vif
*
* Determines if a corresponding prune message has to be generated
*/
void
accept_prune(src, dst, p, datalen)
u_int32 src;
u_int32 dst;
char *p;
int datalen;
{
u_int32 prun_src;
u_int32 prun_grp;
u_int32 prun_tmr;
vifi_t vifi;
int i;
struct rtentry *r;
struct gtable *g;
struct ptable *pt;
if ((vifi = find_vif(src, dst)) == NO_VIF) {
log(LOG_INFO, 0,
"ignoring prune report from non-neighbor %s",
inet_fmt(src, s1));
return;
}
/* Check if enough data is present */
if (datalen < 12)
{
log(LOG_WARNING, 0,
"non-decipherable prune from %s",
inet_fmt(src, s1));
return;
}
for (i = 0; i< 4; i++)
((char *)&prun_src)[i] = *p++;
for (i = 0; i< 4; i++)
((char *)&prun_grp)[i] = *p++;
for (i = 0; i< 4; i++)
((char *)&prun_tmr)[i] = *p++;
prun_tmr = ntohl(prun_tmr);
if (prun_tmr <= MIN_PRUNE_LIFE) {
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "ignoring prune from %s on vif %d for (%s %s)/%d because its lifetime is too short",
inet_fmt(src, s1), vifi,
inet_fmt(prun_src, s2), inet_fmt(prun_grp, s3), prun_tmr);
return;
}
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "%s on vif %d prunes (%s %s)/%d",
inet_fmt(src, s1), vifi,
inet_fmt(prun_src, s2), inet_fmt(prun_grp, s3), prun_tmr);
/*
* Find the subnet for the prune
*/
if (find_src_grp(prun_src, 0, prun_grp)) {
g = gtp ? gtp->gt_gnext : kernel_table;
r = g->gt_route;
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "found grp state, (%s %s), metric is %d, children are %x, subords are %08x%08x",
RT_FMT(r, s1), inet_fmt(g->gt_mcastgrp, s2), r->rt_metric,
r->rt_children, r->rt_subordinates.hi, r->rt_subordinates.lo);
if (!VIFM_ISSET(vifi, r->rt_children)) {
IF_DEBUG(DEBUG_PRUNE)
log(LOG_WARNING, 0, "prune received from non-child %s for (%s %s) (dominant on vif %d is %s)",
inet_fmt(src, s1), inet_fmt(prun_src, s2),
inet_fmt(prun_grp, s3), vifi,
inet_fmt(r->rt_dominants[vifi], s4));
#ifdef RINGBUFFER
printringbuf();
#endif
return;
}
if (VIFM_ISSET(vifi, g->gt_scope)) {
log(LOG_WARNING, 0, "prune received from %s on scoped grp (%s %s)",
inet_fmt(src, s1), inet_fmt(prun_src, s2),
inet_fmt(prun_grp, s3));
return;
}
if ((pt = find_prune_entry(src, g->gt_pruntbl)) != NULL) {
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "%s %d from %s for (%s %s)/%d %s %d %s %x",
"duplicate prune received on vif",
vifi, inet_fmt(src, s1), inet_fmt(prun_src, s2),
inet_fmt(prun_grp, s3), prun_tmr,
"old timer:", pt->pt_timer, "cur gm:", g->gt_grpmems);
pt->pt_timer = prun_tmr;
} else {
struct listaddr *n = neighbor_info(vifi, src);
if (!n) {
log(LOG_WARNING, 0, "Prune from non-neighbor %s on vif %d!?",
inet_fmt(src, s1), vifi);
return;
}
/* allocate space for the prune structure */
pt = (struct ptable *)(malloc(sizeof(struct ptable)));
if (pt == NULL)
log(LOG_ERR, 0, "pt: ran out of memory");
pt->pt_vifi = vifi;
pt->pt_router = src;
pt->pt_timer = prun_tmr;
pt->pt_next = g->gt_pruntbl;
g->gt_pruntbl = pt;
if (n) {
pt->pt_index = n->al_index;
NBRM_SET(n->al_index, g->gt_prunes);
}
}
/*
* check if any more packets need to be sent on the
* vif which sent this message
*/
if (SUBS_ARE_PRUNED(r->rt_subordinates,
uvifs[vifi].uv_nbrmap, g->gt_prunes) &&
!grplst_mem(vifi, prun_grp)) {
nbrbitmap_t tmp;
VIFM_CLR(vifi, g->gt_grpmems);
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "vifnbrs=0x%08x%08x, subord=0x%08x%08x prunes=0x%08x%08x",
uvifs[vifi].uv_nbrmap.hi,uvifs[vifi].uv_nbrmap.lo,
r->rt_subordinates.hi, r->rt_subordinates.lo,
g->gt_prunes.hi, g->gt_prunes.lo);
/* XXX debugging */
NBRM_COPY(r->rt_subordinates, tmp);
NBRM_MASK(tmp, uvifs[vifi].uv_nbrmap);
if (!NBRM_ISSETALLMASK(g->gt_prunes, tmp))
log(LOG_WARNING, 0, "subordinate error");
/* XXX end debugging */
IF_DEBUG(DEBUG_PRUNE|DEBUG_CACHE)
log(LOG_DEBUG, 0, "prune (%s %s), stop sending on vif %d, gm:%x",
RT_FMT(r, s1),
inet_fmt(g->gt_mcastgrp, s2), vifi, g->gt_grpmems);
prun_add_ttls(g);
update_kernel(g);
#ifdef RSRR
/* Send route change notification to reservation protocol. */
rsrr_cache_send(g,1);
#endif /* RSRR */
}
/*
* check if all the child routers have expressed no interest
* in this group and if this group does not exist in the
* interface
* Send a prune message then upstream
*/
if (VIFM_ISEMPTY(g->gt_grpmems) && r->rt_gateway) {
send_prune(g);
}
} else {
/*
* There is no kernel entry for this group. Therefore, we can
* simply ignore the prune, as we are not forwarding this traffic
* downstream.
*/
IF_DEBUG(DEBUG_PRUNE|DEBUG_CACHE)
log(LOG_DEBUG, 0, "%s (%s %s)/%d from %s",
"prune message received with no kernel entry for",
inet_fmt(prun_src, s1), inet_fmt(prun_grp, s2),
prun_tmr, inet_fmt(src, s3));
return;
}
}
/*
* Checks if this mcastgrp is present in the kernel table
* If so and if a prune was sent, it sends a graft upwards
*/
void
chkgrp_graft(vifi, mcastgrp)
vifi_t vifi;
u_int32 mcastgrp;
{
struct rtentry *r;
struct gtable *g;
for (g = kernel_table; g; g = g->gt_gnext) {
if (ntohl(mcastgrp) < ntohl(g->gt_mcastgrp))
break;
r = g->gt_route;
if (g->gt_mcastgrp == mcastgrp && VIFM_ISSET(vifi, r->rt_children))
if (g->gt_prsent_timer) {
VIFM_SET(vifi, g->gt_grpmems);
/*
* If the vif that was joined was a scoped vif,
* ignore it ; don't graft back
*/
APPLY_SCOPE(g);
if (VIFM_ISEMPTY(g->gt_grpmems))
continue;
/* send graft upwards */
send_graft(g);
/* update cache timer*/
g->gt_timer = CACHE_LIFETIME(cache_lifetime);
IF_DEBUG(DEBUG_PRUNE|DEBUG_CACHE)
log(LOG_DEBUG, 0, "chkgrp graft (%s %s) gm:%x",
RT_FMT(r, s1),
inet_fmt(g->gt_mcastgrp, s2), g->gt_grpmems);
prun_add_ttls(g);
update_kernel(g);
#ifdef RSRR
/* Send route change notification to reservation protocol. */
rsrr_cache_send(g,1);
#endif /* RSRR */
}
}
}
/* determine the multicast group and src
*
* if it does, then determine if a prune was sent
* upstream.
* if prune sent upstream, send graft upstream and send
* ack downstream.
*
* if no prune sent upstream, change the forwarding bit
* for this interface and send ack downstream.
*
* if no entry exists for this group send ack downstream.
*/
void
accept_graft(src, dst, p, datalen)
u_int32 src;
u_int32 dst;
char *p;
int datalen;
{
vifi_t vifi;
u_int32 graft_src;
u_int32 graft_grp;
int i;
struct rtentry *r;
struct gtable *g;
struct ptable *pt, **ptnp;
if (datalen < 8) {
log(LOG_WARNING, 0,
"received non-decipherable graft from %s",
inet_fmt(src, s1));
return;
}
for (i = 0; i< 4; i++)
((char *)&graft_src)[i] = *p++;
for (i = 0; i< 4; i++)
((char *)&graft_grp)[i] = *p++;
vifi = find_vif(src, dst);
send_graft_ack(dst, src, graft_src, graft_grp, vifi);
if (vifi == NO_VIF) {
log(LOG_INFO, 0,
"ignoring graft for (%s %s) from non-neighbor %s",
inet_fmt(graft_src, s2), inet_fmt(graft_grp, s3),
inet_fmt(src, s1));
return;
}
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "%s on vif %d grafts (%s %s)",
inet_fmt(src, s1), vifi,
inet_fmt(graft_src, s2), inet_fmt(graft_grp, s3));
/*
* Find the subnet for the graft
*/
if (find_src_grp(graft_src, 0, graft_grp)) {
g = gtp ? gtp->gt_gnext : kernel_table;
r = g->gt_route;
if (VIFM_ISSET(vifi, g->gt_scope)) {
log(LOG_WARNING, 0, "graft received from %s on scoped grp (%s %s)",
inet_fmt(src, s1), inet_fmt(graft_src, s2),
inet_fmt(graft_grp, s3));
return;
}
ptnp = &g->gt_pruntbl;
while ((pt = *ptnp) != NULL) {
if ((pt->pt_vifi == vifi) && (pt->pt_router == src)) {
NBRM_CLR(pt->pt_index, g->gt_prunes);
*ptnp = pt->pt_next;
free(pt);
VIFM_SET(vifi, g->gt_grpmems);
IF_DEBUG(DEBUG_PRUNE|DEBUG_CACHE)
log(LOG_DEBUG, 0, "accept graft (%s %s) gm:%x",
RT_FMT(r, s1),
inet_fmt(g->gt_mcastgrp, s2), g->gt_grpmems);
prun_add_ttls(g);
update_kernel(g);
#ifdef RSRR
/* Send route change notification to reservation protocol. */
rsrr_cache_send(g,1);
#endif /* RSRR */
break;
} else {
ptnp = &pt->pt_next;
}
}
g->gt_timer = CACHE_LIFETIME(cache_lifetime);
if (g->gt_prsent_timer)
/* send graft upwards */
send_graft(g);
} else {
/*
* We have no state for the source and group in question.
* This is fine, since we know that we have no prune state, and
* grafts are requests to remove prune state.
*/
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "%s (%s %s) from %s",
"graft received with no kernel entry for",
inet_fmt(graft_src, s1), inet_fmt(graft_grp, s2),
inet_fmt(src, s3));
return;
}
}
/*
* find out which group is involved first of all
* then determine if a graft was sent.
* if no graft sent, ignore the message
* if graft was sent and the ack is from the right
* source, remove the graft timer so that we don't
* have send a graft again
*/
void
accept_g_ack(src, dst, p, datalen)
u_int32 src;
u_int32 dst;
char *p;
int datalen;
{
struct gtable *g;
vifi_t vifi;
u_int32 grft_src;
u_int32 grft_grp;
int i;
if ((vifi = find_vif(src, dst)) == NO_VIF) {
log(LOG_INFO, 0,
"ignoring graft ack from non-neighbor %s",
inet_fmt(src, s1));
return;
}
if (datalen < 0 || datalen > 8) {
log(LOG_WARNING, 0,
"received non-decipherable graft ack from %s",
inet_fmt(src, s1));
return;
}
for (i = 0; i< 4; i++)
((char *)&grft_src)[i] = *p++;
for (i = 0; i< 4; i++)
((char *)&grft_grp)[i] = *p++;
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "%s on vif %d acks graft (%s, %s)",
inet_fmt(src, s1), vifi,
inet_fmt(grft_src, s2), inet_fmt(grft_grp, s3));
/*
* Find the subnet for the graft ack
*/
if (find_src_grp(grft_src, 0, grft_grp)) {
g = gtp ? gtp->gt_gnext : kernel_table;
g->gt_grftsnt = 0;
} else {
log(LOG_WARNING, 0, "%s (%s, %s) from %s",
"rcvd graft ack with no kernel entry for",
inet_fmt(grft_src, s1), inet_fmt(grft_grp, s2),
inet_fmt(src, s3));
#ifdef RINGBUFFER
printringbuf();
#endif
return;
}
}
/*
* free all prune entries and kernel routes
* normally, this should inform the kernel that all of its routes
* are going away, but this is only called by restart(), which is
* about to call MRT_DONE which does that anyway.
*/
void
free_all_prunes()
{
register struct rtentry *r;
register struct gtable *g, *prev_g;
register struct stable *s, *prev_s;
register struct ptable *p, *prev_p;
for (r = routing_table; r; r = r->rt_next) {
g = r->rt_groups;
while (g) {
s = g->gt_srctbl;
while (s) {
prev_s = s;
s = s->st_next;
free(prev_s);
}
p = g->gt_pruntbl;
while (p) {
prev_p = p;
p = p->pt_next;
free(prev_p);
}
prev_g = g;
g = g->gt_next;
if (prev_g->gt_rexmit_timer)
timer_clearTimer(prev_g->gt_rexmit_timer);
free(prev_g);
}
r->rt_groups = NULL;
}
kernel_table = NULL;
g = kernel_no_route;
while (g) {
if (g->gt_srctbl)
free(g->gt_srctbl);
prev_g = g;
g = g->gt_next;
if (prev_g->gt_rexmit_timer)
timer_clearTimer(prev_g->gt_rexmit_timer);
free(prev_g);
}
kernel_no_route = NULL;
}
/*
* When a new route is created, search
* a) The less-specific part of the routing table
* b) The route-less kernel table
* for sources that the new route might want to handle.
*
* "Inheriting" these sources might be cleanest, but simply deleting
* them is easier, and letting the kernel re-request them.
*/
void
steal_sources(rt)
struct rtentry *rt;
{
register struct rtentry *rp;
register struct gtable *gt, **gtnp;
register struct stable *st, **stnp;
for (rp = rt->rt_next; rp; rp = rp->rt_next) {
if (rp->rt_groups == NULL)
continue;
if ((rt->rt_origin & rp->rt_originmask) == rp->rt_origin) {
IF_DEBUG(DEBUG_ROUTE)
log(LOG_DEBUG, 0, "Route for %s stealing sources from %s",
RT_FMT(rt, s1), RT_FMT(rp, s2));
for (gt = rp->rt_groups; gt; gt = gt->gt_next) {
stnp = &gt->gt_srctbl;
while ((st = *stnp) != NULL) {
if ((st->st_origin & rt->rt_originmask) == rt->rt_origin) {
IF_DEBUG(DEBUG_ROUTE)
log(LOG_DEBUG, 0, "%s stealing (%s %s) from %s",
RT_FMT(rt, s1),
inet_fmt(st->st_origin, s3),
inet_fmt(gt->gt_mcastgrp, s4),
RT_FMT(rp, s2));
if (st->st_ctime != 0) {
if (k_del_rg(st->st_origin, gt) < 0) {
log(LOG_WARNING, errno, "%s (%s, %s)",
"steal_sources trying to delete",
inet_fmt(st->st_origin, s1),
inet_fmt(gt->gt_mcastgrp, s2));
}
kroutes--;
}
*stnp = st->st_next;
free(st);
} else {
stnp = &st->st_next;
}
}
}
}
}
gtnp = &kernel_no_route;
while ((gt = *gtnp) != NULL) {
if (gt->gt_srctbl && ((gt->gt_srctbl->st_origin & rt->rt_originmask)
== rt->rt_origin)) {
IF_DEBUG(DEBUG_ROUTE)
log(LOG_DEBUG, 0, "%s stealing (%s %s) from %s",
RT_FMT(rt, s1),
inet_fmt(gt->gt_srctbl->st_origin, s3),
inet_fmt(gt->gt_mcastgrp, s4),
"no_route table");
if (gt->gt_srctbl->st_ctime != 0) {
if (k_del_rg(gt->gt_srctbl->st_origin, gt) < 0) {
log(LOG_WARNING, errno, "%s (%s %s)",
"steal_sources trying to delete",
inet_fmt(gt->gt_srctbl->st_origin, s1),
inet_fmt(gt->gt_mcastgrp, s2));
}
kroutes--;
}
free(gt->gt_srctbl);
*gtnp = gt->gt_next;
if (gt->gt_next)
gt->gt_next->gt_prev = gt->gt_prev;
if (gt->gt_rexmit_timer)
timer_clearTimer(gt->gt_rexmit_timer);
free(gt);
} else {
gtnp = &gt->gt_next;
}
}
}
/*
* Advance the timers on all the cache entries.
* If there are any entries whose timers have expired,
* remove these entries from the kernel cache.
*/
void
age_table_entry()
{
struct rtentry *r;
struct gtable *gt, **gtnptr;
struct stable *st, **stnp;
struct ptable *pt, **ptnp;
struct sioc_sg_req sg_req;
IF_DEBUG(DEBUG_PRUNE|DEBUG_CACHE)
log(LOG_DEBUG, 0, "aging forwarding cache entries");
gtnptr = &kernel_table;
while ((gt = *gtnptr) != NULL) {
vifi_t i; /* XXX Debugging */
int fixit = 0; /* XXX Debugging */
r = gt->gt_route;
/* XXX Debugging... */
for (i = 0; i < numvifs; i++) {
/*
* If we're not sending on this vif,
* And this group isn't scoped on this vif,
* And I'm the parent for this route on this vif,
* And there are subordinates on this vif,
* And all of the subordinates haven't pruned,
* YELL LOUDLY
* and remember to fix it up later
*/
if (!VIFM_ISSET(i, gt->gt_grpmems) &&
!VIFM_ISSET(i, gt->gt_scope) &&
VIFM_ISSET(i, r->rt_children) &&
NBRM_ISSETMASK(uvifs[i].uv_nbrmap, r->rt_subordinates) &&
!SUBS_ARE_PRUNED(r->rt_subordinates, uvifs[i].uv_nbrmap, gt->gt_prunes)) {
log(LOG_WARNING, 0, "(%s %s) is blackholing on vif %d",
RT_FMT(r, s1), inet_fmt(gt->gt_mcastgrp, s2), i);
fixit = 1;
}
}
if (fixit) {
log(LOG_WARNING, 0, "fixing membership for (%s %s) gm:%x",
RT_FMT(r, s1), inet_fmt(gt->gt_mcastgrp, s2), gt->gt_grpmems);
determine_forwvifs(gt);
send_prune_or_graft(gt);
log(LOG_WARNING, 0, "fixed membership for (%s %s) gm:%x",
RT_FMT(r, s1), inet_fmt(gt->gt_mcastgrp, s2), gt->gt_grpmems);
#ifdef RINGBUFFER
printringbuf();
#endif
}
/*DEBUG2*/
/* If there are group members,
* and there are recent sources,
* and we have a route,
* and it's not directly connected,
* and we haven't sent a prune,
* if there are any cache entries in the kernel
* [if there aren't we're probably waiting to rexmit],
* YELL LOUDLY
* and send a prune
*/
if (VIFM_ISEMPTY(gt->gt_grpmems) && gt->gt_srctbl && r && r->rt_gateway && gt->gt_prsent_timer == 0) {
for (st = gt->gt_srctbl; st; st = st->st_next)
if (st->st_ctime != 0)
break;
if (st != NULL) {
log(LOG_WARNING, 0, "grpmems for (%s %s) is empty but no prune state!", RT_FMT(r, s1), inet_fmt(gt->gt_mcastgrp, s2));
send_prune_or_graft(gt);
#ifdef RINGBUFFER
printringbuf();
#endif
}
}
/* XXX ...Debugging */
/* advance the timer for the kernel entry */
gt->gt_timer -= TIMER_INTERVAL;
/* decrement prune timer if need be */
if (gt->gt_prsent_timer > 0) {
gt->gt_prsent_timer -= TIMER_INTERVAL;
if (gt->gt_prsent_timer <= 0) {
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "upstream prune tmo (%s %s)",
RT_FMT(r, s1),
inet_fmt(gt->gt_mcastgrp, s2));
gt->gt_prsent_timer = -1;
/* Reset the prune retransmission timer to its initial value */
gt->gt_prune_rexmit = PRUNE_REXMIT_VAL;
}
}
/* retransmit graft with exponential backoff */
if (gt->gt_grftsnt) {
register int y;
y = ++gt->gt_grftsnt;
while (y && !(y & 1))
y >>= 1;
if (y == 1)
send_graft(gt);
}
/*
* Age prunes
*
* If a prune expires, forward again on that vif.
*/
ptnp = &gt->gt_pruntbl;
while ((pt = *ptnp) != NULL) {
if ((pt->pt_timer -= TIMER_INTERVAL) <= 0) {
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "expire prune (%s %s) from %s on vif %d",
RT_FMT(r, s1),
inet_fmt(gt->gt_mcastgrp, s2),
inet_fmt(pt->pt_router, s3),
pt->pt_vifi);
if (gt->gt_prsent_timer > 0) {
log(LOG_WARNING, 0, "prune (%s %s) from %s on vif %d expires with %d left on prsent timer",
RT_FMT(r, s1),
inet_fmt(gt->gt_mcastgrp, s2),
inet_fmt(pt->pt_router, s3),
pt->pt_vifi, gt->gt_prsent_timer);
/* Send a graft to heal the tree. */
send_graft(gt);
}
NBRM_CLR(pt->pt_index, gt->gt_prunes);
expire_prune(pt->pt_vifi, gt);
/* remove the router's prune entry and await new one */
*ptnp = pt->pt_next;
free(pt);
} else {
ptnp = &pt->pt_next;
}
}
/*
* If the cache entry has expired, delete source table entries for
* silent sources. If there are no source entries left, and there
* are no downstream prunes, then the entry is deleted.
* Otherwise, the cache entry's timer is refreshed.
*/
if (gt->gt_timer <= 0) {
IF_DEBUG(DEBUG_CACHE)
log(LOG_DEBUG, 0, "(%s %s) timed out, checking for traffic",
RT_FMT(gt->gt_route, s1),
inet_fmt(gt->gt_mcastgrp, s2));
/* Check for traffic before deleting source entries */
sg_req.grp.s_addr = gt->gt_mcastgrp;
stnp = &gt->gt_srctbl;
while ((st = *stnp) != NULL) {
/*
* Source entries with no ctime are not actually in the
* kernel; they have been removed by rexmit_prune() so
* are safe to remove from the list at this point.
*/
if (st->st_ctime) {
sg_req.src.s_addr = st->st_origin;
if (ioctl(udp_socket, SIOCGETSGCNT, (char *)&sg_req) < 0) {
log(LOG_WARNING, errno, "%s (%s %s)",
"age_table_entry: SIOCGETSGCNT failing for",
inet_fmt(st->st_origin, s1),
inet_fmt(gt->gt_mcastgrp, s2));
/* Make sure it gets deleted below */
sg_req.pktcnt = st->st_pktcnt;
}
} else {
sg_req.pktcnt = st->st_pktcnt;
}
if (sg_req.pktcnt == st->st_pktcnt) {
*stnp = st->st_next;
IF_DEBUG(DEBUG_CACHE)
log(LOG_DEBUG, 0, "age_table_entry deleting (%s %s)",
inet_fmt(st->st_origin, s1),
inet_fmt(gt->gt_mcastgrp, s2));
if (st->st_ctime != 0) {
if (k_del_rg(st->st_origin, gt) < 0) {
log(LOG_WARNING, errno,
"age_table_entry trying to delete (%s %s)",
inet_fmt(st->st_origin, s1),
inet_fmt(gt->gt_mcastgrp, s2));
}
kroutes--;
}
free(st);
} else {
st->st_pktcnt = sg_req.pktcnt;
stnp = &st->st_next;
}
}
/*
* Retain the group entry if we have downstream prunes or if
* there is at least one source in the list that still has
* traffic, or if our upstream prune timer or graft
* retransmission timer is running.
*/
if (gt->gt_pruntbl != NULL || gt->gt_srctbl != NULL ||
gt->gt_prsent_timer > 0 || gt->gt_grftsnt > 0) {
IF_DEBUG(DEBUG_CACHE)
log(LOG_DEBUG, 0, "refresh lifetim of cache entry %s%s%s%s(%s, %s)",
gt->gt_pruntbl ? "(dstrm prunes) " : "",
gt->gt_srctbl ? "(trfc flow) " : "",
gt->gt_prsent_timer > 0 ? "(upstrm prune) " : "",
gt->gt_grftsnt > 0 ? "(grft rexmit) " : "",
RT_FMT(r, s1),
inet_fmt(gt->gt_mcastgrp, s2));
gt->gt_timer = CACHE_LIFETIME(cache_lifetime);
if (gt->gt_prsent_timer == -1) {
/*
* The upstream prune timed out. Remove any kernel
* state.
*/
gt->gt_prsent_timer = 0;
if (gt->gt_pruntbl) {
log(LOG_WARNING, 0, "upstream prune for (%s %s) expires with downstream prunes active",
RT_FMT(r, s1), inet_fmt(gt->gt_mcastgrp, s2));
}
remove_sources(gt);
}
gtnptr = &gt->gt_gnext;
continue;
}
IF_DEBUG(DEBUG_CACHE)
log(LOG_DEBUG, 0, "timeout cache entry (%s, %s)",
RT_FMT(r, s1),
inet_fmt(gt->gt_mcastgrp, s2));
if (gt->gt_prev)
gt->gt_prev->gt_next = gt->gt_next;
else
gt->gt_route->rt_groups = gt->gt_next;
if (gt->gt_next)
gt->gt_next->gt_prev = gt->gt_prev;
if (gt->gt_gprev) {
gt->gt_gprev->gt_gnext = gt->gt_gnext;
gtnptr = &gt->gt_gprev->gt_gnext;
} else {
kernel_table = gt->gt_gnext;
gtnptr = &kernel_table;
}
if (gt->gt_gnext)
gt->gt_gnext->gt_gprev = gt->gt_gprev;
#ifdef RSRR
/* Send route change notification to reservation protocol. */
rsrr_cache_send(gt,0);
rsrr_cache_clean(gt);
#endif /* RSRR */
if (gt->gt_rexmit_timer)
timer_clearTimer(gt->gt_rexmit_timer);
free((char *)gt);
} else {
if (gt->gt_prsent_timer == -1) {
/*
* The upstream prune timed out. Remove any kernel
* state.
*/
gt->gt_prsent_timer = 0;
if (gt->gt_pruntbl) {
log(LOG_WARNING, 0, "upstream prune for (%s %s) expires with downstream prunes active",
RT_FMT(r, s1), inet_fmt(gt->gt_mcastgrp, s2));
}
remove_sources(gt);
}
gtnptr = &gt->gt_gnext;
}
}
/*
* When traversing the no_route table, the decision is much easier.
* Just delete it if it has timed out.
*/
gtnptr = &kernel_no_route;
while ((gt = *gtnptr) != NULL) {
/* advance the timer for the kernel entry */
gt->gt_timer -= TIMER_INTERVAL;
if (gt->gt_timer < 0) {
if (gt->gt_srctbl) {
if (gt->gt_srctbl->st_ctime != 0) {
if (k_del_rg(gt->gt_srctbl->st_origin, gt) < 0) {
log(LOG_WARNING, errno, "%s (%s %s)",
"age_table_entry trying to delete no-route",
inet_fmt(gt->gt_srctbl->st_origin, s1),
inet_fmt(gt->gt_mcastgrp, s2));
}
kroutes--;
}
free(gt->gt_srctbl);
}
*gtnptr = gt->gt_next;
if (gt->gt_next)
gt->gt_next->gt_prev = gt->gt_prev;
if (gt->gt_rexmit_timer)
timer_clearTimer(gt->gt_rexmit_timer);
free((char *)gt);
} else {
gtnptr = &gt->gt_next;
}
}
}
/*
* Modify the kernel to forward packets when one or multiple prunes that
* were received on the vif given by vifi, for the group given by gt,
* have expired.
*/
static void
expire_prune(vifi, gt)
vifi_t vifi;
struct gtable *gt;
{
/*
* No need to send a graft, any prunes that we sent
* will expire before any prunes that we have received.
* However, in the case that we did make a mistake,
* send a graft to compensate.
*/
if (gt->gt_prsent_timer >= MIN_PRUNE_LIFE) {
IF_DEBUG(DEBUG_PRUNE)
log(LOG_DEBUG, 0, "prune expired with %d left on %s",
gt->gt_prsent_timer, "prsent_timer");
gt->gt_prsent_timer = 0;
send_graft(gt);
}
/* modify the kernel entry to forward packets */
if (!VIFM_ISSET(vifi, gt->gt_grpmems)) {
struct rtentry *rt = gt->gt_route;
VIFM_SET(vifi, gt->gt_grpmems);
IF_DEBUG(DEBUG_CACHE)
log(LOG_DEBUG, 0, "forw again (%s %s) gm:%x vif:%d",
RT_FMT(rt, s1),
inet_fmt(gt->gt_mcastgrp, s2), gt->gt_grpmems, vifi);
prun_add_ttls(gt);
update_kernel(gt);
#ifdef RSRR
/* Send route change notification to reservation protocol. */
rsrr_cache_send(gt,1);
#endif /* RSRR */
}
}
/*
* Print the contents of the cache table on file 'fp2'.
*/
void
dump_cache(fp2)
FILE *fp2;
{
register struct rtentry *r;
register struct gtable *gt;
register struct stable *st;
register struct ptable *pt;
register vifi_t i;
char c;
register time_t thyme = time(0);
fprintf(fp2,
"Multicast Routing Cache Table (%d entries)\n%s", kroutes,
" Origin Mcast-group CTmr Age Ptmr Rx IVif Forwvifs\n");
fprintf(fp2,
"<(prunesrc:vif[idx]/tmr) prunebitmap\n%s",
">Source Lifetime SavPkt Pkts Bytes RPFf\n");
for (gt = kernel_no_route; gt; gt = gt->gt_next) {
if (gt->gt_srctbl) {
fprintf(fp2, " %-18s %-15s %-8s %-8s - -1 (no route)\n",
inet_fmts(gt->gt_srctbl->st_origin, 0xffffffff, s1),
inet_fmt(gt->gt_mcastgrp, s2), scaletime(gt->gt_timer),
scaletime(thyme - gt->gt_ctime));
fprintf(fp2, ">%s\n", inet_fmt(gt->gt_srctbl->st_origin, s1));
}
}
for (gt = kernel_table; gt; gt = gt->gt_gnext) {
r = gt->gt_route;
fprintf(fp2, " %-18s %-15s",
RT_FMT(r, s1),
inet_fmt(gt->gt_mcastgrp, s2));
fprintf(fp2, " %-8s", scaletime(gt->gt_timer));
fprintf(fp2, " %-8s %-8s ", scaletime(thyme - gt->gt_ctime),
gt->gt_prsent_timer ? scaletime(gt->gt_prsent_timer) :
" -");
if (gt->gt_prune_rexmit) {
int i = gt->gt_prune_rexmit;
int n = 0;
while (i > PRUNE_REXMIT_VAL) {
n++;
i /= 2;
}
if (n == 0 && gt->gt_prsent_timer == 0)
fprintf(fp2, " -");
else
fprintf(fp2, "%2d", n);
} else {
fprintf(fp2, " -");
}
fprintf(fp2, " %2u%c%c", r->rt_parent,
gt->gt_prsent_timer ? 'P' :
gt->gt_grftsnt ? 'G' : ' ',
VIFM_ISSET(r->rt_parent, gt->gt_scope) ? 'B' : ' ');
for (i = 0; i < numvifs; ++i) {
if (VIFM_ISSET(i, gt->gt_grpmems))
fprintf(fp2, " %u ", i);
else if (VIFM_ISSET(i, r->rt_children) &&
NBRM_ISSETMASK(uvifs[i].uv_nbrmap, r->rt_subordinates))
fprintf(fp2, " %u%c", i,
VIFM_ISSET(i, gt->gt_scope) ? 'b' :
SUBS_ARE_PRUNED(r->rt_subordinates,
uvifs[i].uv_nbrmap, gt->gt_prunes) ? 'p' : '!');
}
fprintf(fp2, "\n");
if (gt->gt_pruntbl) {
fprintf(fp2, "<");
c = '(';
for (pt = gt->gt_pruntbl; pt; pt = pt->pt_next) {
fprintf(fp2, "%c%s:%d[%d]/%d", c, inet_fmt(pt->pt_router, s1),
pt->pt_vifi, pt->pt_index, pt->pt_timer);
c = ',';
}
fprintf(fp2, ")");
fprintf(fp2, " 0x%08lx%08lx\n",/*XXX*/
gt->gt_prunes.hi, gt->gt_prunes.lo);
}
for (st = gt->gt_srctbl; st; st = st->st_next) {
fprintf(fp2, ">%-18s %-8s %6ld", inet_fmt(st->st_origin, s1),
st->st_ctime ? scaletime(thyme - st->st_ctime) : "-",
st->st_savpkt);
if (st->st_ctime) {
struct sioc_sg_req sg_req;
sg_req.src.s_addr = st->st_origin;
sg_req.grp.s_addr = gt->gt_mcastgrp;
if (ioctl(udp_socket, SIOCGETSGCNT, (char *)&sg_req) < 0) {
log(LOG_WARNING, errno, "SIOCGETSGCNT on (%s %s)",
inet_fmt(st->st_origin, s1),
inet_fmt(gt->gt_mcastgrp, s2));
} else {
fprintf(fp2, " %8ld %8ld %4ld", sg_req.pktcnt,
sg_req.bytecnt, sg_req.wrong_if);
}
}
fprintf(fp2, "\n");
}
}
}
/*
* Traceroute function which returns traceroute replies to the requesting
* router. Also forwards the request to downstream routers.
*/
void
accept_mtrace(src, dst, group, data, no, datalen)
u_int32 src;
u_int32 dst;
u_int32 group;
char *data;
u_int no; /* promoted u_char */
int datalen;
{
u_char type;
struct rtentry *rt;
struct gtable *gt;
struct tr_query *qry;
struct tr_resp *resp;
int vifi;
char *p;
int rcount;
int errcode = TR_NO_ERR;
int resptype;
struct timeval tp;
struct sioc_vif_req v_req;
struct sioc_sg_req sg_req;
/* Remember qid across invocations */
static u_int32 oqid = 0;
/* timestamp the request/response */
gettimeofday(&tp, 0);
/*
* Check if it is a query or a response
*/
if (datalen == QLEN) {
type = QUERY;
IF_DEBUG(DEBUG_TRACE)
log(LOG_DEBUG, 0, "Initial traceroute query rcvd from %s to %s",
inet_fmt(src, s1), inet_fmt(dst, s2));
}
else if ((datalen - QLEN) % RLEN == 0) {
type = RESP;
IF_DEBUG(DEBUG_TRACE)
log(LOG_DEBUG, 0, "In-transit traceroute query rcvd from %s to %s",
inet_fmt(src, s1), inet_fmt(dst, s2));
if (IN_MULTICAST(ntohl(dst))) {
IF_DEBUG(DEBUG_TRACE)
log(LOG_DEBUG, 0, "Dropping multicast response");
return;
}
}
else {
log(LOG_WARNING, 0, "%s from %s to %s",
"Non decipherable traceroute request recieved",
inet_fmt(src, s1), inet_fmt(dst, s2));
return;
}
qry = (struct tr_query *)data;
/*
* if it is a packet with all reports filled, drop it
*/
if ((rcount = (datalen - QLEN)/RLEN) == no) {
IF_DEBUG(DEBUG_TRACE)
log(LOG_DEBUG, 0, "packet with all reports filled in");
return;
}
IF_DEBUG(DEBUG_TRACE) {
log(LOG_DEBUG, 0, "s: %s g: %s d: %s ", inet_fmt(qry->tr_src, s1),
inet_fmt(group, s2), inet_fmt(qry->tr_dst, s3));
log(LOG_DEBUG, 0, "rttl: %d rd: %s", qry->tr_rttl,
inet_fmt(qry->tr_raddr, s1));
log(LOG_DEBUG, 0, "rcount:%d, qid:%06x", rcount, qry->tr_qid);
}
/* determine the routing table entry for this traceroute */
rt = determine_route(qry->tr_src);
IF_DEBUG(DEBUG_TRACE)
if (rt) {
log(LOG_DEBUG, 0, "rt parent vif: %d rtr: %s metric: %d",
rt->rt_parent, inet_fmt(rt->rt_gateway, s1), rt->rt_metric);
log(LOG_DEBUG, 0, "rt origin %s",
RT_FMT(rt, s1));
} else
log(LOG_DEBUG, 0, "...no route");
/*
* Query type packet - check if rte exists
* Check if the query destination is a vif connected to me.
* and if so, whether I should start response back
*/
if (type == QUERY) {
if (oqid == qry->tr_qid) {
/*
* If the multicast router is a member of the group being
* queried, and the query is multicasted, then the router can
* recieve multiple copies of the same query. If we have already
* replied to this traceroute, just ignore it this time.
*
* This is not a total solution, but since if this fails you
* only get N copies, N <= the number of interfaces on the router,
* it is not fatal.
*/
IF_DEBUG(DEBUG_TRACE)
log(LOG_DEBUG, 0, "ignoring duplicate traceroute packet");
return;
}
if (rt == NULL) {
IF_DEBUG(DEBUG_TRACE)
log(LOG_DEBUG, 0, "Mcast traceroute: no route entry %s",
inet_fmt(qry->tr_src, s1));
if (IN_MULTICAST(ntohl(dst)))
return;
}
vifi = find_vif(qry->tr_dst, 0);
if (vifi == NO_VIF) {
/* The traceroute destination is not on one of my subnet vifs. */
IF_DEBUG(DEBUG_TRACE)
log(LOG_DEBUG, 0, "Destination %s not an interface",
inet_fmt(qry->tr_dst, s1));
if (IN_MULTICAST(ntohl(dst)))
return;
errcode = TR_WRONG_IF;
} else if (rt != NULL && !VIFM_ISSET(vifi, rt->rt_children)) {
IF_DEBUG(DEBUG_TRACE)
log(LOG_DEBUG, 0, "Destination %s not on forwarding tree for src %s",
inet_fmt(qry->tr_dst, s1), inet_fmt(qry->tr_src, s2));
if (IN_MULTICAST(ntohl(dst)))
return;
errcode = TR_WRONG_IF;
}
}
else {
/*
* determine which interface the packet came in on
* RESP packets travel hop-by-hop so this either traversed
* a tunnel or came from a directly attached mrouter.
*/
if ((vifi = find_vif(src, dst)) == NO_VIF) {
IF_DEBUG(DEBUG_TRACE)
log(LOG_DEBUG, 0, "Wrong interface for packet");
errcode = TR_WRONG_IF;
}
}
/* Now that we've decided to send a response, save the qid */
oqid = qry->tr_qid;
IF_DEBUG(DEBUG_TRACE)
log(LOG_DEBUG, 0, "Sending traceroute response");
/* copy the packet to the sending buffer */
p = send_buf + MIN_IP_HEADER_LEN + IGMP_MINLEN;
bcopy(data, p, datalen);
p += datalen;
/*
* If there is no room to insert our reply, coopt the previous hop
* error indication to relay this fact.
*/
if (p + sizeof(struct tr_resp) > send_buf + RECV_BUF_SIZE) {
resp = (struct tr_resp *)p - 1;
resp->tr_rflags = TR_NO_SPACE;
rt = NULL;
goto sendit;
}
/*
* fill in initial response fields
*/
resp = (struct tr_resp *)p;
bzero(resp, sizeof(struct tr_resp));
datalen += RLEN;
resp->tr_qarr = htonl(((tp.tv_sec + JAN_1970) << 16) +
((tp.tv_usec << 10) / 15625));
resp->tr_rproto = PROTO_DVMRP;
resp->tr_outaddr = (vifi == NO_VIF) ? dst : uvifs[vifi].uv_lcl_addr;
resp->tr_fttl = (vifi == NO_VIF) ? 0 : uvifs[vifi].uv_threshold;
resp->tr_rflags = errcode;
/*
* obtain # of packets out on interface
*/
v_req.vifi = vifi;
if (vifi != NO_VIF && ioctl(udp_socket, SIOCGETVIFCNT, (char *)&v_req) >= 0)
resp->tr_vifout = htonl(v_req.ocount);
else
resp->tr_vifout = 0xffffffff;
/*
* fill in scoping & pruning information
*/
if (rt)
for (gt = rt->rt_groups; gt; gt = gt->gt_next) {
if (gt->gt_mcastgrp >= group)
break;
}
else
gt = NULL;
if (gt && gt->gt_mcastgrp == group) {
struct stable *st;
for (st = gt->gt_srctbl; st; st = st->st_next)
if (qry->tr_src == st->st_origin)
break;
sg_req.src.s_addr = qry->tr_src;
sg_req.grp.s_addr = group;
if (st && st->st_ctime != 0 &&
ioctl(udp_socket, SIOCGETSGCNT, (char *)&sg_req) >= 0)
resp->tr_pktcnt = htonl(sg_req.pktcnt + st->st_savpkt);
else
resp->tr_pktcnt = htonl(st ? st->st_savpkt : 0xffffffff);
if (VIFM_ISSET(vifi, gt->gt_scope))
resp->tr_rflags = TR_SCOPED;
else if (gt->gt_prsent_timer)
resp->tr_rflags = TR_PRUNED;
else if (!VIFM_ISSET(vifi, gt->gt_grpmems))
if (!NBRM_ISEMPTY(uvifs[vifi].uv_nbrmap) &&
SUBS_ARE_PRUNED(rt->rt_subordinates,
uvifs[vifi].uv_nbrmap, gt->gt_prunes))
resp->tr_rflags = TR_OPRUNED;
else
resp->tr_rflags = TR_NO_FWD;
} else {
if ((vifi != NO_VIF && scoped_addr(vifi, group)) ||
(rt && scoped_addr(rt->rt_parent, group)))
resp->tr_rflags = TR_SCOPED;
else if (rt && !VIFM_ISSET(vifi, rt->rt_children))
resp->tr_rflags = TR_NO_FWD;
}
/*
* if no rte exists, set NO_RTE error
*/
if (rt == NULL) {
src = dst; /* the dst address of resp. pkt */
resp->tr_inaddr = 0;
resp->tr_rflags = TR_NO_RTE;
resp->tr_rmtaddr = 0;
} else {
/* get # of packets in on interface */
v_req.vifi = rt->rt_parent;
if (ioctl(udp_socket, SIOCGETVIFCNT, (char *)&v_req) >= 0)
resp->tr_vifin = htonl(v_req.icount);
else
resp->tr_vifin = 0xffffffff;
MASK_TO_VAL(rt->rt_originmask, resp->tr_smask);
src = uvifs[rt->rt_parent].uv_lcl_addr;
resp->tr_inaddr = src;
resp->tr_rmtaddr = rt->rt_gateway;
if (!VIFM_ISSET(vifi, rt->rt_children)) {
IF_DEBUG(DEBUG_TRACE)
log(LOG_DEBUG, 0, "Destination %s not on forwarding tree for src %s",
inet_fmt(qry->tr_dst, s1), inet_fmt(qry->tr_src, s2));
resp->tr_rflags = TR_WRONG_IF;
}
if (rt->rt_metric >= UNREACHABLE) {
resp->tr_rflags = TR_NO_RTE;
/* Hack to send reply directly */
rt = NULL;
}
}
sendit:
/*
* if metric is 1 or no. of reports is 1, send response to requestor
* else send to upstream router. If the upstream router can't handle
* mtrace, set an error code and send to requestor anyway.
*/
IF_DEBUG(DEBUG_TRACE)
log(LOG_DEBUG, 0, "rcount:%d, no:%d", rcount, no);
if ((rcount + 1 == no) || (rt == NULL) || (rt->rt_metric == 1)) {
resptype = IGMP_MTRACE_RESP;
dst = qry->tr_raddr;
} else
if (!can_mtrace(rt->rt_parent, rt->rt_gateway)) {
dst = qry->tr_raddr;
resp->tr_rflags = TR_OLD_ROUTER;
resptype = IGMP_MTRACE_RESP;
} else {
dst = rt->rt_gateway;
resptype = IGMP_MTRACE;
}
if (IN_MULTICAST(ntohl(dst))) {
/*
* Send the reply on a known multicast capable vif.
* If we don't have one, we can't source any multicasts anyway.
*/
if (phys_vif != -1) {
IF_DEBUG(DEBUG_TRACE)
log(LOG_DEBUG, 0, "Sending reply to %s from %s",
inet_fmt(dst, s1), inet_fmt(uvifs[phys_vif].uv_lcl_addr, s2));
k_set_ttl(qry->tr_rttl);
send_igmp(uvifs[phys_vif].uv_lcl_addr, dst,
resptype, no, group,
datalen);
k_set_ttl(1);
} else
log(LOG_INFO, 0, "No enabled phyints -- %s",
"dropping traceroute reply");
} else {
IF_DEBUG(DEBUG_TRACE)
log(LOG_DEBUG, 0, "Sending %s to %s from %s",
resptype == IGMP_MTRACE_RESP ? "reply" : "request on",
inet_fmt(dst, s1), inet_fmt(src, s2));
send_igmp(src, dst,
resptype, no, group,
datalen);
}
return;
}