1 /* $NetBSD: prune.c,v 1.3 1995/12/10 10:07:09 mycroft Exp $ */ 2 3 /* 4 * The mrouted program is covered by the license in the accompanying file 5 * named "LICENSE". Use of the mrouted program represents acceptance of 6 * the terms and conditions listed in that file. 7 * 8 * The mrouted program is COPYRIGHT 1989 by The Board of Trustees of 9 * Leland Stanford Junior University. 10 */ 11 12 13 #include "defs.h" 14 15 extern int cache_lifetime; 16 extern int max_prune_lifetime; 17 extern struct rtentry *routing_table; 18 19 extern int phys_vif; 20 21 /* 22 * dither cache lifetime to obtain a value between x and 2*x 23 */ 24 #define CACHE_LIFETIME(x) ((x) + (arc4random_uniform(x))) 25 26 #define CHK_GS(x, y) { \ 27 switch(x) { \ 28 case 2: \ 29 case 4: \ 30 case 8: \ 31 case 16: \ 32 case 32: \ 33 case 64: \ 34 case 128: \ 35 case 256: y = 1; \ 36 break; \ 37 default: y = 0; \ 38 } \ 39 } 40 41 struct gtable *kernel_table; /* ptr to list of kernel grp entries*/ 42 static struct gtable *kernel_no_route; /* list of grp entries w/o routes */ 43 struct gtable *gtp; /* pointer for kernel rt entries */ 44 unsigned int kroutes; /* current number of cache entries */ 45 46 /**************************************************************************** 47 Functions that are local to prune.c 48 ****************************************************************************/ 49 static void prun_add_ttls(struct gtable *gt); 50 static int pruning_neighbor(vifi_t vifi, u_int32_t addr); 51 static int can_mtrace(vifi_t vifi, u_int32_t addr); 52 static struct ptable * find_prune_entry(u_int32_t vr, struct ptable *pt); 53 static void expire_prune(vifi_t vifi, struct gtable *gt); 54 static void send_prune(struct gtable *gt); 55 static void send_graft(struct gtable *gt); 56 static void send_graft_ack(u_int32_t src, u_int32_t dst, 57 u_int32_t origin, u_int32_t grp); 58 static void update_kernel(struct gtable *g); 59 static char * scaletime(time_t t); 60 61 /* 62 * Updates the ttl values for each vif. 63 */ 64 static void 65 prun_add_ttls(struct gtable *gt) 66 { 67 struct uvif *v; 68 vifi_t vifi; 69 70 for (vifi = 0, v = uvifs; vifi < numvifs; ++vifi, ++v) { 71 if (VIFM_ISSET(vifi, gt->gt_grpmems)) 72 gt->gt_ttls[vifi] = v->uv_threshold; 73 else 74 gt->gt_ttls[vifi] = 0; 75 } 76 } 77 78 /* 79 * checks for scoped multicast addresses 80 */ 81 #define GET_SCOPE(gt) { \ 82 vifi_t _i; \ 83 if ((ntohl((gt)->gt_mcastgrp) & 0xff000000) == 0xef000000) \ 84 for (_i = 0; _i < numvifs; _i++) \ 85 if (scoped_addr(_i, (gt)->gt_mcastgrp)) \ 86 VIFM_SET(_i, (gt)->gt_scope); \ 87 } 88 89 int 90 scoped_addr(vifi_t vifi, u_int32_t addr) 91 { 92 struct vif_acl *acl; 93 94 for (acl = uvifs[vifi].uv_acl; acl; acl = acl->acl_next) 95 if ((addr & acl->acl_mask) == acl->acl_addr) 96 return 1; 97 98 return 0; 99 } 100 101 /* 102 * Determine if mcastgrp has a listener on vifi 103 */ 104 int 105 grplst_mem(vifi_t vifi, u_int32_t mcastgrp) 106 { 107 struct listaddr *g; 108 struct uvif *v; 109 110 v = &uvifs[vifi]; 111 112 for (g = v->uv_groups; g != NULL; g = g->al_next) 113 if (mcastgrp == g->al_addr) 114 return 1; 115 116 return 0; 117 } 118 119 /* 120 * Finds the group entry with the specified source and netmask. 121 * If netmask is 0, it uses the route's netmask. 122 * 123 * Returns TRUE if found a match, and the global variable gtp is left 124 * pointing to entry before the found entry. 125 * Returns FALSE if no exact match found, gtp is left pointing to before 126 * the entry in question belongs, or is NULL if the it belongs at the 127 * head of the list. 128 */ 129 int 130 find_src_grp(u_int32_t src, u_int32_t mask, u_int32_t grp) 131 { 132 struct gtable *gt; 133 134 gtp = NULL; 135 gt = kernel_table; 136 while (gt != NULL) { 137 if (grp == gt->gt_mcastgrp && 138 (mask ? (gt->gt_route->rt_origin == src && 139 gt->gt_route->rt_originmask == mask) : 140 ((src & gt->gt_route->rt_originmask) == 141 gt->gt_route->rt_origin))) 142 return TRUE; 143 if (ntohl(grp) > ntohl(gt->gt_mcastgrp) || 144 (grp == gt->gt_mcastgrp && 145 (ntohl(mask) < ntohl(gt->gt_route->rt_originmask) || 146 (mask == gt->gt_route->rt_originmask && 147 (ntohl(src) > ntohl(gt->gt_route->rt_origin)))))) { 148 gtp = gt; 149 gt = gt->gt_gnext; 150 } 151 else break; 152 } 153 return FALSE; 154 } 155 156 /* 157 * Check if the neighbor supports pruning 158 */ 159 static int 160 pruning_neighbor(vifi_t vifi, u_int32_t addr) 161 { 162 struct listaddr *n = neighbor_info(vifi, addr); 163 int vers; 164 165 if (n == NULL) 166 return 0; 167 168 if (n->al_flags & NF_PRUNE) 169 return 1; 170 171 /* 172 * Versions from 3.0 to 3.4 relied on the version number to identify 173 * that they could handle pruning. 174 */ 175 vers = NBR_VERS(n); 176 return (vers >= 0x0300 && vers <= 0x0304); 177 } 178 179 /* 180 * Can the neighbor in question handle multicast traceroute? 181 */ 182 static int 183 can_mtrace(vifi_t vifi, u_int32_t addr) 184 { 185 struct listaddr *n = neighbor_info(vifi, addr); 186 int vers; 187 188 if (n == NULL) 189 return 0; 190 191 if (n->al_flags & NF_MTRACE) 192 return 1; 193 194 /* 195 * Versions 3.3 and 3.4 relied on the version number to identify 196 * that they could handle traceroute. 197 */ 198 vers = NBR_VERS(n); 199 return (vers >= 0x0303 && vers <= 0x0304); 200 } 201 202 /* 203 * Returns the prune entry of the router, or NULL if none exists 204 */ 205 static struct ptable * 206 find_prune_entry(u_int32_t vr, struct ptable *pt) 207 { 208 while (pt) { 209 if (pt->pt_router == vr) 210 return pt; 211 pt = pt->pt_next; 212 } 213 214 return NULL; 215 } 216 217 /* 218 * Send a prune message to the dominant router for 219 * this source. 220 * 221 * Record an entry that a prune was sent for this group 222 */ 223 static void 224 send_prune(struct gtable *gt) 225 { 226 struct ptable *pt; 227 char *p; 228 int i; 229 int datalen; 230 u_int32_t src; 231 u_int32_t dst; 232 u_int32_t tmp; 233 234 /* Don't process any prunes if router is not pruning */ 235 if (pruning == 0) 236 return; 237 238 /* Can't process a prune if we don't have an associated route */ 239 if (gt->gt_route == NULL) 240 return; 241 242 /* Don't send a prune to a non-pruning router */ 243 if (!pruning_neighbor(gt->gt_route->rt_parent, gt->gt_route->rt_gateway)) 244 return; 245 246 /* 247 * sends a prune message to the router upstream. 248 */ 249 src = uvifs[gt->gt_route->rt_parent].uv_lcl_addr; 250 dst = gt->gt_route->rt_gateway; 251 252 p = send_buf + MIN_IP_HEADER_LEN + IGMP_MINLEN; 253 datalen = 0; 254 255 /* 256 * determine prune lifetime 257 */ 258 gt->gt_prsent_timer = gt->gt_timer; 259 for (pt = gt->gt_pruntbl; pt; pt = pt->pt_next) 260 if (pt->pt_timer < gt->gt_prsent_timer) 261 gt->gt_prsent_timer = pt->pt_timer; 262 263 /* 264 * If we have a graft pending, cancel graft retransmission 265 */ 266 gt->gt_grftsnt = 0; 267 268 for (i = 0; i < 4; i++) 269 *p++ = ((char *)&(gt->gt_route->rt_origin))[i]; 270 for (i = 0; i < 4; i++) 271 *p++ = ((char *)&(gt->gt_mcastgrp))[i]; 272 tmp = htonl(gt->gt_prsent_timer); 273 for (i = 0; i < 4; i++) 274 *p++ = ((char *)&(tmp))[i]; 275 datalen += 12; 276 277 send_igmp(src, dst, IGMP_DVMRP, DVMRP_PRUNE, 278 htonl(MROUTED_LEVEL), datalen); 279 280 logit(LOG_DEBUG, 0, "sent prune for (%s %s)/%d on vif %d to %s", 281 inet_fmts(gt->gt_route->rt_origin, gt->gt_route->rt_originmask, s1), 282 inet_fmt(gt->gt_mcastgrp, s2), 283 gt->gt_prsent_timer, gt->gt_route->rt_parent, 284 inet_fmt(gt->gt_route->rt_gateway, s3)); 285 } 286 287 /* 288 * a prune was sent upstream 289 * so, a graft has to be sent to annul the prune 290 * set up a graft timer so that if an ack is not 291 * heard within that time, another graft request 292 * is sent out. 293 */ 294 static void 295 send_graft(struct gtable *gt) 296 { 297 char *p; 298 int i; 299 int datalen; 300 u_int32_t src; 301 u_int32_t dst; 302 303 /* Can't send a graft without an associated route */ 304 if (gt->gt_route == NULL) 305 return; 306 307 src = uvifs[gt->gt_route->rt_parent].uv_lcl_addr; 308 dst = gt->gt_route->rt_gateway; 309 310 p = send_buf + MIN_IP_HEADER_LEN + IGMP_MINLEN; 311 datalen = 0; 312 313 for (i = 0; i < 4; i++) 314 *p++ = ((char *)&(gt->gt_route->rt_origin))[i]; 315 for (i = 0; i < 4; i++) 316 *p++ = ((char *)&(gt->gt_mcastgrp))[i]; 317 datalen += 8; 318 319 if (datalen != 0) { 320 send_igmp(src, dst, IGMP_DVMRP, DVMRP_GRAFT, 321 htonl(MROUTED_LEVEL), datalen); 322 } 323 logit(LOG_DEBUG, 0, "sent graft for (%s %s) to %s on vif %d", 324 inet_fmts(gt->gt_route->rt_origin, gt->gt_route->rt_originmask, s1), 325 inet_fmt(gt->gt_mcastgrp, s2), 326 inet_fmt(gt->gt_route->rt_gateway, s3), gt->gt_route->rt_parent); 327 } 328 329 /* 330 * Send an ack that a graft was received 331 */ 332 static void 333 send_graft_ack(u_int32_t src, u_int32_t dst, u_int32_t origin, u_int32_t grp) 334 { 335 char *p; 336 int i; 337 int datalen; 338 339 p = send_buf + MIN_IP_HEADER_LEN + IGMP_MINLEN; 340 datalen = 0; 341 342 for (i = 0; i < 4; i++) 343 *p++ = ((char *)&(origin))[i]; 344 for (i = 0; i < 4; i++) 345 *p++ = ((char *)&(grp))[i]; 346 datalen += 8; 347 348 send_igmp(src, dst, IGMP_DVMRP, DVMRP_GRAFT_ACK, 349 htonl(MROUTED_LEVEL), datalen); 350 351 logit(LOG_DEBUG, 0, "sent graft ack for (%s, %s) to %s", 352 inet_fmt(origin, s1), inet_fmt(grp, s2), inet_fmt(dst, s3)); 353 } 354 355 /* 356 * Update the kernel cache with all the routes hanging off the group entry 357 */ 358 static void 359 update_kernel(struct gtable *g) 360 { 361 struct stable *st; 362 363 for (st = g->gt_srctbl; st; st = st->st_next) 364 k_add_rg(st->st_origin, g); 365 } 366 367 /**************************************************************************** 368 Functions that are used externally 369 ****************************************************************************/ 370 371 /* 372 * Initialize the kernel table structure 373 */ 374 void 375 init_ktable(void) 376 { 377 kernel_table = NULL; 378 kernel_no_route = NULL; 379 kroutes = 0; 380 } 381 382 /* 383 * Add a new table entry for (origin, mcastgrp) 384 */ 385 void 386 add_table_entry(u_int32_t origin, u_int32_t mcastgrp) 387 { 388 struct rtentry *r; 389 struct gtable *gt,**gtnp,*prev_gt; 390 struct stable *st,**stnp; 391 vifi_t i; 392 393 #ifdef DEBUG_MFC 394 md_logit(MD_MISS, origin, mcastgrp); 395 #endif 396 397 r = determine_route(origin); 398 prev_gt = NULL; 399 if (r == NULL) { 400 /* 401 * Look for it on the no_route table; if it is found then 402 * it will be detected as a duplicate below. 403 */ 404 for (gt = kernel_no_route; gt; gt = gt->gt_next) 405 if (mcastgrp == gt->gt_mcastgrp && 406 gt->gt_srctbl && gt->gt_srctbl->st_origin == origin) 407 break; 408 gtnp = &kernel_no_route; 409 } else { 410 gtnp = &r->rt_groups; 411 while ((gt = *gtnp) != NULL) { 412 if (gt->gt_mcastgrp >= mcastgrp) 413 break; 414 gtnp = >->gt_next; 415 prev_gt = gt; 416 } 417 } 418 419 if (gt == NULL || gt->gt_mcastgrp != mcastgrp) { 420 gt = malloc(sizeof(struct gtable)); 421 if (gt == NULL) 422 logit(LOG_ERR, 0, "ran out of memory"); 423 424 gt->gt_mcastgrp = mcastgrp; 425 gt->gt_timer = CACHE_LIFETIME(cache_lifetime); 426 time(>->gt_ctime); 427 gt->gt_grpmems = 0; 428 gt->gt_scope = 0; 429 gt->gt_prsent_timer = 0; 430 gt->gt_grftsnt = 0; 431 gt->gt_srctbl = NULL; 432 gt->gt_pruntbl = NULL; 433 gt->gt_route = r; 434 #ifdef RSRR 435 gt->gt_rsrr_cache = NULL; 436 #endif 437 438 if (r != NULL) { 439 /* obtain the multicast group membership list */ 440 for (i = 0; i < numvifs; i++) { 441 if (VIFM_ISSET(i, r->rt_children) && 442 !(VIFM_ISSET(i, r->rt_leaves))) 443 VIFM_SET(i, gt->gt_grpmems); 444 445 if (VIFM_ISSET(i, r->rt_leaves) && grplst_mem(i, mcastgrp)) 446 VIFM_SET(i, gt->gt_grpmems); 447 } 448 GET_SCOPE(gt); 449 if (VIFM_ISSET(r->rt_parent, gt->gt_scope)) 450 gt->gt_scope = -1; 451 gt->gt_grpmems &= ~gt->gt_scope; 452 } else { 453 gt->gt_scope = -1; 454 gt->gt_grpmems = 0; 455 } 456 457 /* update ttls */ 458 prun_add_ttls(gt); 459 460 gt->gt_next = *gtnp; 461 *gtnp = gt; 462 if (gt->gt_next) 463 gt->gt_next->gt_prev = gt; 464 gt->gt_prev = prev_gt; 465 466 if (r) { 467 if (find_src_grp(r->rt_origin, r->rt_originmask, gt->gt_mcastgrp)) { 468 struct gtable *g; 469 470 g = gtp ? gtp->gt_gnext : kernel_table; 471 logit(LOG_WARNING, 0, "Entry for (%s %s) (rt:%x) exists (rt:%x)", 472 inet_fmts(r->rt_origin, r->rt_originmask, s1), 473 inet_fmt(g->gt_mcastgrp, s2), 474 r, g->gt_route); 475 } else { 476 if (gtp) { 477 gt->gt_gnext = gtp->gt_gnext; 478 gt->gt_gprev = gtp; 479 gtp->gt_gnext = gt; 480 } else { 481 gt->gt_gnext = kernel_table; 482 gt->gt_gprev = NULL; 483 kernel_table = gt; 484 } 485 if (gt->gt_gnext) 486 gt->gt_gnext->gt_gprev = gt; 487 } 488 } else { 489 gt->gt_gnext = gt->gt_gprev = NULL; 490 } 491 } 492 493 stnp = >->gt_srctbl; 494 while ((st = *stnp) != NULL) { 495 if (ntohl(st->st_origin) >= ntohl(origin)) 496 break; 497 stnp = &st->st_next; 498 } 499 500 if (st == NULL || st->st_origin != origin) { 501 st = malloc(sizeof(struct stable)); 502 if (st == NULL) 503 logit(LOG_ERR, 0, "ran out of memory"); 504 505 st->st_origin = origin; 506 st->st_pktcnt = 0; 507 st->st_next = *stnp; 508 *stnp = st; 509 } else { 510 #ifdef DEBUG_MFC 511 md_logit(MD_DUPE, origin, mcastgrp); 512 #endif 513 logit(LOG_WARNING, 0, "kernel entry already exists for (%s %s)", 514 inet_fmt(origin, s1), inet_fmt(mcastgrp, s2)); 515 /* XXX Doing this should cause no harm, and may ensure 516 * kernel<>mrouted synchronization */ 517 k_add_rg(origin, gt); 518 return; 519 } 520 521 kroutes++; 522 k_add_rg(origin, gt); 523 524 logit(LOG_DEBUG, 0, "add cache entry (%s %s) gm:%x, parent-vif:%d", 525 inet_fmt(origin, s1), 526 inet_fmt(mcastgrp, s2), 527 gt->gt_grpmems, r ? r->rt_parent : -1); 528 529 /* If there are no leaf vifs 530 * which have this group, then 531 * mark this src-grp as a prune candidate. 532 */ 533 if (!gt->gt_prsent_timer && !gt->gt_grpmems && r && r->rt_gateway) 534 send_prune(gt); 535 } 536 537 /* 538 * An mrouter has gone down and come up on an interface 539 * Forward on that interface immediately 540 */ 541 void 542 reset_neighbor_state(vifi_t vifi, u_int32_t addr) 543 { 544 struct rtentry *r; 545 struct gtable *g; 546 struct ptable *pt, **ptnp; 547 struct stable *st; 548 549 for (g = kernel_table; g; g = g->gt_gnext) { 550 r = g->gt_route; 551 552 /* 553 * If neighbor was the parent, remove the prune sent state 554 * and all of the source cache info so that prunes get 555 * regenerated. 556 */ 557 if (vifi == r->rt_parent) { 558 if (addr == r->rt_gateway) { 559 logit(LOG_DEBUG, 0, "reset_neighbor_state parent reset (%s %s)", 560 inet_fmts(r->rt_origin, r->rt_originmask, s1), 561 inet_fmt(g->gt_mcastgrp, s2)); 562 563 g->gt_prsent_timer = 0; 564 g->gt_grftsnt = 0; 565 while (st = g->gt_srctbl) { 566 g->gt_srctbl = st->st_next; 567 k_del_rg(st->st_origin, g); 568 kroutes--; 569 free(st); 570 } 571 } 572 } else { 573 /* 574 * Neighbor was not the parent, send grafts to join the groups 575 */ 576 if (g->gt_prsent_timer) { 577 g->gt_grftsnt = 1; 578 send_graft(g); 579 g->gt_prsent_timer = 0; 580 } 581 582 /* 583 * Remove any prunes that this router has sent us. 584 */ 585 ptnp = &g->gt_pruntbl; 586 while ((pt = *ptnp) != NULL) { 587 if (pt->pt_vifi == vifi && pt->pt_router == addr) { 588 *ptnp = pt->pt_next; 589 free(pt); 590 } else 591 ptnp = &pt->pt_next; 592 } 593 594 /* 595 * And see if we want to forward again. 596 */ 597 if (!VIFM_ISSET(vifi, g->gt_grpmems)) { 598 if (VIFM_ISSET(vifi, r->rt_children) && 599 !(VIFM_ISSET(vifi, r->rt_leaves))) 600 VIFM_SET(vifi, g->gt_grpmems); 601 602 if (VIFM_ISSET(vifi, r->rt_leaves) && 603 grplst_mem(vifi, g->gt_mcastgrp)) 604 VIFM_SET(vifi, g->gt_grpmems); 605 606 g->gt_grpmems &= ~g->gt_scope; 607 prun_add_ttls(g); 608 609 /* Update kernel state */ 610 update_kernel(g); 611 #ifdef RSRR 612 /* Send route change notification to reservation protocol. */ 613 rsrr_cache_send(g,1); 614 #endif /* RSRR */ 615 616 logit(LOG_DEBUG, 0, "reset member state (%s %s) gm:%x", 617 inet_fmts(r->rt_origin, r->rt_originmask, s1), 618 inet_fmt(g->gt_mcastgrp, s2), g->gt_grpmems); 619 } 620 } 621 } 622 } 623 624 /* 625 * Delete table entry from the kernel 626 * del_flag determines how many entries to delete 627 */ 628 void 629 del_table_entry(struct rtentry *r, u_int32_t mcastgrp, u_int del_flag) 630 { 631 struct gtable *g, *prev_g; 632 struct stable *st, *prev_st; 633 struct ptable *pt, *prev_pt; 634 635 if (del_flag == DEL_ALL_ROUTES) { 636 g = r->rt_groups; 637 while (g) { 638 logit(LOG_DEBUG, 0, "del_table_entry deleting (%s %s)", 639 inet_fmts(r->rt_origin, r->rt_originmask, s1), 640 inet_fmt(g->gt_mcastgrp, s2)); 641 st = g->gt_srctbl; 642 while (st) { 643 if (k_del_rg(st->st_origin, g) < 0) { 644 logit(LOG_WARNING, errno, 645 "del_table_entry trying to delete (%s, %s)", 646 inet_fmt(st->st_origin, s1), 647 inet_fmt(g->gt_mcastgrp, s2)); 648 } 649 kroutes--; 650 prev_st = st; 651 st = st->st_next; 652 free(prev_st); 653 } 654 g->gt_srctbl = NULL; 655 656 pt = g->gt_pruntbl; 657 while (pt) { 658 prev_pt = pt; 659 pt = pt->pt_next; 660 free(prev_pt); 661 } 662 g->gt_pruntbl = NULL; 663 664 if (g->gt_gnext) 665 g->gt_gnext->gt_gprev = g->gt_gprev; 666 if (g->gt_gprev) 667 g->gt_gprev->gt_gnext = g->gt_gnext; 668 else 669 kernel_table = g->gt_gnext; 670 671 #ifdef RSRR 672 /* Send route change notification to reservation protocol. */ 673 rsrr_cache_send(g,0); 674 rsrr_cache_clean(g); 675 #endif /* RSRR */ 676 prev_g = g; 677 g = g->gt_next; 678 free(prev_g); 679 } 680 r->rt_groups = NULL; 681 } 682 683 /* 684 * Dummy routine - someday this may be needed, so it is just there 685 */ 686 if (del_flag == DEL_RTE_GROUP) { 687 prev_g = (struct gtable *)&r->rt_groups; 688 for (g = r->rt_groups; g; g = g->gt_next) { 689 if (g->gt_mcastgrp == mcastgrp) { 690 logit(LOG_DEBUG, 0, "del_table_entry deleting (%s %s)", 691 inet_fmts(r->rt_origin, r->rt_originmask, s1), 692 inet_fmt(g->gt_mcastgrp, s2)); 693 st = g->gt_srctbl; 694 while (st) { 695 if (k_del_rg(st->st_origin, g) < 0) { 696 logit(LOG_WARNING, errno, 697 "del_table_entry trying to delete (%s, %s)", 698 inet_fmt(st->st_origin, s1), 699 inet_fmt(g->gt_mcastgrp, s2)); 700 } 701 kroutes--; 702 prev_st = st; 703 st = st->st_next; 704 free(prev_st); 705 } 706 g->gt_srctbl = NULL; 707 708 pt = g->gt_pruntbl; 709 while (pt) { 710 prev_pt = pt; 711 pt = pt->pt_next; 712 free(prev_pt); 713 } 714 g->gt_pruntbl = NULL; 715 716 if (g->gt_gnext) 717 g->gt_gnext->gt_gprev = g->gt_gprev; 718 if (g->gt_gprev) 719 g->gt_gprev->gt_gnext = g->gt_gnext; 720 else 721 kernel_table = g->gt_gnext; 722 723 if (prev_g != (struct gtable *)&r->rt_groups) 724 g->gt_next->gt_prev = prev_g; 725 else 726 g->gt_next->gt_prev = NULL; 727 prev_g->gt_next = g->gt_next; 728 729 #ifdef RSRR 730 /* Send route change notification to reservation protocol. */ 731 rsrr_cache_send(g,0); 732 rsrr_cache_clean(g); 733 #endif /* RSRR */ 734 free(g); 735 g = prev_g; 736 } else { 737 prev_g = g; 738 } 739 } 740 } 741 } 742 743 /* 744 * update kernel table entry when a route entry changes 745 */ 746 void 747 update_table_entry(struct rtentry *r) 748 { 749 struct gtable *g; 750 struct ptable *pt, *prev_pt; 751 vifi_t i; 752 753 for (g = r->rt_groups; g; g = g->gt_next) { 754 pt = g->gt_pruntbl; 755 while (pt) { 756 prev_pt = pt->pt_next; 757 free(pt); 758 pt = prev_pt; 759 } 760 g->gt_pruntbl = NULL; 761 762 g->gt_grpmems = 0; 763 764 /* obtain the multicast group membership list */ 765 for (i = 0; i < numvifs; i++) { 766 if (VIFM_ISSET(i, r->rt_children) && 767 !(VIFM_ISSET(i, r->rt_leaves))) 768 VIFM_SET(i, g->gt_grpmems); 769 770 if (VIFM_ISSET(i, r->rt_leaves) && grplst_mem(i, g->gt_mcastgrp)) 771 VIFM_SET(i, g->gt_grpmems); 772 } 773 if (VIFM_ISSET(r->rt_parent, g->gt_scope)) 774 g->gt_scope = -1; 775 g->gt_grpmems &= ~g->gt_scope; 776 777 logit(LOG_DEBUG, 0, "updating cache entries (%s %s) gm:%x", 778 inet_fmts(r->rt_origin, r->rt_originmask, s1), 779 inet_fmt(g->gt_mcastgrp, s2), 780 g->gt_grpmems); 781 782 if (g->gt_grpmems && g->gt_prsent_timer) { 783 g->gt_grftsnt = 1; 784 send_graft(g); 785 g->gt_prsent_timer = 0; 786 } 787 788 /* update ttls and add entry into kernel */ 789 prun_add_ttls(g); 790 update_kernel(g); 791 #ifdef RSRR 792 /* Send route change notification to reservation protocol. */ 793 rsrr_cache_send(g,1); 794 #endif /* RSRR */ 795 796 /* Check if we want to prune this group */ 797 if (!g->gt_prsent_timer && g->gt_grpmems == 0 && r->rt_gateway) { 798 g->gt_timer = CACHE_LIFETIME(cache_lifetime); 799 send_prune(g); 800 } 801 } 802 } 803 804 /* 805 * set the forwarding flag for all mcastgrps on this vifi 806 */ 807 void 808 update_lclgrp(vifi_t vifi, u_int32_t mcastgrp) 809 { 810 struct rtentry *r; 811 struct gtable *g; 812 813 logit(LOG_DEBUG, 0, "group %s joined on vif %d", 814 inet_fmt(mcastgrp, s1), vifi); 815 816 for (g = kernel_table; g; g = g->gt_gnext) { 817 if (ntohl(mcastgrp) < ntohl(g->gt_mcastgrp)) 818 break; 819 820 r = g->gt_route; 821 if (g->gt_mcastgrp == mcastgrp && 822 VIFM_ISSET(vifi, r->rt_children)) { 823 824 VIFM_SET(vifi, g->gt_grpmems); 825 g->gt_grpmems &= ~g->gt_scope; 826 if (g->gt_grpmems == 0) 827 continue; 828 829 prun_add_ttls(g); 830 logit(LOG_DEBUG, 0, "update lclgrp (%s %s) gm:%x", 831 inet_fmts(r->rt_origin, r->rt_originmask, s1), 832 inet_fmt(g->gt_mcastgrp, s2), g->gt_grpmems); 833 834 update_kernel(g); 835 #ifdef RSRR 836 /* Send route change notification to reservation protocol. */ 837 rsrr_cache_send(g,1); 838 #endif /* RSRR */ 839 } 840 } 841 } 842 843 /* 844 * reset forwarding flag for all mcastgrps on this vifi 845 */ 846 void 847 delete_lclgrp(vifi_t vifi, u_int32_t mcastgrp) 848 { 849 struct rtentry *r; 850 struct gtable *g; 851 852 logit(LOG_DEBUG, 0, "group %s left on vif %d", 853 inet_fmt(mcastgrp, s1), vifi); 854 855 for (g = kernel_table; g; g = g->gt_gnext) { 856 if (ntohl(mcastgrp) < ntohl(g->gt_mcastgrp)) 857 break; 858 859 if (g->gt_mcastgrp == mcastgrp) { 860 int stop_sending = 1; 861 862 r = g->gt_route; 863 /* 864 * If this is not a leaf, then we have router neighbors on this 865 * vif. Only turn off forwarding if they have all pruned. 866 */ 867 if (!VIFM_ISSET(vifi, r->rt_leaves)) { 868 struct listaddr *vr; 869 870 for (vr = uvifs[vifi].uv_neighbors; vr; vr = vr->al_next) 871 if (find_prune_entry(vr->al_addr, g->gt_pruntbl) == NULL) { 872 stop_sending = 0; 873 break; 874 } 875 } 876 877 if (stop_sending) { 878 VIFM_CLR(vifi, g->gt_grpmems); 879 logit(LOG_DEBUG, 0, "delete lclgrp (%s %s) gm:%x", 880 inet_fmts(r->rt_origin, r->rt_originmask, s1), 881 inet_fmt(g->gt_mcastgrp, s2), g->gt_grpmems); 882 883 prun_add_ttls(g); 884 update_kernel(g); 885 #ifdef RSRR 886 /* Send route change notification to reservation protocol. */ 887 rsrr_cache_send(g,1); 888 #endif /* RSRR */ 889 890 /* 891 * If there are no more members of this particular group, 892 * send prune upstream 893 */ 894 if (!g->gt_prsent_timer && g->gt_grpmems == 0 && r->rt_gateway) 895 send_prune(g); 896 } 897 } 898 } 899 } 900 901 /* 902 * Takes the prune message received and then strips it to 903 * determine the (src, grp) pair to be pruned. 904 * 905 * Adds the router to the (src, grp) entry then. 906 * 907 * Determines if further packets have to be sent down that vif 908 * 909 * Determines if a corresponding prune message has to be generated 910 */ 911 void 912 accept_prune(u_int32_t src, u_int32_t dst, char *p, int datalen) 913 { 914 u_int32_t prun_src; 915 u_int32_t prun_grp; 916 u_int32_t prun_tmr; 917 vifi_t vifi; 918 int i; 919 int stop_sending; 920 struct rtentry *r; 921 struct gtable *g; 922 struct ptable *pt; 923 struct listaddr *vr; 924 925 /* Don't process any prunes if router is not pruning */ 926 if (pruning == 0) 927 return; 928 929 if ((vifi = find_vif(src, dst)) == NO_VIF) { 930 logit(LOG_INFO, 0, 931 "ignoring prune report from non-neighbor %s", 932 inet_fmt(src, s1)); 933 return; 934 } 935 936 /* Check if enough data is present */ 937 if (datalen < 12) 938 { 939 logit(LOG_WARNING, 0, 940 "non-decipherable prune from %s", 941 inet_fmt(src, s1)); 942 return; 943 } 944 945 for (i = 0; i< 4; i++) 946 ((char *)&prun_src)[i] = *p++; 947 for (i = 0; i< 4; i++) 948 ((char *)&prun_grp)[i] = *p++; 949 for (i = 0; i< 4; i++) 950 ((char *)&prun_tmr)[i] = *p++; 951 prun_tmr = ntohl(prun_tmr); 952 953 logit(LOG_DEBUG, 0, "%s on vif %d prunes (%s %s)/%d", 954 inet_fmt(src, s1), vifi, 955 inet_fmt(prun_src, s2), inet_fmt(prun_grp, s3), prun_tmr); 956 957 /* 958 * Find the subnet for the prune 959 */ 960 if (find_src_grp(prun_src, 0, prun_grp)) { 961 g = gtp ? gtp->gt_gnext : kernel_table; 962 r = g->gt_route; 963 964 if (!VIFM_ISSET(vifi, r->rt_children)) { 965 logit(LOG_WARNING, 0, "prune received from non-child %s for (%s %s)", 966 inet_fmt(src, s1), inet_fmt(prun_src, s2), 967 inet_fmt(prun_grp, s3)); 968 return; 969 } 970 if (VIFM_ISSET(vifi, g->gt_scope)) { 971 logit(LOG_WARNING, 0, "prune received from %s on scoped grp (%s %s)", 972 inet_fmt(src, s1), inet_fmt(prun_src, s2), 973 inet_fmt(prun_grp, s3)); 974 return; 975 } 976 if ((pt = find_prune_entry(src, g->gt_pruntbl)) != NULL) { 977 /* 978 * If it's about to expire, then it's only still around because 979 * of timer granularity, so don't warn about it. 980 */ 981 if (pt->pt_timer > 10) { 982 logit(LOG_WARNING, 0, "%s %d from %s for (%s %s)/%d %s %d %s %x", 983 "duplicate prune received on vif", 984 vifi, inet_fmt(src, s1), inet_fmt(prun_src, s2), 985 inet_fmt(prun_grp, s3), prun_tmr, 986 "old timer:", pt->pt_timer, "cur gm:", g->gt_grpmems); 987 } 988 pt->pt_timer = prun_tmr; 989 } else { 990 /* allocate space for the prune structure */ 991 pt = malloc(sizeof(struct ptable)); 992 if (pt == NULL) 993 logit(LOG_ERR, 0, "pt: ran out of memory"); 994 995 pt->pt_vifi = vifi; 996 pt->pt_router = src; 997 pt->pt_timer = prun_tmr; 998 999 pt->pt_next = g->gt_pruntbl; 1000 g->gt_pruntbl = pt; 1001 } 1002 1003 /* Refresh the group's lifetime */ 1004 g->gt_timer = CACHE_LIFETIME(cache_lifetime); 1005 if (g->gt_timer < prun_tmr) 1006 g->gt_timer = prun_tmr; 1007 1008 /* 1009 * check if any more packets need to be sent on the 1010 * vif which sent this message 1011 */ 1012 stop_sending = 1; 1013 for (vr = uvifs[vifi].uv_neighbors; vr; vr = vr->al_next) 1014 if (find_prune_entry(vr->al_addr, g->gt_pruntbl) == NULL) { 1015 stop_sending = 0; 1016 break; 1017 } 1018 1019 if (stop_sending && !grplst_mem(vifi, prun_grp)) { 1020 VIFM_CLR(vifi, g->gt_grpmems); 1021 logit(LOG_DEBUG, 0, "prune (%s %s), stop sending on vif %d, gm:%x", 1022 inet_fmts(r->rt_origin, r->rt_originmask, s1), 1023 inet_fmt(g->gt_mcastgrp, s2), vifi, g->gt_grpmems); 1024 1025 prun_add_ttls(g); 1026 update_kernel(g); 1027 #ifdef RSRR 1028 /* Send route change notification to reservation protocol. */ 1029 rsrr_cache_send(g,1); 1030 #endif /* RSRR */ 1031 } 1032 1033 /* 1034 * check if all the child routers have expressed no interest 1035 * in this group and if this group does not exist in the 1036 * interface 1037 * Send a prune message then upstream 1038 */ 1039 if (!g->gt_prsent_timer && g->gt_grpmems == 0 && r->rt_gateway) { 1040 send_prune(g); 1041 } 1042 } else { 1043 /* 1044 * There is no kernel entry for this group. Therefore, we can 1045 * simply ignore the prune, as we are not forwarding this traffic 1046 * downstream. 1047 */ 1048 logit(LOG_DEBUG, 0, "%s (%s %s)/%d from %s", 1049 "prune message received with no kernel entry for", 1050 inet_fmt(prun_src, s1), inet_fmt(prun_grp, s2), 1051 prun_tmr, inet_fmt(src, s3)); 1052 return; 1053 } 1054 } 1055 1056 /* 1057 * Checks if this mcastgrp is present in the kernel table 1058 * If so and if a prune was sent, it sends a graft upwards 1059 */ 1060 void 1061 chkgrp_graft(vifi_t vifi, u_int32_t mcastgrp) 1062 { 1063 struct rtentry *r; 1064 struct gtable *g; 1065 1066 for (g = kernel_table; g; g = g->gt_gnext) { 1067 if (ntohl(mcastgrp) < ntohl(g->gt_mcastgrp)) 1068 break; 1069 1070 r = g->gt_route; 1071 if (g->gt_mcastgrp == mcastgrp && VIFM_ISSET(vifi, r->rt_children)) 1072 if (g->gt_prsent_timer) { 1073 VIFM_SET(vifi, g->gt_grpmems); 1074 1075 /* 1076 * If the vif that was joined was a scoped vif, 1077 * ignore it ; don't graft back 1078 */ 1079 g->gt_grpmems &= ~g->gt_scope; 1080 if (g->gt_grpmems == 0) 1081 continue; 1082 1083 /* set the flag for graft retransmission */ 1084 g->gt_grftsnt = 1; 1085 1086 /* send graft upwards */ 1087 send_graft(g); 1088 1089 /* reset the prune timer and update cache timer*/ 1090 g->gt_prsent_timer = 0; 1091 g->gt_timer = max_prune_lifetime; 1092 1093 logit(LOG_DEBUG, 0, "chkgrp graft (%s %s) gm:%x", 1094 inet_fmts(r->rt_origin, r->rt_originmask, s1), 1095 inet_fmt(g->gt_mcastgrp, s2), g->gt_grpmems); 1096 1097 prun_add_ttls(g); 1098 update_kernel(g); 1099 #ifdef RSRR 1100 /* Send route change notification to reservation protocol. */ 1101 rsrr_cache_send(g,1); 1102 #endif /* RSRR */ 1103 } 1104 } 1105 } 1106 1107 /* determine the multicast group and src 1108 * 1109 * if it does, then determine if a prune was sent 1110 * upstream. 1111 * if prune sent upstream, send graft upstream and send 1112 * ack downstream. 1113 * 1114 * if no prune sent upstream, change the forwarding bit 1115 * for this interface and send ack downstream. 1116 * 1117 * if no entry exists for this group send ack downstream. 1118 */ 1119 void 1120 accept_graft(u_int32_t src, u_int32_t dst, char *p, int datalen) 1121 { 1122 vifi_t vifi; 1123 u_int32_t graft_src; 1124 u_int32_t graft_grp; 1125 int i; 1126 struct rtentry *r; 1127 struct gtable *g; 1128 struct ptable *pt, **ptnp; 1129 1130 if ((vifi = find_vif(src, dst)) == NO_VIF) { 1131 logit(LOG_INFO, 0, 1132 "ignoring graft from non-neighbor %s", 1133 inet_fmt(src, s1)); 1134 return; 1135 } 1136 1137 if (datalen < 8) { 1138 logit(LOG_WARNING, 0, 1139 "received non-decipherable graft from %s", 1140 inet_fmt(src, s1)); 1141 return; 1142 } 1143 1144 for (i = 0; i< 4; i++) 1145 ((char *)&graft_src)[i] = *p++; 1146 for (i = 0; i< 4; i++) 1147 ((char *)&graft_grp)[i] = *p++; 1148 1149 logit(LOG_DEBUG, 0, "%s on vif %d grafts (%s %s)", 1150 inet_fmt(src, s1), vifi, 1151 inet_fmt(graft_src, s2), inet_fmt(graft_grp, s3)); 1152 1153 /* 1154 * Find the subnet for the graft 1155 */ 1156 if (find_src_grp(graft_src, 0, graft_grp)) { 1157 g = gtp ? gtp->gt_gnext : kernel_table; 1158 r = g->gt_route; 1159 1160 if (VIFM_ISSET(vifi, g->gt_scope)) { 1161 logit(LOG_WARNING, 0, "graft received from %s on scoped grp (%s %s)", 1162 inet_fmt(src, s1), inet_fmt(graft_src, s2), 1163 inet_fmt(graft_grp, s3)); 1164 return; 1165 } 1166 1167 ptnp = &g->gt_pruntbl; 1168 while ((pt = *ptnp) != NULL) { 1169 if ((pt->pt_vifi == vifi) && (pt->pt_router == src)) { 1170 *ptnp = pt->pt_next; 1171 free(pt); 1172 1173 VIFM_SET(vifi, g->gt_grpmems); 1174 logit(LOG_DEBUG, 0, "accept graft (%s %s) gm:%x", 1175 inet_fmts(r->rt_origin, r->rt_originmask, s1), 1176 inet_fmt(g->gt_mcastgrp, s2), g->gt_grpmems); 1177 1178 prun_add_ttls(g); 1179 update_kernel(g); 1180 #ifdef RSRR 1181 /* Send route change notification to reservation protocol. */ 1182 rsrr_cache_send(g,1); 1183 #endif /* RSRR */ 1184 break; 1185 } else { 1186 ptnp = &pt->pt_next; 1187 } 1188 } 1189 1190 /* send ack downstream */ 1191 send_graft_ack(dst, src, graft_src, graft_grp); 1192 g->gt_timer = max_prune_lifetime; 1193 1194 if (g->gt_prsent_timer) { 1195 /* set the flag for graft retransmission */ 1196 g->gt_grftsnt = 1; 1197 1198 /* send graft upwards */ 1199 send_graft(g); 1200 1201 /* reset the prune sent timer */ 1202 g->gt_prsent_timer = 0; 1203 } 1204 } else { 1205 /* 1206 * We have no state for the source and group in question. 1207 * We can simply acknowledge the graft, since we know 1208 * that we have no prune state, and grafts are requests 1209 * to remove prune state. 1210 */ 1211 send_graft_ack(dst, src, graft_src, graft_grp); 1212 logit(LOG_DEBUG, 0, "%s (%s %s) from %s", 1213 "graft received with no kernel entry for", 1214 inet_fmt(graft_src, s1), inet_fmt(graft_grp, s2), 1215 inet_fmt(src, s3)); 1216 return; 1217 } 1218 } 1219 1220 /* 1221 * find out which group is involved first of all 1222 * then determine if a graft was sent. 1223 * if no graft sent, ignore the message 1224 * if graft was sent and the ack is from the right 1225 * source, remove the graft timer so that we don't 1226 * have send a graft again 1227 */ 1228 void 1229 accept_g_ack(u_int32_t src, u_int32_t dst, char *p, int datalen) 1230 { 1231 struct gtable *g; 1232 vifi_t vifi; 1233 u_int32_t grft_src; 1234 u_int32_t grft_grp; 1235 int i; 1236 1237 if ((vifi = find_vif(src, dst)) == NO_VIF) { 1238 logit(LOG_INFO, 0, 1239 "ignoring graft ack from non-neighbor %s", 1240 inet_fmt(src, s1)); 1241 return; 1242 } 1243 1244 if (datalen < 0 || datalen > 8) { 1245 logit(LOG_WARNING, 0, 1246 "received non-decipherable graft ack from %s", 1247 inet_fmt(src, s1)); 1248 return; 1249 } 1250 1251 for (i = 0; i< 4; i++) 1252 ((char *)&grft_src)[i] = *p++; 1253 for (i = 0; i< 4; i++) 1254 ((char *)&grft_grp)[i] = *p++; 1255 1256 logit(LOG_DEBUG, 0, "%s on vif %d acks graft (%s, %s)", 1257 inet_fmt(src, s1), vifi, 1258 inet_fmt(grft_src, s2), inet_fmt(grft_grp, s3)); 1259 1260 /* 1261 * Find the subnet for the graft ack 1262 */ 1263 if (find_src_grp(grft_src, 0, grft_grp)) { 1264 g = gtp ? gtp->gt_gnext : kernel_table; 1265 g->gt_grftsnt = 0; 1266 } else { 1267 logit(LOG_WARNING, 0, "%s (%s, %s) from %s", 1268 "rcvd graft ack with no kernel entry for", 1269 inet_fmt(grft_src, s1), inet_fmt(grft_grp, s2), 1270 inet_fmt(src, s3)); 1271 return; 1272 } 1273 } 1274 1275 1276 /* 1277 * free all prune entries and kernel routes 1278 * normally, this should inform the kernel that all of its routes 1279 * are going away, but this is only called by restart(), which is 1280 * about to call MRT_DONE which does that anyway. 1281 */ 1282 void 1283 free_all_prunes(void) 1284 { 1285 struct rtentry *r; 1286 struct gtable *g, *prev_g; 1287 struct stable *s, *prev_s; 1288 struct ptable *p, *prev_p; 1289 1290 for (r = routing_table; r; r = r->rt_next) { 1291 g = r->rt_groups; 1292 while (g) { 1293 s = g->gt_srctbl; 1294 while (s) { 1295 prev_s = s; 1296 s = s->st_next; 1297 free(prev_s); 1298 } 1299 1300 p = g->gt_pruntbl; 1301 while (p) { 1302 prev_p = p; 1303 p = p->pt_next; 1304 free(prev_p); 1305 } 1306 1307 prev_g = g; 1308 g = g->gt_next; 1309 free(prev_g); 1310 } 1311 r->rt_groups = NULL; 1312 } 1313 kernel_table = NULL; 1314 1315 g = kernel_no_route; 1316 while (g) { 1317 free(g->gt_srctbl); 1318 1319 prev_g = g; 1320 g = g->gt_next; 1321 free(prev_g); 1322 } 1323 kernel_no_route = NULL; 1324 } 1325 1326 /* 1327 * When a new route is created, search 1328 * a) The less-specific part of the routing table 1329 * b) The route-less kernel table 1330 * for sources that the new route might want to handle. 1331 * 1332 * "Inheriting" these sources might be cleanest, but simply deleting 1333 * them is easier, and letting the kernel re-request them. 1334 */ 1335 void 1336 steal_sources(struct rtentry *rt) 1337 { 1338 struct rtentry *rp; 1339 struct gtable *gt, **gtnp; 1340 struct stable *st, **stnp; 1341 1342 for (rp = rt->rt_next; rp; rp = rp->rt_next) { 1343 if ((rt->rt_origin & rp->rt_originmask) == rp->rt_origin) { 1344 logit(LOG_DEBUG, 0, "Route for %s stealing sources from %s", 1345 inet_fmts(rt->rt_origin, rt->rt_originmask, s1), 1346 inet_fmts(rp->rt_origin, rp->rt_originmask, s2)); 1347 for (gt = rp->rt_groups; gt; gt = gt->gt_next) { 1348 stnp = >->gt_srctbl; 1349 while ((st = *stnp) != NULL) { 1350 if ((st->st_origin & rt->rt_originmask) == rt->rt_origin) { 1351 logit(LOG_DEBUG, 0, "%s stealing (%s %s) from %s", 1352 inet_fmts(rt->rt_origin, rt->rt_originmask, s1), 1353 inet_fmt(st->st_origin, s3), 1354 inet_fmt(gt->gt_mcastgrp, s4), 1355 inet_fmts(rp->rt_origin, rp->rt_originmask, s2)); 1356 if (k_del_rg(st->st_origin, gt) < 0) { 1357 logit(LOG_WARNING, errno, "%s (%s, %s)", 1358 "steal_sources trying to delete", 1359 inet_fmt(st->st_origin, s1), 1360 inet_fmt(gt->gt_mcastgrp, s2)); 1361 } 1362 *stnp = st->st_next; 1363 kroutes--; 1364 free(st); 1365 } else { 1366 stnp = &st->st_next; 1367 } 1368 } 1369 } 1370 } 1371 } 1372 1373 gtnp = &kernel_no_route; 1374 while ((gt = *gtnp) != NULL) { 1375 if (gt->gt_srctbl && ((gt->gt_srctbl->st_origin & rt->rt_originmask) 1376 == rt->rt_origin)) { 1377 logit(LOG_DEBUG, 0, "%s stealing (%s %s) from %s", 1378 inet_fmts(rt->rt_origin, rt->rt_originmask, s1), 1379 inet_fmt(gt->gt_srctbl->st_origin, s3), 1380 inet_fmt(gt->gt_mcastgrp, s4), 1381 "no_route table"); 1382 if (k_del_rg(gt->gt_srctbl->st_origin, gt) < 0) { 1383 logit(LOG_WARNING, errno, "%s (%s %s)", 1384 "steal_sources trying to delete", 1385 inet_fmt(gt->gt_srctbl->st_origin, s1), 1386 inet_fmt(gt->gt_mcastgrp, s2)); 1387 } 1388 kroutes--; 1389 free(gt->gt_srctbl); 1390 *gtnp = gt->gt_next; 1391 if (gt->gt_next) 1392 gt->gt_next->gt_prev = gt->gt_prev; 1393 free(gt); 1394 } else { 1395 gtnp = >->gt_next; 1396 } 1397 } 1398 } 1399 1400 /* 1401 * Advance the timers on all the cache entries. 1402 * If there are any entries whose timers have expired, 1403 * remove these entries from the kernel cache. 1404 */ 1405 void 1406 age_table_entry(void) 1407 { 1408 struct rtentry *r; 1409 struct gtable *gt, **gtnptr; 1410 struct stable *st, **stnp; 1411 struct ptable *pt, **ptnp; 1412 struct sioc_sg_req sg_req; 1413 1414 logit(LOG_DEBUG, 0, "ageing entries"); 1415 1416 gtnptr = &kernel_table; 1417 while ((gt = *gtnptr) != NULL) { 1418 r = gt->gt_route; 1419 1420 /* advance the timer for the kernel entry */ 1421 gt->gt_timer -= ROUTE_MAX_REPORT_DELAY; 1422 1423 /* decrement prune timer if need be */ 1424 if (gt->gt_prsent_timer > 0) { 1425 gt->gt_prsent_timer -= ROUTE_MAX_REPORT_DELAY; 1426 if (gt->gt_prsent_timer <= 0) { 1427 logit(LOG_DEBUG, 0, "upstream prune tmo (%s %s)", 1428 inet_fmts(r->rt_origin, r->rt_originmask, s1), 1429 inet_fmt(gt->gt_mcastgrp, s2)); 1430 gt->gt_prsent_timer = -1; 1431 } 1432 } 1433 1434 /* retransmit graft if graft sent flag is still set */ 1435 if (gt->gt_grftsnt) { 1436 int y; 1437 CHK_GS(gt->gt_grftsnt++, y); 1438 if (y) 1439 send_graft(gt); 1440 } 1441 1442 /* 1443 * Age prunes 1444 * 1445 * If a prune expires, forward again on that vif. 1446 */ 1447 ptnp = >->gt_pruntbl; 1448 while ((pt = *ptnp) != NULL) { 1449 if ((pt->pt_timer -= ROUTE_MAX_REPORT_DELAY) <= 0) { 1450 logit(LOG_DEBUG, 0, "expire prune (%s %s) from %s on vif %d", 1451 inet_fmts(r->rt_origin, r->rt_originmask, s1), 1452 inet_fmt(gt->gt_mcastgrp, s2), 1453 inet_fmt(pt->pt_router, s3), 1454 pt->pt_vifi); 1455 1456 expire_prune(pt->pt_vifi, gt); 1457 1458 /* remove the router's prune entry and await new one */ 1459 *ptnp = pt->pt_next; 1460 free(pt); 1461 } else { 1462 ptnp = &pt->pt_next; 1463 } 1464 } 1465 1466 /* 1467 * If the cache entry has expired, delete source table entries for 1468 * silent sources. If there are no source entries left, and there 1469 * are no downstream prunes, then the entry is deleted. 1470 * Otherwise, the cache entry's timer is refreshed. 1471 */ 1472 if (gt->gt_timer <= 0) { 1473 /* Check for traffic before deleting source entries */ 1474 sg_req.grp.s_addr = gt->gt_mcastgrp; 1475 stnp = >->gt_srctbl; 1476 while ((st = *stnp) != NULL) { 1477 sg_req.src.s_addr = st->st_origin; 1478 if (ioctl(udp_socket, SIOCGETSGCNT, (char *)&sg_req) < 0) { 1479 logit(LOG_WARNING, errno, "%s (%s %s)", 1480 "age_table_entry: SIOCGETSGCNT failing for", 1481 inet_fmt(st->st_origin, s1), 1482 inet_fmt(gt->gt_mcastgrp, s2)); 1483 /* Make sure it gets deleted below */ 1484 sg_req.pktcnt = st->st_pktcnt; 1485 } 1486 if (sg_req.pktcnt == st->st_pktcnt) { 1487 *stnp = st->st_next; 1488 logit(LOG_DEBUG, 0, "age_table_entry deleting (%s %s)", 1489 inet_fmt(st->st_origin, s1), 1490 inet_fmt(gt->gt_mcastgrp, s2)); 1491 if (k_del_rg(st->st_origin, gt) < 0) { 1492 logit(LOG_WARNING, errno, 1493 "age_table_entry trying to delete (%s %s)", 1494 inet_fmt(st->st_origin, s1), 1495 inet_fmt(gt->gt_mcastgrp, s2)); 1496 } 1497 kroutes--; 1498 free(st); 1499 } else { 1500 st->st_pktcnt = sg_req.pktcnt; 1501 stnp = &st->st_next; 1502 } 1503 } 1504 1505 /* 1506 * Retain the group entry if we have downstream prunes or if 1507 * there is at least one source in the list that still has 1508 * traffic, or if our upstream prune timer is running. 1509 */ 1510 if (gt->gt_pruntbl != NULL || gt->gt_srctbl != NULL || 1511 gt->gt_prsent_timer > 0) { 1512 gt->gt_timer = CACHE_LIFETIME(cache_lifetime); 1513 if (gt->gt_prsent_timer == -1) 1514 if (gt->gt_grpmems == 0) 1515 send_prune(gt); 1516 else 1517 gt->gt_prsent_timer = 0; 1518 gtnptr = >->gt_gnext; 1519 continue; 1520 } 1521 1522 logit(LOG_DEBUG, 0, "timeout cache entry (%s, %s)", 1523 inet_fmts(r->rt_origin, r->rt_originmask, s1), 1524 inet_fmt(gt->gt_mcastgrp, s2)); 1525 1526 if (gt->gt_prev) 1527 gt->gt_prev->gt_next = gt->gt_next; 1528 else 1529 gt->gt_route->rt_groups = gt->gt_next; 1530 if (gt->gt_next) 1531 gt->gt_next->gt_prev = gt->gt_prev; 1532 1533 if (gt->gt_gprev) { 1534 gt->gt_gprev->gt_gnext = gt->gt_gnext; 1535 gtnptr = >->gt_gprev->gt_gnext; 1536 } else { 1537 kernel_table = gt->gt_gnext; 1538 gtnptr = &kernel_table; 1539 } 1540 if (gt->gt_gnext) 1541 gt->gt_gnext->gt_gprev = gt->gt_gprev; 1542 1543 #ifdef RSRR 1544 /* Send route change notification to reservation protocol. */ 1545 rsrr_cache_send(gt,0); 1546 rsrr_cache_clean(gt); 1547 #endif /* RSRR */ 1548 free((char *)gt); 1549 } else { 1550 if (gt->gt_prsent_timer == -1) 1551 if (gt->gt_grpmems == 0) 1552 send_prune(gt); 1553 else 1554 gt->gt_prsent_timer = 0; 1555 gtnptr = >->gt_gnext; 1556 } 1557 } 1558 1559 /* 1560 * When traversing the no_route table, the decision is much easier. 1561 * Just delete it if it has timed out. 1562 */ 1563 gtnptr = &kernel_no_route; 1564 while ((gt = *gtnptr) != NULL) { 1565 /* advance the timer for the kernel entry */ 1566 gt->gt_timer -= ROUTE_MAX_REPORT_DELAY; 1567 1568 if (gt->gt_timer < 0) { 1569 if (gt->gt_srctbl) { 1570 if (k_del_rg(gt->gt_srctbl->st_origin, gt) < 0) { 1571 logit(LOG_WARNING, errno, "%s (%s %s)", 1572 "age_table_entry trying to delete no-route", 1573 inet_fmt(gt->gt_srctbl->st_origin, s1), 1574 inet_fmt(gt->gt_mcastgrp, s2)); 1575 } 1576 free(gt->gt_srctbl); 1577 } 1578 *gtnptr = gt->gt_next; 1579 if (gt->gt_next) 1580 gt->gt_next->gt_prev = gt->gt_prev; 1581 1582 free((char *)gt); 1583 } else { 1584 gtnptr = >->gt_next; 1585 } 1586 } 1587 } 1588 1589 /* 1590 * Modify the kernel to forward packets when one or multiple prunes that 1591 * were received on the vif given by vifi, for the group given by gt, 1592 * have expired. 1593 */ 1594 static void 1595 expire_prune(vifi_t vifi, struct gtable *gt) 1596 { 1597 /* 1598 * No need to send a graft, any prunes that we sent 1599 * will expire before any prunes that we have received. 1600 */ 1601 if (gt->gt_prsent_timer > 0) { 1602 logit(LOG_DEBUG, 0, "prune expired with %d left on %s", 1603 gt->gt_prsent_timer, "prsent_timer"); 1604 gt->gt_prsent_timer = 0; 1605 } 1606 1607 /* modify the kernel entry to forward packets */ 1608 if (!VIFM_ISSET(vifi, gt->gt_grpmems)) { 1609 struct rtentry *rt = gt->gt_route; 1610 VIFM_SET(vifi, gt->gt_grpmems); 1611 logit(LOG_DEBUG, 0, "forw again (%s %s) gm:%x vif:%d", 1612 inet_fmts(rt->rt_origin, rt->rt_originmask, s1), 1613 inet_fmt(gt->gt_mcastgrp, s2), gt->gt_grpmems, vifi); 1614 1615 prun_add_ttls(gt); 1616 update_kernel(gt); 1617 #ifdef RSRR 1618 /* Send route change notification to reservation protocol. */ 1619 rsrr_cache_send(gt,1); 1620 #endif /* RSRR */ 1621 } 1622 } 1623 1624 1625 static char * 1626 scaletime(time_t t) 1627 { 1628 static char buf1[5]; 1629 static char buf2[5]; 1630 static char *buf=buf1; 1631 char s; 1632 char *p; 1633 1634 p = buf; 1635 if (buf == buf1) 1636 buf = buf2; 1637 else 1638 buf = buf1; 1639 1640 if (t < 120) { 1641 s = 's'; 1642 } else if (t < 3600) { 1643 t /= 60; 1644 s = 'm'; 1645 } else if (t < 86400) { 1646 t /= 3600; 1647 s = 'h'; 1648 } else if (t < 864000) { 1649 t /= 86400; 1650 s = 'd'; 1651 } else { 1652 t /= 604800; 1653 s = 'w'; 1654 } 1655 if (t > 999) 1656 return "*** "; 1657 1658 snprintf(p, 5, "%3d%c", (int)t, s); 1659 1660 return p; 1661 } 1662 1663 /* 1664 * Print the contents of the cache table on file 'fp2'. 1665 */ 1666 void 1667 dump_cache(FILE *fp2) 1668 { 1669 struct rtentry *r; 1670 struct gtable *gt; 1671 struct stable *st; 1672 vifi_t i; 1673 time_t thyme = time(NULL); 1674 1675 fprintf(fp2, 1676 "Multicast Routing Cache Table (%d entries)\n%s", kroutes, 1677 " Origin Mcast-group CTmr Age Ptmr IVif Forwvifs\n"); 1678 1679 for (gt = kernel_no_route; gt; gt = gt->gt_next) { 1680 if (gt->gt_srctbl) { 1681 fprintf(fp2, " %-18s %-15s %-4s %-4s - -1\n", 1682 inet_fmts(gt->gt_srctbl->st_origin, 0xffffffff, s1), 1683 inet_fmt(gt->gt_mcastgrp, s2), scaletime(gt->gt_timer), 1684 scaletime(thyme - gt->gt_ctime)); 1685 fprintf(fp2, ">%s\n", inet_fmt(gt->gt_srctbl->st_origin, s1)); 1686 } 1687 } 1688 1689 for (gt = kernel_table; gt; gt = gt->gt_gnext) { 1690 r = gt->gt_route; 1691 fprintf(fp2, " %-18s %-15s", 1692 inet_fmts(r->rt_origin, r->rt_originmask, s1), 1693 inet_fmt(gt->gt_mcastgrp, s2)); 1694 1695 fprintf(fp2, " %-4s", scaletime(gt->gt_timer)); 1696 1697 fprintf(fp2, " %-4s %-4s ", scaletime(thyme - gt->gt_ctime), 1698 gt->gt_prsent_timer ? scaletime(gt->gt_prsent_timer) : 1699 " -"); 1700 1701 fprintf(fp2, "%2u%c%c ", r->rt_parent, 1702 gt->gt_prsent_timer ? 'P' : ' ', 1703 VIFM_ISSET(r->rt_parent, gt->gt_scope) ? 'B' : ' '); 1704 1705 for (i = 0; i < numvifs; ++i) { 1706 if (VIFM_ISSET(i, gt->gt_grpmems)) 1707 fprintf(fp2, " %u ", i); 1708 else if (VIFM_ISSET(i, r->rt_children) && 1709 !VIFM_ISSET(i, r->rt_leaves)) 1710 fprintf(fp2, " %u%c", i, 1711 VIFM_ISSET(i, gt->gt_scope) ? 'b' : 'p'); 1712 } 1713 fprintf(fp2, "\n"); 1714 for (st = gt->gt_srctbl; st; st = st->st_next) { 1715 fprintf(fp2, ">%s\n", inet_fmt(st->st_origin, s1)); 1716 } 1717 #ifdef DEBUG_PRUNES 1718 for (pt = gt->gt_pruntbl; pt; pt = pt->pt_next) { 1719 fprintf(fp2, "<r:%s v:%d t:%d\n", inet_fmt(pt->pt_router, s1), 1720 pt->pt_vifi, pt->pt_timer); 1721 } 1722 #endif 1723 } 1724 } 1725 1726 /* 1727 * Traceroute function which returns traceroute replies to the requesting 1728 * router. Also forwards the request to downstream routers. 1729 * NOTE: u_int no is narrowed to u_char 1730 */ 1731 void 1732 accept_mtrace(u_int32_t src, u_int32_t dst, u_int32_t group, 1733 char *data, u_int no, int datalen) 1734 { 1735 u_char type; 1736 struct rtentry *rt; 1737 struct gtable *gt; 1738 struct tr_query *qry; 1739 struct tr_resp *resp; 1740 int vifi; 1741 char *p; 1742 int rcount; 1743 int errcode = TR_NO_ERR; 1744 int resptype; 1745 struct timeval tp; 1746 struct sioc_vif_req v_req; 1747 struct sioc_sg_req sg_req; 1748 1749 /* Remember qid across invocations */ 1750 static u_int32_t oqid = 0; 1751 1752 /* timestamp the request/response */ 1753 gettimeofday(&tp, 0); 1754 1755 /* 1756 * Check if it is a query or a response 1757 */ 1758 if (datalen == QLEN) { 1759 type = QUERY; 1760 logit(LOG_DEBUG, 0, "Initial traceroute query rcvd from %s to %s", 1761 inet_fmt(src, s1), inet_fmt(dst, s2)); 1762 } 1763 else if ((datalen - QLEN) % RLEN == 0) { 1764 type = RESP; 1765 logit(LOG_DEBUG, 0, "In-transit traceroute query rcvd from %s to %s", 1766 inet_fmt(src, s1), inet_fmt(dst, s2)); 1767 if (IN_MULTICAST(ntohl(dst))) { 1768 logit(LOG_DEBUG, 0, "Dropping multicast response"); 1769 return; 1770 } 1771 } 1772 else { 1773 logit(LOG_WARNING, 0, "%s from %s to %s", 1774 "Non decipherable traceroute request received", 1775 inet_fmt(src, s1), inet_fmt(dst, s2)); 1776 return; 1777 } 1778 1779 qry = (struct tr_query *)data; 1780 1781 /* 1782 * if it is a packet with all reports filled, drop it 1783 */ 1784 if ((rcount = (datalen - QLEN)/RLEN) == no) { 1785 logit(LOG_DEBUG, 0, "packet with all reports filled in"); 1786 return; 1787 } 1788 1789 logit(LOG_DEBUG, 0, "s: %s g: %s d: %s ", inet_fmt(qry->tr_src, s1), 1790 inet_fmt(group, s2), inet_fmt(qry->tr_dst, s3)); 1791 logit(LOG_DEBUG, 0, "rttl: %d rd: %s", qry->tr_rttl, 1792 inet_fmt(qry->tr_raddr, s1)); 1793 logit(LOG_DEBUG, 0, "rcount:%d, qid:%06x", rcount, qry->tr_qid); 1794 1795 /* determine the routing table entry for this traceroute */ 1796 rt = determine_route(qry->tr_src); 1797 if (rt) { 1798 logit(LOG_DEBUG, 0, "rt parent vif: %d rtr: %s metric: %d", 1799 rt->rt_parent, inet_fmt(rt->rt_gateway, s1), rt->rt_metric); 1800 logit(LOG_DEBUG, 0, "rt origin %s", 1801 inet_fmts(rt->rt_origin, rt->rt_originmask, s1)); 1802 } else 1803 logit(LOG_DEBUG, 0, "...no route"); 1804 1805 /* 1806 * Query type packet - check if rte exists 1807 * Check if the query destination is a vif connected to me. 1808 * and if so, whether I should start response back 1809 */ 1810 if (type == QUERY) { 1811 if (oqid == qry->tr_qid) { 1812 /* 1813 * If the multicast router is a member of the group being 1814 * queried, and the query is multicasted, then the router can 1815 * receive multiple copies of the same query. If we have already 1816 * replied to this traceroute, just ignore it this time. 1817 * 1818 * This is not a total solution, but since if this fails you 1819 * only get N copies, N <= the number of interfaces on the router, 1820 * it is not fatal. 1821 */ 1822 logit(LOG_DEBUG, 0, "ignoring duplicate traceroute packet"); 1823 return; 1824 } 1825 1826 if (rt == NULL) { 1827 logit(LOG_DEBUG, 0, "Mcast traceroute: no route entry %s", 1828 inet_fmt(qry->tr_src, s1)); 1829 if (IN_MULTICAST(ntohl(dst))) 1830 return; 1831 } 1832 vifi = find_vif(qry->tr_dst, 0); 1833 1834 if (vifi == NO_VIF) { 1835 /* The traceroute destination is not on one of my subnet vifs. */ 1836 logit(LOG_DEBUG, 0, "Destination %s not an interface", 1837 inet_fmt(qry->tr_dst, s1)); 1838 if (IN_MULTICAST(ntohl(dst))) 1839 return; 1840 errcode = TR_WRONG_IF; 1841 } else if (rt != NULL && !VIFM_ISSET(vifi, rt->rt_children)) { 1842 logit(LOG_DEBUG, 0, "Destination %s not on forwarding tree for src %s", 1843 inet_fmt(qry->tr_dst, s1), inet_fmt(qry->tr_src, s2)); 1844 if (IN_MULTICAST(ntohl(dst))) 1845 return; 1846 errcode = TR_WRONG_IF; 1847 } 1848 } 1849 else { 1850 /* 1851 * determine which interface the packet came in on 1852 * RESP packets travel hop-by-hop so this either traversed 1853 * a tunnel or came from a directly attached mrouter. 1854 */ 1855 if ((vifi = find_vif(src, dst)) == NO_VIF) { 1856 logit(LOG_DEBUG, 0, "Wrong interface for packet"); 1857 errcode = TR_WRONG_IF; 1858 } 1859 } 1860 1861 /* Now that we've decided to send a response, save the qid */ 1862 oqid = qry->tr_qid; 1863 1864 logit(LOG_DEBUG, 0, "Sending traceroute response"); 1865 1866 /* copy the packet to the sending buffer */ 1867 p = send_buf + MIN_IP_HEADER_LEN + IGMP_MINLEN; 1868 1869 bcopy(data, p, datalen); 1870 1871 p += datalen; 1872 1873 /* 1874 * If there is no room to insert our reply, coopt the previous hop 1875 * error indication to relay this fact. 1876 */ 1877 if (p + sizeof(struct tr_resp) > send_buf + RECV_BUF_SIZE) { 1878 resp = (struct tr_resp *)p - 1; 1879 resp->tr_rflags = TR_NO_SPACE; 1880 rt = NULL; 1881 goto sendit; 1882 } 1883 1884 /* 1885 * fill in initial response fields 1886 */ 1887 resp = (struct tr_resp *)p; 1888 bzero(resp, sizeof(struct tr_resp)); 1889 datalen += RLEN; 1890 1891 resp->tr_qarr = htonl((tp.tv_sec + JAN_1970) << 16) + 1892 ((tp.tv_usec >> 4) & 0xffff); 1893 1894 resp->tr_rproto = PROTO_DVMRP; 1895 if (errcode != TR_NO_ERR) { 1896 resp->tr_rflags = errcode; 1897 rt = NULL; /* hack to enforce send straight to requestor */ 1898 goto sendit; 1899 } 1900 resp->tr_outaddr = uvifs[vifi].uv_lcl_addr; 1901 resp->tr_fttl = uvifs[vifi].uv_threshold; 1902 resp->tr_rflags = TR_NO_ERR; 1903 1904 /* 1905 * obtain # of packets out on interface 1906 */ 1907 v_req.vifi = vifi; 1908 if (ioctl(udp_socket, SIOCGETVIFCNT, (char *)&v_req) >= 0) 1909 resp->tr_vifout = htonl(v_req.ocount); 1910 1911 /* 1912 * fill in scoping & pruning information 1913 */ 1914 if (rt) 1915 for (gt = rt->rt_groups; gt; gt = gt->gt_next) { 1916 if (gt->gt_mcastgrp >= group) 1917 break; 1918 } 1919 else 1920 gt = NULL; 1921 1922 if (gt && gt->gt_mcastgrp == group) { 1923 sg_req.src.s_addr = qry->tr_src; 1924 sg_req.grp.s_addr = group; 1925 if (ioctl(udp_socket, SIOCGETSGCNT, (char *)&sg_req) >= 0) 1926 resp->tr_pktcnt = htonl(sg_req.pktcnt); 1927 1928 if (VIFM_ISSET(vifi, gt->gt_scope)) 1929 resp->tr_rflags = TR_SCOPED; 1930 else if (gt->gt_prsent_timer) 1931 resp->tr_rflags = TR_PRUNED; 1932 else if (!VIFM_ISSET(vifi, gt->gt_grpmems)) 1933 if (VIFM_ISSET(vifi, rt->rt_children) && 1934 !VIFM_ISSET(vifi, rt->rt_leaves)) 1935 resp->tr_rflags = TR_OPRUNED; 1936 else 1937 resp->tr_rflags = TR_NO_FWD; 1938 } else { 1939 if (scoped_addr(vifi, group)) 1940 resp->tr_rflags = TR_SCOPED; 1941 else if (rt && !VIFM_ISSET(vifi, rt->rt_children)) 1942 resp->tr_rflags = TR_NO_FWD; 1943 } 1944 1945 /* 1946 * if no rte exists, set NO_RTE error 1947 */ 1948 if (rt == NULL) { 1949 src = dst; /* the dst address of resp. pkt */ 1950 resp->tr_inaddr = 0; 1951 resp->tr_rflags = TR_NO_RTE; 1952 resp->tr_rmtaddr = 0; 1953 } else { 1954 /* get # of packets in on interface */ 1955 v_req.vifi = rt->rt_parent; 1956 if (ioctl(udp_socket, SIOCGETVIFCNT, (char *)&v_req) >= 0) 1957 resp->tr_vifin = htonl(v_req.icount); 1958 1959 MASK_TO_VAL(rt->rt_originmask, resp->tr_smask); 1960 src = uvifs[rt->rt_parent].uv_lcl_addr; 1961 resp->tr_inaddr = src; 1962 resp->tr_rmtaddr = rt->rt_gateway; 1963 if (!VIFM_ISSET(vifi, rt->rt_children)) { 1964 logit(LOG_DEBUG, 0, "Destination %s not on forwarding tree for src %s", 1965 inet_fmt(qry->tr_dst, s1), inet_fmt(qry->tr_src, s2)); 1966 resp->tr_rflags = TR_WRONG_IF; 1967 } 1968 if (rt->rt_metric >= UNREACHABLE) { 1969 resp->tr_rflags = TR_NO_RTE; 1970 /* Hack to send reply directly */ 1971 rt = NULL; 1972 } 1973 } 1974 1975 sendit: 1976 /* 1977 * if metric is 1 or no. of reports is 1, send response to requestor 1978 * else send to upstream router. If the upstream router can't handle 1979 * mtrace, set an error code and send to requestor anyway. 1980 */ 1981 logit(LOG_DEBUG, 0, "rcount:%d, no:%d", rcount, no); 1982 1983 if ((rcount + 1 == no) || (rt == NULL) || (rt->rt_metric == 1)) { 1984 resptype = IGMP_MTRACE_REPLY; 1985 dst = qry->tr_raddr; 1986 } else 1987 if (!can_mtrace(rt->rt_parent, rt->rt_gateway)) { 1988 dst = qry->tr_raddr; 1989 resp->tr_rflags = TR_OLD_ROUTER; 1990 resptype = IGMP_MTRACE_REPLY; 1991 } else { 1992 dst = rt->rt_gateway; 1993 resptype = IGMP_MTRACE_QUERY; 1994 } 1995 1996 if (IN_MULTICAST(ntohl(dst))) { 1997 /* 1998 * Send the reply on a known multicast capable vif. 1999 * If we don't have one, we can't source any multicasts anyway. 2000 */ 2001 if (phys_vif != -1) { 2002 logit(LOG_DEBUG, 0, "Sending reply to %s from %s", 2003 inet_fmt(dst, s1), inet_fmt(uvifs[phys_vif].uv_lcl_addr, s2)); 2004 k_set_ttl(qry->tr_rttl); 2005 send_igmp(uvifs[phys_vif].uv_lcl_addr, dst, 2006 resptype, no, group, 2007 datalen); 2008 k_set_ttl(1); 2009 } else 2010 logit(LOG_INFO, 0, "No enabled phyints -- %s", 2011 "dropping traceroute reply"); 2012 } else { 2013 logit(LOG_DEBUG, 0, "Sending %s to %s from %s", 2014 resptype == IGMP_MTRACE_REPLY ? "reply" : "request on", 2015 inet_fmt(dst, s1), inet_fmt(src, s2)); 2016 2017 send_igmp(src, dst, 2018 resptype, no, group, 2019 datalen); 2020 } 2021 return; 2022 } 2023