xref: /dragonfly/sbin/routed/output.c (revision 1765f996)
1 /*
2  * Copyright (c) 1983, 1988, 1993
3  *	The Regents of the University of California.  All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  * 1. Redistributions of source code must retain the above copyright
9  *    notice, this list of conditions and the following disclaimer.
10  * 2. Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in the
12  *    documentation and/or other materials provided with the distribution.
13  * 3. Neither the name of the University nor the names of its contributors
14  *    may be used to endorse or promote products derived from this software
15  *    without specific prior written permission.
16  *
17  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
18  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
19  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
20  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
21  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
22  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
23  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
24  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
25  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
26  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
27  * SUCH DAMAGE.
28  *
29  * @(#)output.c	8.1 (Berkeley) 6/5/93
30  * $FreeBSD: src/sbin/routed/output.c,v 1.5.2.1 2000/08/14 17:00:03 sheldonh Exp $
31  */
32 
33 #include "defs.h"
34 
35 u_int update_seqno;
36 
37 
38 /* walk the tree of routes with this for output
39  */
40 struct {
41 	struct sockaddr_in to;
42 	naddr	to_mask;
43 	naddr	to_net;
44 	naddr	to_std_mask;
45 	naddr	to_std_net;
46 	struct interface *ifp;		/* usually output interface */
47 	struct auth *a;
48 	char	metric;			/* adjust metrics by interface */
49 	int	npackets;
50 	int	gen_limit;
51 	u_int	state;
52 #define	    WS_ST_FLASH	    0x001	/* send only changed routes */
53 #define	    WS_ST_RIP2_ALL  0x002	/* send full featured RIPv2 */
54 #define	    WS_ST_AG	    0x004	/* ok to aggregate subnets */
55 #define	    WS_ST_SUPER_AG  0x008	/* ok to aggregate networks */
56 #define	    WS_ST_QUERY	    0x010	/* responding to a query */
57 #define	    WS_ST_TO_ON_NET 0x020	/* sending onto one of our nets */
58 #define	    WS_ST_DEFAULT   0x040	/* faking a default */
59 } ws;
60 
61 /* A buffer for what can be heard by both RIPv1 and RIPv2 listeners */
62 struct ws_buf v12buf;
63 union pkt_buf ripv12_buf;
64 
65 /* Another for only RIPv2 listeners */
66 struct ws_buf v2buf;
67 union pkt_buf rip_v2_buf;
68 
69 
70 
71 void
72 bufinit(void)
73 {
74 	ripv12_buf.rip.rip_cmd = RIPCMD_RESPONSE;
75 	v12buf.buf = &ripv12_buf.rip;
76 	v12buf.base = &v12buf.buf->rip_nets[0];
77 
78 	rip_v2_buf.rip.rip_cmd = RIPCMD_RESPONSE;
79 	rip_v2_buf.rip.rip_vers = RIPv2;
80 	v2buf.buf = &rip_v2_buf.rip;
81 	v2buf.base = &v2buf.buf->rip_nets[0];
82 }
83 
84 
85 /* Send the contents of the global buffer via the non-multicast socket
86  */
87 int					/* <0 on failure */
88 output(enum output_type type,
89        struct sockaddr_in *dst,		/* send to here */
90        struct interface *ifp,
91        struct rip *buf,
92        int size)			/* this many bytes */
93 {
94 	struct sockaddr_in in;
95 	int flags;
96 	const char *msg;
97 	int res;
98 	naddr tgt_mcast;
99 	int soc;
100 	int serrno;
101 
102 	in = *dst;
103 	if (in.sin_port == 0)
104 		in.sin_port = htons(RIP_PORT);
105 #ifdef _HAVE_SIN_LEN
106 	if (in.sin_len == 0)
107 		in.sin_len = sizeof(in);
108 #endif
109 
110 	soc = rip_sock;
111 	flags = 0;
112 
113 	switch (type) {
114 	case OUT_QUERY:
115 		msg = "Answer Query";
116 		if (soc < 0)
117 			soc = ifp->int_rip_sock;
118 		break;
119 	case OUT_UNICAST:
120 		msg = "Send";
121 		if (soc < 0)
122 			soc = ifp->int_rip_sock;
123 		flags = MSG_DONTROUTE;
124 		break;
125 	case OUT_BROADCAST:
126 		if (ifp->int_if_flags & IFF_POINTOPOINT) {
127 			msg = "Send";
128 		} else {
129 			msg = "Send bcast";
130 		}
131 		flags = MSG_DONTROUTE;
132 		break;
133 	case OUT_MULTICAST:
134 		if (ifp->int_if_flags & IFF_POINTOPOINT) {
135 			msg = "Send pt-to-pt";
136 		} else if (ifp->int_state & IS_DUP) {
137 			trace_act("abort multicast output via %s"
138 				  " with duplicate address",
139 				  ifp->int_name);
140 			return 0;
141 		} else {
142 			msg = "Send mcast";
143 			if (rip_sock_mcast != ifp) {
144 #ifdef MCAST_PPP_BUG
145 				/* Do not specify the primary interface
146 				 * explicitly if we have the multicast
147 				 * point-to-point kernel bug, since the
148 				 * kernel will do the wrong thing if the
149 				 * local address of a point-to-point link
150 				 * is the same as the address of an ordinary
151 				 * interface.
152 				 */
153 				if (ifp->int_addr == myaddr) {
154 					tgt_mcast = 0;
155 				} else
156 #endif
157 				tgt_mcast = ifp->int_addr;
158 				if (0 > setsockopt(rip_sock,
159 						   IPPROTO_IP, IP_MULTICAST_IF,
160 						   &tgt_mcast,
161 						   sizeof(tgt_mcast))) {
162 					serrno = errno;
163 					LOGERR("setsockopt(rip_sock,"
164 					       "IP_MULTICAST_IF)");
165 					errno = serrno;
166 					ifp = NULL;
167 					return -1;
168 				}
169 				rip_sock_mcast = ifp;
170 			}
171 			in.sin_addr.s_addr = htonl(INADDR_RIP_GROUP);
172 		}
173 		break;
174 
175 	case NO_OUT_MULTICAST:
176 	case NO_OUT_RIPV2:
177 	default:
178 #ifdef DEBUG
179 		abort();
180 #endif
181 		return -1;
182 	}
183 
184 	trace_rip(msg, "to", &in, ifp, buf, size);
185 
186 	res = sendto(soc, buf, size, flags,
187 		     (struct sockaddr *)&in, sizeof(in));
188 	if (res < 0
189 	    && (ifp == NULL || !(ifp->int_state & IS_BROKE))) {
190 		serrno = errno;
191 		msglog("%s sendto(%s%s%s.%d): %s", msg,
192 		       ifp != NULL ? ifp->int_name : "",
193 		       ifp != NULL ? ", " : "",
194 		       inet_ntoa(in.sin_addr),
195 		       ntohs(in.sin_port),
196 		       strerror(errno));
197 		errno = serrno;
198 	}
199 
200 	return res;
201 }
202 
203 
204 /* Find the first key for a packet to send.
205  * Try for a key that is eligible and has not expired, but settle for
206  * the last key if they have all expired.
207  * If no key is ready yet, give up.
208  */
209 struct auth *
210 find_auth(struct interface *ifp)
211 {
212 	struct auth *ap, *res;
213 	int i;
214 
215 
216 	if (ifp == NULL)
217 		return 0;
218 
219 	res = NULL;
220 	ap = ifp->int_auth;
221 	for (i = 0; i < MAX_AUTH_KEYS; i++, ap++) {
222 		/* stop looking after the last key */
223 		if (ap->type == RIP_AUTH_NONE)
224 			break;
225 
226 		/* ignore keys that are not ready yet */
227 		if ((u_long)ap->start > (u_long)clk.tv_sec)
228 			continue;
229 
230 		if ((u_long)ap->end < (u_long)clk.tv_sec) {
231 			/* note best expired password as a fall-back */
232 			if (res == NULL || (u_long)ap->end > (u_long)res->end)
233 				res = ap;
234 			continue;
235 		}
236 
237 		/* note key with the best future */
238 		if (res == NULL || (u_long)res->end < (u_long)ap->end)
239 			res = ap;
240 	}
241 	return res;
242 }
243 
244 
245 void
246 clr_ws_buf(struct ws_buf *wb,
247 	   struct auth *ap)
248 {
249 	struct netauth *na;
250 
251 	wb->lim = wb->base + NETS_LEN;
252 	wb->n = wb->base;
253 	memset(wb->n, 0, NETS_LEN*sizeof(*wb->n));
254 
255 	/* (start to) install authentication if appropriate
256 	 */
257 	if (ap == NULL)
258 		return;
259 
260 	na = (struct netauth*)wb->n;
261 	if (ap->type == RIP_AUTH_PW) {
262 		na->a_family = RIP_AF_AUTH;
263 		na->a_type = RIP_AUTH_PW;
264 		memcpy(na->au.au_pw, ap->key, sizeof(na->au.au_pw));
265 		wb->n++;
266 
267 	} else if (ap->type ==  RIP_AUTH_MD5) {
268 		na->a_family = RIP_AF_AUTH;
269 		na->a_type = RIP_AUTH_MD5;
270 		na->au.a_md5.md5_keyid = ap->keyid;
271 		na->au.a_md5.md5_auth_len = RIP_AUTH_MD5_LEN;
272 		na->au.a_md5.md5_seqno = htonl(clk.tv_sec);
273 		wb->n++;
274 		wb->lim--;		/* make room for trailer */
275 	}
276 }
277 
278 
279 void
280 end_md5_auth(struct ws_buf *wb,
281 	     struct auth *ap)
282 {
283 	struct netauth *na, *na2;
284 	MD5_CTX md5_ctx;
285 	int len;
286 
287 
288 	na = (struct netauth*)wb->base;
289 	na2 = (struct netauth*)wb->n;
290 	len = (char *)na2-(char *)wb->buf;
291 	na2->a_family = RIP_AF_AUTH;
292 	na2->a_type = htons(1);
293 	na->au.a_md5.md5_pkt_len = htons(len);
294 	MD5Init(&md5_ctx);
295 	MD5Update(&md5_ctx, (u_char *)wb->buf, len);
296 	MD5Update(&md5_ctx, ap->key, RIP_AUTH_MD5_LEN);
297 	MD5Final(na2->au.au_pw, &md5_ctx);
298 	wb->n++;
299 }
300 
301 
302 /* Send the buffer
303  */
304 static void
305 supply_write(struct ws_buf *wb)
306 {
307 	/* Output multicast only if legal.
308 	 * If we would multicast and it would be illegal, then discard the
309 	 * packet.
310 	 */
311 	switch (wb->type) {
312 	case NO_OUT_MULTICAST:
313 		trace_pkt("skip multicast to %s because impossible",
314 			  naddr_ntoa(ws.to.sin_addr.s_addr));
315 		break;
316 	case NO_OUT_RIPV2:
317 		break;
318 	default:
319 		if (ws.a != NULL && ws.a->type == RIP_AUTH_MD5)
320 			end_md5_auth(wb,ws.a);
321 		if (output(wb->type, &ws.to, ws.ifp, wb->buf,
322 			   ((char *)wb->n - (char*)wb->buf)) < 0
323 		    && ws.ifp != NULL)
324 			if_sick(ws.ifp);
325 		ws.npackets++;
326 		break;
327 	}
328 
329 	clr_ws_buf(wb,ws.a);
330 }
331 
332 
333 /* put an entry into the packet
334  */
335 static void
336 supply_out(struct ag_info *ag)
337 {
338 	int i;
339 	naddr mask, v1_mask, dst_h, ddst_h = 0;
340 	struct ws_buf *wb;
341 
342 
343 	/* Skip this route if doing a flash update and it and the routes
344 	 * it aggregates have not changed recently.
345 	 */
346 	if (ag->ag_seqno < update_seqno
347 	    && (ws.state & WS_ST_FLASH))
348 		return;
349 
350 	dst_h = ag->ag_dst_h;
351 	mask = ag->ag_mask;
352 	v1_mask = ripv1_mask_host(htonl(dst_h),
353 				  (ws.state & WS_ST_TO_ON_NET) ? ws.ifp : 0);
354 	i = 0;
355 
356 	/* If we are sending RIPv2 packets that cannot (or must not) be
357 	 * heard by RIPv1 listeners, do not worry about sub- or supernets.
358 	 * Subnets (from other networks) can only be sent via multicast.
359 	 * A pair of subnet routes might have been promoted so that they
360 	 * are legal to send by RIPv1.
361 	 * If RIPv1 is off, use the multicast buffer.
362 	 */
363 	if ((ws.state & WS_ST_RIP2_ALL)
364 	    || ((ag->ag_state & AGS_RIPV2) && v1_mask != mask)) {
365 		/* use the RIPv2-only buffer */
366 		wb = &v2buf;
367 
368 	} else {
369 		/* use the RIPv1-or-RIPv2 buffer */
370 		wb = &v12buf;
371 
372 		/* Convert supernet route into corresponding set of network
373 		 * routes for RIPv1, but leave non-contiguous netmasks
374 		 * to ag_check().
375 		 */
376 		if (v1_mask > mask
377 		    && mask + (mask & -mask) == 0) {
378 			ddst_h = v1_mask & -v1_mask;
379 			i = (v1_mask & ~mask)/ddst_h;
380 
381 			if (i > ws.gen_limit) {
382 				/* Punt if we would have to generate an
383 				 * unreasonable number of routes.
384 				 */
385 				if (TRACECONTENTS)
386 					trace_misc("sending %s-->%s as 1"
387 						   " instead of %d routes",
388 						   addrname(htonl(dst_h), mask,
389 							1),
390 						   naddr_ntoa(ws.to.sin_addr
391 							.s_addr),
392 						   i+1);
393 				i = 0;
394 
395 			} else {
396 				mask = v1_mask;
397 				ws.gen_limit -= i;
398 			}
399 		}
400 	}
401 
402 	do {
403 		wb->n->n_family = RIP_AF_INET;
404 		wb->n->n_dst = htonl(dst_h);
405 		/* If the route is from router-discovery or we are
406 		 * shutting down, admit only a bad metric.
407 		 */
408 		wb->n->n_metric = ((stopint || ag->ag_metric < 1)
409 				   ? HOPCNT_INFINITY
410 				   : ag->ag_metric);
411 		wb->n->n_metric = htonl(wb->n->n_metric);
412 		/* Any non-zero bits in the supposedly unused RIPv1 fields
413 		 * cause the old `routed` to ignore the route.
414 		 * That means the mask and so forth cannot be sent
415 		 * in the hybrid RIPv1/RIPv2 mode.
416 		 */
417 		if (ws.state & WS_ST_RIP2_ALL) {
418 			if (ag->ag_nhop != 0
419 			    && ((ws.state & WS_ST_QUERY)
420 				|| (ag->ag_nhop != ws.ifp->int_addr
421 				    && on_net(ag->ag_nhop,
422 					      ws.ifp->int_net,
423 					      ws.ifp->int_mask))))
424 				wb->n->n_nhop = ag->ag_nhop;
425 			wb->n->n_mask = htonl(mask);
426 			wb->n->n_tag = ag->ag_tag;
427 		}
428 		dst_h += ddst_h;
429 
430 		if (++wb->n >= wb->lim)
431 			supply_write(wb);
432 	} while (i-- != 0);
433 }
434 
435 
436 /* supply one route from the table
437  */
438 /* ARGSUSED */
439 static int
440 walk_supply(struct radix_node *rn, __unused struct walkarg *argp)
441 {
442 #define RT ((struct rt_entry *)rn)
443 	u_short ags;
444 	char metric, pref;
445 	naddr dst, nhop;
446 	struct rt_spare *rts;
447 	int i;
448 
449 
450 	/* Do not advertise external remote interfaces or passive interfaces.
451 	 */
452 	if ((RT->rt_state & RS_IF)
453 	    && RT->rt_ifp != 0
454 	    && (RT->rt_ifp->int_state & IS_PASSIVE)
455 	    && !(RT->rt_state & RS_MHOME))
456 		return 0;
457 
458 	/* If being quiet about our ability to forward, then
459 	 * do not say anything unless responding to a query,
460 	 * except about our main interface.
461 	 */
462 	if (!supplier && !(ws.state & WS_ST_QUERY)
463 	    && !(RT->rt_state & RS_MHOME))
464 		return 0;
465 
466 	dst = RT->rt_dst;
467 
468 	/* do not collide with the fake default route */
469 	if (dst == RIP_DEFAULT
470 	    && (ws.state & WS_ST_DEFAULT))
471 		return 0;
472 
473 	if (RT->rt_state & RS_NET_SYN) {
474 		if (RT->rt_state & RS_NET_INT) {
475 			/* Do not send manual synthetic network routes
476 			 * into the subnet.
477 			 */
478 			if (on_net(ws.to.sin_addr.s_addr,
479 				   ntohl(dst), RT->rt_mask))
480 				return 0;
481 
482 		} else {
483 			/* Do not send automatic synthetic network routes
484 			 * if they are not needed because no RIPv1 listeners
485 			 * can hear them.
486 			 */
487 			if (ws.state & WS_ST_RIP2_ALL)
488 				return 0;
489 
490 			/* Do not send automatic synthetic network routes to
491 			 * the real subnet.
492 			 */
493 			if (on_net(ws.to.sin_addr.s_addr,
494 				   ntohl(dst), RT->rt_mask))
495 				return 0;
496 		}
497 		nhop = 0;
498 
499 	} else {
500 		/* Advertise the next hop if this is not a route for one
501 		 * of our interfaces and the next hop is on the same
502 		 * network as the target.
503 		 * The final determination is made by supply_out().
504 		 */
505 		if (!(RT->rt_state & RS_IF)
506 		    && RT->rt_gate != myaddr
507 		    && RT->rt_gate != loopaddr)
508 			nhop = RT->rt_gate;
509 		else
510 			nhop = 0;
511 	}
512 
513 	metric = RT->rt_metric;
514 	ags = 0;
515 
516 	if (RT->rt_state & RS_MHOME) {
517 		/* retain host route of multi-homed servers */
518 		;
519 
520 	} else if (RT_ISHOST(RT)) {
521 		/* We should always suppress (into existing network routes)
522 		 * the host routes for the local end of our point-to-point
523 		 * links.
524 		 * If we are suppressing host routes in general, then do so.
525 		 * Avoid advertising host routes onto their own network,
526 		 * where they should be handled by proxy-ARP.
527 		 */
528 		if ((RT->rt_state & RS_LOCAL)
529 		    || ridhosts
530 		    || on_net(dst, ws.to_net, ws.to_mask))
531 			ags |= AGS_SUPPRESS;
532 
533 		/* Aggregate stray host routes into network routes if allowed.
534 		 * We cannot aggregate host routes into small network routes
535 		 * without confusing RIPv1 listeners into thinking the
536 		 * network routes are host routes.
537 		 */
538 		if ((ws.state & WS_ST_AG)
539 		    && !(ws.state & WS_ST_RIP2_ALL))
540 			ags |= AGS_AGGREGATE;
541 
542 	} else {
543 		/* Always suppress network routes into other, existing
544 		 * network routes
545 		 */
546 		ags |= AGS_SUPPRESS;
547 
548 		/* Generate supernets if allowed.
549 		 * If we can be heard by RIPv1 systems, we will
550 		 * later convert back to ordinary nets.
551 		 * This unifies dealing with received supernets.
552 		 */
553 		if ((ws.state & WS_ST_AG)
554 		    && ((RT->rt_state & RS_SUBNET)
555 			|| (ws.state & WS_ST_SUPER_AG)))
556 			ags |= AGS_AGGREGATE;
557 	}
558 
559 	/* Do not send RIPv1 advertisements of subnets to other
560 	 * networks. If possible, multicast them by RIPv2.
561 	 */
562 	if ((RT->rt_state & RS_SUBNET)
563 	    && !(ws.state & WS_ST_RIP2_ALL)
564 	    && !on_net(dst, ws.to_std_net, ws.to_std_mask))
565 		ags |= AGS_RIPV2 | AGS_AGGREGATE;
566 
567 
568 	/* Do not send a route back to where it came from, except in
569 	 * response to a query.  This is "split-horizon".  That means not
570 	 * advertising back to the same network	and so via the same interface.
571 	 *
572 	 * We want to suppress routes that might have been fragmented
573 	 * from this route by a RIPv1 router and sent back to us, and so we
574 	 * cannot forget this route here.  Let the split-horizon route
575 	 * suppress the fragmented routes and then itself be forgotten.
576 	 *
577 	 * Include the routes for both ends of point-to-point interfaces
578 	 * among those suppressed by split-horizon, since the other side
579 	 * should knows them as well as we do.
580 	 *
581 	 * Notice spare routes with the same metric that we are about to
582 	 * advertise, to split the horizon on redundant, inactive paths.
583 	 */
584 	if (ws.ifp != 0
585 	    && !(ws.state & WS_ST_QUERY)
586 	    && (ws.state & WS_ST_TO_ON_NET)
587 	    && (!(RT->rt_state & RS_IF)
588 		|| ws.ifp->int_if_flags & IFF_POINTOPOINT)) {
589 		for (rts = RT->rt_spares, i = NUM_SPARES; i != 0; i--, rts++) {
590 			if (rts->rts_metric > metric
591 			    || rts->rts_ifp != ws.ifp)
592 				continue;
593 
594 			/* If we do not mark the route with AGS_SPLIT_HZ here,
595 			 * it will be poisoned-reverse, or advertised back
596 			 * toward its source with an infinite metric.
597 			 * If we have recently advertised the route with a
598 			 * better metric than we now have, then we should
599 			 * poison-reverse the route before suppressing it for
600 			 * split-horizon.
601 			 *
602 			 * In almost all cases, if there is no spare for the
603 			 * route then it is either old and dead or a brand
604 			 * new route. If it is brand new, there is no need
605 			 * for poison-reverse. If it is old and dead, it
606 			 * is already poisoned.
607 			 */
608 			if (RT->rt_poison_time < now_expire
609 			    || RT->rt_poison_metric >= metric
610 			    || RT->rt_spares[1].rts_gate == 0) {
611 				ags |= AGS_SPLIT_HZ;
612 				ags &= ~AGS_SUPPRESS;
613 			}
614 			metric = HOPCNT_INFINITY;
615 			break;
616 		}
617 	}
618 
619 	/* Keep track of the best metric with which the
620 	 * route has been advertised recently.
621 	 */
622 	if (RT->rt_poison_metric >= metric
623 	    || RT->rt_poison_time < now_expire) {
624 		RT->rt_poison_time = now.tv_sec;
625 		RT->rt_poison_metric = metric;
626 	}
627 
628 	/* Adjust the outgoing metric by the cost of the link.
629 	 * Avoid aggregation when a route is counting to infinity.
630 	 */
631 	pref = RT->rt_poison_metric + ws.metric;
632 	metric += ws.metric;
633 
634 	/* Do not advertise stable routes that will be ignored,
635 	 * unless we are answering a query.
636 	 * If the route recently was advertised with a metric that
637 	 * would have been less than infinity through this interface,
638 	 * we need to continue to advertise it in order to poison it.
639 	 */
640 	if (metric >= HOPCNT_INFINITY) {
641 		if (!(ws.state & WS_ST_QUERY)
642 		    && (pref >= HOPCNT_INFINITY
643 			|| RT->rt_poison_time < now_garbage))
644 			return 0;
645 
646 		metric = HOPCNT_INFINITY;
647 	}
648 
649 	ag_check(dst, RT->rt_mask, 0, nhop, metric, pref,
650 		 RT->rt_seqno, RT->rt_tag, ags, supply_out);
651 	return 0;
652 #undef RT
653 }
654 
655 
656 /* Supply dst with the contents of the routing tables.
657  * If this won't fit in one packet, chop it up into several.
658  */
659 void
660 supply(struct sockaddr_in *dst,
661        struct interface *ifp,		/* output interface */
662        enum output_type type,
663        int flash,			/* 1=flash update */
664        int vers,			/* RIP version */
665        int passwd_ok)			/* OK to include cleartext password */
666 {
667 	struct rt_entry *rt;
668 	int def_metric;
669 
670 
671 	ws.state = 0;
672 	ws.gen_limit = 1024;
673 
674 	ws.to = *dst;
675 	ws.to_std_mask = std_mask(ws.to.sin_addr.s_addr);
676 	ws.to_std_net = ntohl(ws.to.sin_addr.s_addr) & ws.to_std_mask;
677 
678 	if (ifp != NULL) {
679 		ws.to_mask = ifp->int_mask;
680 		ws.to_net = ifp->int_net;
681 		if (on_net(ws.to.sin_addr.s_addr, ws.to_net, ws.to_mask))
682 			ws.state |= WS_ST_TO_ON_NET;
683 
684 	} else {
685 		ws.to_mask = ripv1_mask_net(ws.to.sin_addr.s_addr, 0);
686 		ws.to_net = ntohl(ws.to.sin_addr.s_addr) & ws.to_mask;
687 		rt = rtfind(dst->sin_addr.s_addr);
688 		if (rt)
689 			ifp = rt->rt_ifp;
690 	}
691 
692 	ws.npackets = 0;
693 	if (flash)
694 		ws.state |= WS_ST_FLASH;
695 
696 	if ((ws.ifp = ifp) == NULL) {
697 		ws.metric = 1;
698 	} else {
699 		/* Adjust the advertised metric by the outgoing interface
700 		 * metric.
701 		 */
702 		ws.metric = ifp->int_metric+1;
703 	}
704 
705 	ripv12_buf.rip.rip_vers = vers;
706 
707 	switch (type) {
708 	case OUT_MULTICAST:
709 		if (ifp->int_if_flags & IFF_MULTICAST)
710 			v2buf.type = OUT_MULTICAST;
711 		else
712 			v2buf.type = NO_OUT_MULTICAST;
713 		v12buf.type = OUT_BROADCAST;
714 		break;
715 
716 	case OUT_QUERY:
717 		ws.state |= WS_ST_QUERY;
718 		/* fall through */
719 	case OUT_BROADCAST:
720 	case OUT_UNICAST:
721 		v2buf.type = (vers == RIPv2) ? type : NO_OUT_RIPV2;
722 		v12buf.type = type;
723 		break;
724 
725 	case NO_OUT_MULTICAST:
726 	case NO_OUT_RIPV2:
727 		break;			/* no output */
728 	}
729 
730 	if (vers == RIPv2) {
731 		/* full RIPv2 only if cannot be heard by RIPv1 listeners */
732 		if (type != OUT_BROADCAST)
733 			ws.state |= WS_ST_RIP2_ALL;
734 		if ((ws.state & WS_ST_QUERY)
735 		    || !(ws.state & WS_ST_TO_ON_NET)) {
736 			ws.state |= (WS_ST_AG | WS_ST_SUPER_AG);
737 		} else if (ifp == NULL || !(ifp->int_state & IS_NO_AG)) {
738 			ws.state |= WS_ST_AG;
739 			if (type != OUT_BROADCAST
740 			    && (ifp == NULL
741 				|| !(ifp->int_state & IS_NO_SUPER_AG)))
742 				ws.state |= WS_ST_SUPER_AG;
743 		}
744 	}
745 
746 	ws.a = (vers == RIPv2) ? find_auth(ifp) : 0;
747 	if (!passwd_ok && ws.a != NULL && ws.a->type == RIP_AUTH_PW)
748 		ws.a = NULL;
749 	clr_ws_buf(&v12buf,ws.a);
750 	clr_ws_buf(&v2buf,ws.a);
751 
752 	/*  Fake a default route if asked and if there is not already
753 	 * a better, real default route.
754 	 */
755 	if (supplier && (def_metric = ifp->int_d_metric) != 0) {
756 		if (NULL == (rt = rtget(RIP_DEFAULT, 0))
757 		    || rt->rt_metric+ws.metric >= def_metric) {
758 			ws.state |= WS_ST_DEFAULT;
759 			ag_check(0, 0, 0, 0, def_metric, def_metric,
760 				 0, 0, 0, supply_out);
761 		} else {
762 			def_metric = rt->rt_metric+ws.metric;
763 		}
764 
765 		/* If both RIPv2 and the poor-man's router discovery
766 		 * kludge are on, arrange to advertise an extra
767 		 * default route via RIPv1.
768 		 */
769 		if ((ws.state & WS_ST_RIP2_ALL)
770 		    && (ifp->int_state & IS_PM_RDISC)) {
771 			ripv12_buf.rip.rip_vers = RIPv1;
772 			v12buf.n->n_family = RIP_AF_INET;
773 			v12buf.n->n_dst = htonl(RIP_DEFAULT);
774 			v12buf.n->n_metric = htonl(def_metric);
775 			v12buf.n++;
776 		}
777 	}
778 
779 	rn_walktree(rhead, walk_supply, 0);
780 	ag_flush(0,0,supply_out);
781 
782 	/* Flush the packet buffers, provided they are not empty and
783 	 * do not contain only the password.
784 	 */
785 	if (v12buf.n != v12buf.base
786 	    && (v12buf.n > v12buf.base+1
787 		|| v12buf.base->n_family != RIP_AF_AUTH))
788 		supply_write(&v12buf);
789 	if (v2buf.n != v2buf.base
790 	    && (v2buf.n > v2buf.base+1
791 		|| v2buf.base->n_family != RIP_AF_AUTH))
792 		supply_write(&v2buf);
793 
794 	/* If we sent nothing and this is an answer to a query, send
795 	 * an empty buffer.
796 	 */
797 	if (ws.npackets == 0
798 	    && (ws.state & WS_ST_QUERY))
799 		supply_write(&v12buf);
800 }
801 
802 
803 /* send all of the routing table or just do a flash update
804  */
805 void
806 rip_bcast(int flash)
807 {
808 #ifdef _HAVE_SIN_LEN
809 	static struct sockaddr_in dst = {sizeof(dst), AF_INET, 0, {0}, {0}};
810 #else
811 	static struct sockaddr_in dst = {AF_INET};
812 #endif
813 	struct interface *ifp;
814 	enum output_type type;
815 	int vers;
816 	struct timeval rtime;
817 
818 
819 	need_flash = 0;
820 	intvl_random(&rtime, MIN_WAITTIME, MAX_WAITTIME);
821 	no_flash = rtime;
822 	timevaladd(&no_flash, &now);
823 
824 	if (rip_sock < 0)
825 		return;
826 
827 	trace_act("send %s and inhibit dynamic updates for %.3f sec",
828 		  flash ? "dynamic update" : "all routes",
829 		  rtime.tv_sec + ((float)rtime.tv_usec)/1000000.0);
830 
831 	for (ifp = ifnet; ifp != NULL; ifp = ifp->int_next) {
832 		/* Skip interfaces not doing RIP.
833 		 * Do try broken interfaces to see if they have healed.
834 		 */
835 		if (IS_RIP_OUT_OFF(ifp->int_state))
836 			continue;
837 
838 		/* skip turned off interfaces */
839 		if (!iff_up(ifp->int_if_flags))
840 			continue;
841 
842 		vers = (ifp->int_state & IS_NO_RIPV1_OUT) ? RIPv2 : RIPv1;
843 
844 		if (ifp->int_if_flags & IFF_BROADCAST) {
845 			/* ordinary, hardware interface */
846 			dst.sin_addr.s_addr = ifp->int_brdaddr;
847 
848 			if (vers == RIPv2
849 			    && !(ifp->int_state  & IS_NO_RIP_MCAST)) {
850 				type = OUT_MULTICAST;
851 			} else {
852 				type = OUT_BROADCAST;
853 			}
854 
855 		} else if (ifp->int_if_flags & IFF_POINTOPOINT) {
856 			/* point-to-point hardware interface */
857 			dst.sin_addr.s_addr = ifp->int_dstaddr;
858 			type = OUT_UNICAST;
859 
860 		} else if (ifp->int_state & IS_REMOTE) {
861 			/* remote interface */
862 			dst.sin_addr.s_addr = ifp->int_addr;
863 			type = OUT_UNICAST;
864 
865 		} else {
866 			/* ATM, HIPPI, etc. */
867 			continue;
868 		}
869 
870 		supply(&dst, ifp, type, flash, vers, 1);
871 	}
872 
873 	update_seqno++;			/* all routes are up to date */
874 }
875 
876 
877 /* Ask for routes
878  * Do it only once to an interface, and not even after the interface
879  * was broken and recovered.
880  */
881 void
882 rip_query(void)
883 {
884 #ifdef _HAVE_SIN_LEN
885 	static struct sockaddr_in dst = {sizeof(dst), AF_INET, 0, {0}, {0}};
886 #else
887 	static struct sockaddr_in dst = {AF_INET};
888 #endif
889 	struct interface *ifp;
890 	struct rip buf;
891 	enum output_type type;
892 
893 
894 	if (rip_sock < 0)
895 		return;
896 
897 	memset(&buf, 0, sizeof(buf));
898 
899 	for (ifp = ifnet; ifp; ifp = ifp->int_next) {
900 		/* Skip interfaces those already queried.
901 		 * Do not ask via interfaces through which we don't
902 		 * accept input.  Do not ask via interfaces that cannot
903 		 * send RIP packets.
904 		 * Do try broken interfaces to see if they have healed.
905 		 */
906 		if (IS_RIP_IN_OFF(ifp->int_state)
907 		    || ifp->int_query_time != NEVER)
908 			continue;
909 
910 		/* skip turned off interfaces */
911 		if (!iff_up(ifp->int_if_flags))
912 			continue;
913 
914 		buf.rip_vers = (ifp->int_state&IS_NO_RIPV1_OUT) ? RIPv2:RIPv1;
915 		buf.rip_cmd = RIPCMD_REQUEST;
916 		buf.rip_nets[0].n_family = RIP_AF_UNSPEC;
917 		buf.rip_nets[0].n_metric = htonl(HOPCNT_INFINITY);
918 
919 		/* Send a RIPv1 query only if allowed and if we will
920 		 * listen to RIPv1 routers.
921 		 */
922 		if ((ifp->int_state & IS_NO_RIPV1_OUT)
923 		    || (ifp->int_state & IS_NO_RIPV1_IN)) {
924 			buf.rip_vers = RIPv2;
925 		} else {
926 			buf.rip_vers = RIPv1;
927 		}
928 
929 		if (ifp->int_if_flags & IFF_BROADCAST) {
930 			/* ordinary, hardware interface */
931 			dst.sin_addr.s_addr = ifp->int_brdaddr;
932 
933 			/* Broadcast RIPv1 queries and RIPv2 queries
934 			 * when the hardware cannot multicast.
935 			 */
936 			if (buf.rip_vers == RIPv2
937 			    && (ifp->int_if_flags & IFF_MULTICAST)
938 			    && !(ifp->int_state  & IS_NO_RIP_MCAST)) {
939 				type = OUT_MULTICAST;
940 			} else {
941 				type = OUT_BROADCAST;
942 			}
943 
944 		} else if (ifp->int_if_flags & IFF_POINTOPOINT) {
945 			/* point-to-point hardware interface */
946 			dst.sin_addr.s_addr = ifp->int_dstaddr;
947 			type = OUT_UNICAST;
948 
949 		} else if (ifp->int_state & IS_REMOTE) {
950 			/* remote interface */
951 			dst.sin_addr.s_addr = ifp->int_addr;
952 			type = OUT_UNICAST;
953 
954 		} else {
955 			/* ATM, HIPPI, etc. */
956 			continue;
957 		}
958 
959 		ifp->int_query_time = now.tv_sec+SUPPLY_INTERVAL;
960 		if (output(type, &dst, ifp, &buf, sizeof(buf)) < 0)
961 			if_sick(ifp);
962 	}
963 }
964