]> git.ipfire.org Git - thirdparty/bird.git/blob - nest/rt-table.c
Whitespace fixes
[thirdparty/bird.git] / nest / rt-table.c
1 /*
2 * BIRD -- Routing Tables
3 *
4 * (c) 1998--2000 Martin Mares <mj@ucw.cz>
5 *
6 * Can be freely distributed and used under the terms of the GNU GPL.
7 */
8
9 /**
10 * DOC: Routing tables
11 *
12 * Routing tables are probably the most important structures BIRD uses. They
13 * hold all the information about known networks, the associated routes and
14 * their attributes.
15 *
16 * There are multiple routing tables (a primary one together with any
17 * number of secondary ones if requested by the configuration). Each table
18 * is basically a FIB containing entries describing the individual
19 * destination networks. For each network (represented by structure &net),
20 * there is a one-way linked list of route entries (&rte), the first entry
21 * on the list being the best one (i.e., the one we currently use
22 * for routing), the order of the other ones is undetermined.
23 *
24 * The &rte contains information specific to the route (preference, protocol
25 * metrics, time of last modification etc.) and a pointer to a &rta structure
26 * (see the route attribute module for a precise explanation) holding the
27 * remaining route attributes which are expected to be shared by multiple
28 * routes in order to conserve memory.
29 */
30
31 #undef LOCAL_DEBUG
32
33 #include "nest/bird.h"
34 #include "nest/route.h"
35 #include "nest/protocol.h"
36 #include "nest/cli.h"
37 #include "nest/iface.h"
38 #include "lib/resource.h"
39 #include "lib/event.h"
40 #include "lib/string.h"
41 #include "conf/conf.h"
42 #include "filter/filter.h"
43 #include "lib/string.h"
44 #include "lib/alloca.h"
45
46 pool *rt_table_pool;
47
48 static slab *rte_slab;
49 static linpool *rte_update_pool;
50
51 static list routing_tables;
52
53 static byte *rt_format_via(rte *e);
54 static void rt_free_hostcache(rtable *tab);
55 static void rt_notify_hostcache(rtable *tab, net *net);
56 static void rt_update_hostcache(rtable *tab);
57 static void rt_next_hop_update(rtable *tab);
58 static inline int rt_prune_table(rtable *tab);
59 static inline void rt_schedule_gc(rtable *tab);
60 static inline void rt_schedule_prune(rtable *tab);
61
62
63 static inline struct ea_list *
64 make_tmp_attrs(struct rte *rt, struct linpool *pool)
65 {
66 struct ea_list *(*mta)(struct rte *rt, struct linpool *pool);
67 mta = rt->attrs->src->proto->make_tmp_attrs;
68 return mta ? mta(rt, rte_update_pool) : NULL;
69 }
70
71 /* Like fib_route(), but skips empty net entries */
72 static net *
73 net_route(rtable *tab, ip_addr a, int len)
74 {
75 ip_addr a0;
76 net *n;
77
78 while (len >= 0)
79 {
80 a0 = ipa_and(a, ipa_mkmask(len));
81 n = fib_find(&tab->fib, &a0, len);
82 if (n && rte_is_valid(n->routes))
83 return n;
84 len--;
85 }
86 return NULL;
87 }
88
89 static void
90 rte_init(struct fib_node *N)
91 {
92 net *n = (net *) N;
93
94 N->flags = 0;
95 n->routes = NULL;
96 }
97
98 /**
99 * rte_find - find a route
100 * @net: network node
101 * @src: route source
102 *
103 * The rte_find() function returns a route for destination @net
104 * which is from route source @src.
105 */
106 rte *
107 rte_find(net *net, struct rte_src *src)
108 {
109 rte *e = net->routes;
110
111 while (e && e->attrs->src != src)
112 e = e->next;
113 return e;
114 }
115
116 /**
117 * rte_get_temp - get a temporary &rte
118 * @a: attributes to assign to the new route (a &rta; in case it's
119 * un-cached, rte_update() will create a cached copy automatically)
120 *
121 * Create a temporary &rte and bind it with the attributes @a.
122 * Also set route preference to the default preference set for
123 * the protocol.
124 */
125 rte *
126 rte_get_temp(rta *a)
127 {
128 rte *e = sl_alloc(rte_slab);
129
130 e->attrs = a;
131 e->flags = 0;
132 e->pref = a->src->proto->preference;
133 return e;
134 }
135
136 rte *
137 rte_do_cow(rte *r)
138 {
139 rte *e = sl_alloc(rte_slab);
140
141 memcpy(e, r, sizeof(rte));
142 e->attrs = rta_clone(r->attrs);
143 e->flags = 0;
144 return e;
145 }
146
147 /**
148 * rte_cow_rta - get a private writable copy of &rte with writable &rta
149 * @r: a route entry to be copied
150 * @lp: a linpool from which to allocate &rta
151 *
152 * rte_cow_rta() takes a &rte and prepares it and associated &rta for
153 * modification. There are three possibilities: First, both &rte and &rta are
154 * private copies, in that case they are returned unchanged. Second, &rte is
155 * private copy, but &rta is cached, in that case &rta is duplicated using
156 * rta_do_cow(). Third, both &rte is shared and &rta is cached, in that case
157 * both structures are duplicated by rte_do_cow() and rta_do_cow().
158 *
159 * Note that in the second case, cached &rta loses one reference, while private
160 * copy created by rta_do_cow() is a shallow copy sharing indirect data (eattrs,
161 * nexthops, ...) with it. To work properly, original shared &rta should have
162 * another reference during the life of created private copy.
163 *
164 * Result: a pointer to the new writable &rte with writable &rta.
165 */
166 rte *
167 rte_cow_rta(rte *r, linpool *lp)
168 {
169 if (!rta_is_cached(r->attrs))
170 return r;
171
172 rte *e = rte_cow(r);
173 rta *a = rta_do_cow(r->attrs, lp);
174 rta_free(e->attrs);
175 e->attrs = a;
176 return e;
177 }
178
179 static int /* Actually better or at least as good as */
180 rte_better(rte *new, rte *old)
181 {
182 int (*better)(rte *, rte *);
183
184 if (!rte_is_valid(old))
185 return 1;
186 if (!rte_is_valid(new))
187 return 0;
188
189 if (new->pref > old->pref)
190 return 1;
191 if (new->pref < old->pref)
192 return 0;
193 if (new->attrs->src->proto->proto != old->attrs->src->proto->proto)
194 {
195 /*
196 * If the user has configured protocol preferences, so that two different protocols
197 * have the same preference, try to break the tie by comparing addresses. Not too
198 * useful, but keeps the ordering of routes unambiguous.
199 */
200 return new->attrs->src->proto->proto > old->attrs->src->proto->proto;
201 }
202 if (better = new->attrs->src->proto->rte_better)
203 return better(new, old);
204 return 0;
205 }
206
207 static int
208 rte_mergable(rte *pri, rte *sec)
209 {
210 int (*mergable)(rte *, rte *);
211
212 if (!rte_is_valid(pri) || !rte_is_valid(sec))
213 return 0;
214
215 if (pri->pref != sec->pref)
216 return 0;
217
218 if (pri->attrs->src->proto->proto != sec->attrs->src->proto->proto)
219 return 0;
220
221 if (mergable = pri->attrs->src->proto->rte_mergable)
222 return mergable(pri, sec);
223
224 return 0;
225 }
226
227 static void
228 rte_trace(struct proto *p, rte *e, int dir, char *msg)
229 {
230 log(L_TRACE "%s %c %s %I/%d %s", p->name, dir, msg, e->net->n.prefix, e->net->n.pxlen, rt_format_via(e));
231 }
232
233 static inline void
234 rte_trace_in(uint flag, struct proto *p, rte *e, char *msg)
235 {
236 if (p->debug & flag)
237 rte_trace(p, e, '>', msg);
238 }
239
240 static inline void
241 rte_trace_out(uint flag, struct proto *p, rte *e, char *msg)
242 {
243 if (p->debug & flag)
244 rte_trace(p, e, '<', msg);
245 }
246
247 static rte *
248 export_filter(struct announce_hook *ah, rte *rt0, rte **rt_free, ea_list **tmpa, int silent)
249 {
250 struct proto *p = ah->proto;
251 struct filter *filter = ah->out_filter;
252 struct proto_stats *stats = ah->stats;
253 ea_list *tmpb = NULL;
254 rte *rt;
255 int v;
256
257 rt = rt0;
258 *rt_free = NULL;
259
260 if (!tmpa)
261 tmpa = &tmpb;
262
263 *tmpa = make_tmp_attrs(rt, rte_update_pool);
264
265 v = p->import_control ? p->import_control(p, &rt, tmpa, rte_update_pool) : 0;
266 if (v < 0)
267 {
268 if (silent)
269 goto reject;
270
271 stats->exp_updates_rejected++;
272 if (v == RIC_REJECT)
273 rte_trace_out(D_FILTERS, p, rt, "rejected by protocol");
274 goto reject;
275 }
276 if (v > 0)
277 {
278 if (!silent)
279 rte_trace_out(D_FILTERS, p, rt, "forced accept by protocol");
280 goto accept;
281 }
282
283 v = filter && ((filter == FILTER_REJECT) ||
284 (f_run(filter, &rt, tmpa, rte_update_pool, FF_FORCE_TMPATTR) > F_ACCEPT));
285 if (v)
286 {
287 if (silent)
288 goto reject;
289
290 stats->exp_updates_filtered++;
291 rte_trace_out(D_FILTERS, p, rt, "filtered out");
292 goto reject;
293 }
294
295 accept:
296 if (rt != rt0)
297 *rt_free = rt;
298 return rt;
299
300 reject:
301 /* Discard temporary rte */
302 if (rt != rt0)
303 rte_free(rt);
304 return NULL;
305 }
306
307 static void
308 do_rt_notify(struct announce_hook *ah, net *net, rte *new, rte *old, ea_list *tmpa, int refeed)
309 {
310 struct proto *p = ah->proto;
311 struct proto_stats *stats = ah->stats;
312
313
314 /*
315 * First, apply export limit.
316 *
317 * Export route limits has several problems. Because exp_routes
318 * counter is reset before refeed, we don't really know whether
319 * limit is breached and whether the update is new or not. Therefore
320 * the number of really exported routes may exceed the limit
321 * temporarily (routes exported before and new routes in refeed).
322 *
323 * Minor advantage is that if the limit is decreased and refeed is
324 * requested, the number of exported routes really decrease.
325 *
326 * Second problem is that with export limits, we don't know whether
327 * old was really exported (it might be blocked by limit). When a
328 * withdraw is exported, we announce it even when the previous
329 * update was blocked. This is not a big issue, but the same problem
330 * is in updating exp_routes counter. Therefore, to be consistent in
331 * increases and decreases of exp_routes, we count exported routes
332 * regardless of blocking by limits.
333 *
334 * Similar problem is in handling updates - when a new route is
335 * received and blocking is active, the route would be blocked, but
336 * when an update for the route will be received later, the update
337 * would be propagated (as old != NULL). Therefore, we have to block
338 * also non-new updates (contrary to import blocking).
339 */
340
341 struct proto_limit *l = ah->out_limit;
342 if (l && new)
343 {
344 if ((!old || refeed) && (stats->exp_routes >= l->limit))
345 proto_notify_limit(ah, l, PLD_OUT, stats->exp_routes);
346
347 if (l->state == PLS_BLOCKED)
348 {
349 stats->exp_routes++; /* see note above */
350 stats->exp_updates_rejected++;
351 rte_trace_out(D_FILTERS, p, new, "rejected [limit]");
352 new = NULL;
353
354 if (!old)
355 return;
356 }
357 }
358
359
360 if (new)
361 stats->exp_updates_accepted++;
362 else
363 stats->exp_withdraws_accepted++;
364
365 /* Hack: We do not decrease exp_routes during refeed, we instead
366 reset exp_routes at the start of refeed. */
367 if (new)
368 stats->exp_routes++;
369 if (old && !refeed)
370 stats->exp_routes--;
371
372 if (p->debug & D_ROUTES)
373 {
374 if (new && old)
375 rte_trace_out(D_ROUTES, p, new, "replaced");
376 else if (new)
377 rte_trace_out(D_ROUTES, p, new, "added");
378 else if (old)
379 rte_trace_out(D_ROUTES, p, old, "removed");
380 }
381 if (!new)
382 p->rt_notify(p, ah->table, net, NULL, old, NULL);
383 else if (tmpa)
384 {
385 ea_list *t = tmpa;
386 while (t->next)
387 t = t->next;
388 t->next = new->attrs->eattrs;
389 p->rt_notify(p, ah->table, net, new, old, tmpa);
390 t->next = NULL;
391 }
392 else
393 p->rt_notify(p, ah->table, net, new, old, new->attrs->eattrs);
394 }
395
396 static void
397 rt_notify_basic(struct announce_hook *ah, net *net, rte *new0, rte *old0, int refeed)
398 {
399 struct proto *p = ah->proto;
400 struct proto_stats *stats = ah->stats;
401
402 rte *new = new0;
403 rte *old = old0;
404 rte *new_free = NULL;
405 rte *old_free = NULL;
406 ea_list *tmpa = NULL;
407
408 if (new)
409 stats->exp_updates_received++;
410 else
411 stats->exp_withdraws_received++;
412
413 /*
414 * This is a tricky part - we don't know whether route 'old' was
415 * exported to protocol 'p' or was filtered by the export filter.
416 * We try to run the export filter to know this to have a correct
417 * value in 'old' argument of rte_update (and proper filter value)
418 *
419 * FIXME - this is broken because 'configure soft' may change
420 * filters but keep routes. Refeed is expected to be called after
421 * change of the filters and with old == new, therefore we do not
422 * even try to run the filter on an old route, This may lead to
423 * 'spurious withdraws' but ensure that there are no 'missing
424 * withdraws'.
425 *
426 * This is not completely safe as there is a window between
427 * reconfiguration and the end of refeed - if a newly filtered
428 * route disappears during this period, proper withdraw is not
429 * sent (because old would be also filtered) and the route is
430 * not refeeded (because it disappeared before that).
431 */
432
433 if (new)
434 new = export_filter(ah, new, &new_free, &tmpa, 0);
435
436 if (old && !refeed)
437 old = export_filter(ah, old, &old_free, NULL, 1);
438
439 if (!new && !old)
440 {
441 /*
442 * As mentioned above, 'old' value may be incorrect in some race conditions.
443 * We generally ignore it with the exception of withdraw to pipe protocol.
444 * In that case we rather propagate unfiltered withdraws regardless of
445 * export filters to ensure that when a protocol is flushed, its routes are
446 * removed from all tables. Possible spurious unfiltered withdraws are not
447 * problem here as they are ignored if there is no corresponding route at
448 * the other end of the pipe. We directly call rt_notify() hook instead of
449 * do_rt_notify() to avoid logging and stat counters.
450 */
451
452 #ifdef CONFIG_PIPE
453 if ((p->proto == &proto_pipe) && !new0 && (p != old0->sender->proto))
454 p->rt_notify(p, ah->table, net, NULL, old0, NULL);
455 #endif
456
457 return;
458 }
459
460 do_rt_notify(ah, net, new, old, tmpa, refeed);
461
462 /* Discard temporary rte's */
463 if (new_free)
464 rte_free(new_free);
465 if (old_free)
466 rte_free(old_free);
467 }
468
469 static void
470 rt_notify_accepted(struct announce_hook *ah, net *net, rte *new_changed, rte *old_changed, rte *before_old, int feed)
471 {
472 // struct proto *p = ah->proto;
473 struct proto_stats *stats = ah->stats;
474
475 rte *r;
476 rte *new_best = NULL;
477 rte *old_best = NULL;
478 rte *new_free = NULL;
479 rte *old_free = NULL;
480 ea_list *tmpa = NULL;
481
482 /* Used to track whether we met old_changed position. If before_old is NULL
483 old_changed was the first and we met it implicitly before current best route. */
484 int old_meet = old_changed && !before_old;
485
486 /* Note that before_old is either NULL or valid (not rejected) route.
487 If old_changed is valid, before_old have to be too. If old changed route
488 was not valid, caller must use NULL for both old_changed and before_old. */
489
490 if (new_changed)
491 stats->exp_updates_received++;
492 else
493 stats->exp_withdraws_received++;
494
495 /* First, find the new_best route - first accepted by filters */
496 for (r=net->routes; rte_is_valid(r); r=r->next)
497 {
498 if (new_best = export_filter(ah, r, &new_free, &tmpa, 0))
499 break;
500
501 /* Note if we walked around the position of old_changed route */
502 if (r == before_old)
503 old_meet = 1;
504 }
505
506 /*
507 * Second, handle the feed case. That means we do not care for
508 * old_best. It is NULL for feed, and the new_best for refeed.
509 * For refeed, there is a hack similar to one in rt_notify_basic()
510 * to ensure withdraws in case of changed filters
511 */
512 if (feed)
513 {
514 if (feed == 2) /* refeed */
515 old_best = new_best ? new_best :
516 (rte_is_valid(net->routes) ? net->routes : NULL);
517 else
518 old_best = NULL;
519
520 if (!new_best && !old_best)
521 return;
522
523 goto found;
524 }
525
526 /*
527 * Now, we find the old_best route. Generally, it is the same as the
528 * new_best, unless new_best is the same as new_changed or
529 * old_changed is accepted before new_best.
530 *
531 * There are four cases:
532 *
533 * - We would find and accept old_changed before new_best, therefore
534 * old_changed is old_best. In remaining cases we suppose this
535 * is not true.
536 *
537 * - We found no new_best, therefore there is also no old_best and
538 * we ignore this withdraw.
539 *
540 * - We found new_best different than new_changed, therefore
541 * old_best is the same as new_best and we ignore this update.
542 *
543 * - We found new_best the same as new_changed, therefore it cannot
544 * be old_best and we have to continue search for old_best.
545 */
546
547 /* First case */
548 if (old_meet)
549 if (old_best = export_filter(ah, old_changed, &old_free, NULL, 1))
550 goto found;
551
552 /* Second case */
553 if (!new_best)
554 return;
555
556 /* Third case, we use r instead of new_best, because export_filter() could change it */
557 if (r != new_changed)
558 {
559 if (new_free)
560 rte_free(new_free);
561 return;
562 }
563
564 /* Fourth case */
565 for (r=r->next; rte_is_valid(r); r=r->next)
566 {
567 if (old_best = export_filter(ah, r, &old_free, NULL, 1))
568 goto found;
569
570 if (r == before_old)
571 if (old_best = export_filter(ah, old_changed, &old_free, NULL, 1))
572 goto found;
573 }
574
575 /* Implicitly, old_best is NULL and new_best is non-NULL */
576
577 found:
578 do_rt_notify(ah, net, new_best, old_best, tmpa, (feed == 2));
579
580 /* Discard temporary rte's */
581 if (new_free)
582 rte_free(new_free);
583 if (old_free)
584 rte_free(old_free);
585 }
586
587
588 static struct mpnh *
589 mpnh_merge_rta(struct mpnh *nhs, rta *a, int max)
590 {
591 struct mpnh nh = { .gw = a->gw, .iface = a->iface };
592 struct mpnh *nh2 = (a->dest == RTD_MULTIPATH) ? a->nexthops : &nh;
593 return mpnh_merge(nhs, nh2, 1, 0, max, rte_update_pool);
594 }
595
596 rte *
597 rt_export_merged(struct announce_hook *ah, net *net, rte **rt_free, ea_list **tmpa, int silent)
598 {
599 // struct proto *p = ah->proto;
600 struct mpnh *nhs = NULL;
601 rte *best0, *best, *rt0, *rt, *tmp;
602
603 best0 = net->routes;
604 *rt_free = NULL;
605
606 if (!rte_is_valid(best0))
607 return NULL;
608
609 best = export_filter(ah, best0, rt_free, tmpa, silent);
610
611 if (!best || !rte_is_reachable(best))
612 return best;
613
614 for (rt0 = best0->next; rt0; rt0 = rt0->next)
615 {
616 if (!rte_mergable(best0, rt0))
617 continue;
618
619 rt = export_filter(ah, rt0, &tmp, NULL, 1);
620
621 if (!rt)
622 continue;
623
624 if (rte_is_reachable(rt))
625 nhs = mpnh_merge_rta(nhs, rt->attrs, ah->proto->merge_limit);
626
627 if (tmp)
628 rte_free(tmp);
629 }
630
631 if (nhs)
632 {
633 nhs = mpnh_merge_rta(nhs, best->attrs, ah->proto->merge_limit);
634
635 if (nhs->next)
636 {
637 best = rte_cow_rta(best, rte_update_pool);
638 best->attrs->dest = RTD_MULTIPATH;
639 best->attrs->nexthops = nhs;
640 }
641 }
642
643 if (best != best0)
644 *rt_free = best;
645
646 return best;
647 }
648
649
650 static void
651 rt_notify_merged(struct announce_hook *ah, net *net, rte *new_changed, rte *old_changed,
652 rte *new_best, rte*old_best, int refeed)
653 {
654 // struct proto *p = ah->proto;
655
656 rte *new_best_free = NULL;
657 rte *old_best_free = NULL;
658 rte *new_changed_free = NULL;
659 rte *old_changed_free = NULL;
660 ea_list *tmpa = NULL;
661
662 /* We assume that all rte arguments are either NULL or rte_is_valid() */
663
664 /* This check should be done by the caller */
665 if (!new_best && !old_best)
666 return;
667
668 /* Check whether the change is relevant to the merged route */
669 if ((new_best == old_best) && !refeed)
670 {
671 new_changed = rte_mergable(new_best, new_changed) ?
672 export_filter(ah, new_changed, &new_changed_free, NULL, 1) : NULL;
673
674 old_changed = rte_mergable(old_best, old_changed) ?
675 export_filter(ah, old_changed, &old_changed_free, NULL, 1) : NULL;
676
677 if (!new_changed && !old_changed)
678 return;
679 }
680
681 if (new_best)
682 ah->stats->exp_updates_received++;
683 else
684 ah->stats->exp_withdraws_received++;
685
686 /* Prepare new merged route */
687 if (new_best)
688 new_best = rt_export_merged(ah, net, &new_best_free, &tmpa, 0);
689
690 /* Prepare old merged route (without proper merged next hops) */
691 /* There are some issues with running filter on old route - see rt_notify_basic() */
692 if (old_best && !refeed)
693 old_best = export_filter(ah, old_best, &old_best_free, NULL, 1);
694
695 if (new_best || old_best)
696 do_rt_notify(ah, net, new_best, old_best, tmpa, refeed);
697
698 /* Discard temporary rte's */
699 if (new_best_free)
700 rte_free(new_best_free);
701 if (old_best_free)
702 rte_free(old_best_free);
703 if (new_changed_free)
704 rte_free(new_changed_free);
705 if (old_changed_free)
706 rte_free(old_changed_free);
707 }
708
709
710 /**
711 * rte_announce - announce a routing table change
712 * @tab: table the route has been added to
713 * @type: type of route announcement (RA_OPTIMAL or RA_ANY)
714 * @net: network in question
715 * @new: the new route to be announced
716 * @old: the previous route for the same network
717 * @new_best: the new best route for the same network
718 * @old_best: the previous best route for the same network
719 * @before_old: The previous route before @old for the same network.
720 * If @before_old is NULL @old was the first.
721 *
722 * This function gets a routing table update and announces it
723 * to all protocols that acccepts given type of route announcement
724 * and are connected to the same table by their announcement hooks.
725 *
726 * Route announcement of type %RA_OPTIMAL si generated when optimal
727 * route (in routing table @tab) changes. In that case @old stores the
728 * old optimal route.
729 *
730 * Route announcement of type %RA_ANY si generated when any route (in
731 * routing table @tab) changes In that case @old stores the old route
732 * from the same protocol.
733 *
734 * For each appropriate protocol, we first call its import_control()
735 * hook which performs basic checks on the route (each protocol has a
736 * right to veto or force accept of the route before any filter is
737 * asked) and adds default values of attributes specific to the new
738 * protocol (metrics, tags etc.). Then it consults the protocol's
739 * export filter and if it accepts the route, the rt_notify() hook of
740 * the protocol gets called.
741 */
742 static void
743 rte_announce(rtable *tab, unsigned type, net *net, rte *new, rte *old,
744 rte *new_best, rte *old_best, rte *before_old)
745 {
746 if (!rte_is_valid(new))
747 new = NULL;
748
749 if (!rte_is_valid(old))
750 old = before_old = NULL;
751
752 if (!rte_is_valid(new_best))
753 new_best = NULL;
754
755 if (!rte_is_valid(old_best))
756 old_best = NULL;
757
758 if (!old && !new)
759 return;
760
761 if (type == RA_OPTIMAL)
762 {
763 if (new)
764 new->attrs->src->proto->stats.pref_routes++;
765 if (old)
766 old->attrs->src->proto->stats.pref_routes--;
767
768 if (tab->hostcache)
769 rt_notify_hostcache(tab, net);
770 }
771
772 struct announce_hook *a;
773 WALK_LIST(a, tab->hooks)
774 {
775 ASSERT(a->proto->export_state != ES_DOWN);
776 if (a->proto->accept_ra_types == type)
777 if (type == RA_ACCEPTED)
778 rt_notify_accepted(a, net, new, old, before_old, 0);
779 else if (type == RA_MERGED)
780 rt_notify_merged(a, net, new, old, new_best, old_best, 0);
781 else
782 rt_notify_basic(a, net, new, old, 0);
783 }
784 }
785
786 static inline int
787 rte_validate(rte *e)
788 {
789 int c;
790 net *n = e->net;
791
792 if ((n->n.pxlen > BITS_PER_IP_ADDRESS) || !ip_is_prefix(n->n.prefix,n->n.pxlen))
793 {
794 log(L_WARN "Ignoring bogus prefix %I/%d received via %s",
795 n->n.prefix, n->n.pxlen, e->sender->proto->name);
796 return 0;
797 }
798
799 c = ipa_classify_net(n->n.prefix);
800 if ((c < 0) || !(c & IADDR_HOST) || ((c & IADDR_SCOPE_MASK) <= SCOPE_LINK))
801 {
802 log(L_WARN "Ignoring bogus route %I/%d received via %s",
803 n->n.prefix, n->n.pxlen, e->sender->proto->name);
804 return 0;
805 }
806
807 return 1;
808 }
809
810 /**
811 * rte_free - delete a &rte
812 * @e: &rte to be deleted
813 *
814 * rte_free() deletes the given &rte from the routing table it's linked to.
815 */
816 void
817 rte_free(rte *e)
818 {
819 if (rta_is_cached(e->attrs))
820 rta_free(e->attrs);
821 sl_free(rte_slab, e);
822 }
823
824 static inline void
825 rte_free_quick(rte *e)
826 {
827 rta_free(e->attrs);
828 sl_free(rte_slab, e);
829 }
830
831 static int
832 rte_same(rte *x, rte *y)
833 {
834 return
835 x->attrs == y->attrs &&
836 x->flags == y->flags &&
837 x->pflags == y->pflags &&
838 x->pref == y->pref &&
839 (!x->attrs->src->proto->rte_same || x->attrs->src->proto->rte_same(x, y));
840 }
841
842 static inline int rte_is_ok(rte *e) { return e && !rte_is_filtered(e); }
843
844 static void
845 rte_recalculate(struct announce_hook *ah, net *net, rte *new, struct rte_src *src)
846 {
847 struct proto *p = ah->proto;
848 struct rtable *table = ah->table;
849 struct proto_stats *stats = ah->stats;
850 static struct tbf rl_pipe = TBF_DEFAULT_LOG_LIMITS;
851 rte *before_old = NULL;
852 rte *old_best = net->routes;
853 rte *old = NULL;
854 rte **k;
855
856 k = &net->routes; /* Find and remove original route from the same protocol */
857 while (old = *k)
858 {
859 if (old->attrs->src == src)
860 {
861 /* If there is the same route in the routing table but from
862 * a different sender, then there are two paths from the
863 * source protocol to this routing table through transparent
864 * pipes, which is not allowed.
865 *
866 * We log that and ignore the route. If it is withdraw, we
867 * ignore it completely (there might be 'spurious withdraws',
868 * see FIXME in do_rte_announce())
869 */
870 if (old->sender->proto != p)
871 {
872 if (new)
873 {
874 log_rl(&rl_pipe, L_ERR "Pipe collision detected when sending %I/%d to table %s",
875 net->n.prefix, net->n.pxlen, table->name);
876 rte_free_quick(new);
877 }
878 return;
879 }
880
881 if (new && rte_same(old, new))
882 {
883 /* No changes, ignore the new route */
884
885 if (!rte_is_filtered(new))
886 {
887 stats->imp_updates_ignored++;
888 rte_trace_in(D_ROUTES, p, new, "ignored");
889 }
890
891 rte_free_quick(new);
892 return;
893 }
894 *k = old->next;
895 break;
896 }
897 k = &old->next;
898 before_old = old;
899 }
900
901 if (!old)
902 before_old = NULL;
903
904 if (!old && !new)
905 {
906 stats->imp_withdraws_ignored++;
907 return;
908 }
909
910 int new_ok = rte_is_ok(new);
911 int old_ok = rte_is_ok(old);
912
913 struct proto_limit *l = ah->rx_limit;
914 if (l && !old && new)
915 {
916 u32 all_routes = stats->imp_routes + stats->filt_routes;
917
918 if (all_routes >= l->limit)
919 proto_notify_limit(ah, l, PLD_RX, all_routes);
920
921 if (l->state == PLS_BLOCKED)
922 {
923 /* In receive limit the situation is simple, old is NULL so
924 we just free new and exit like nothing happened */
925
926 stats->imp_updates_ignored++;
927 rte_trace_in(D_FILTERS, p, new, "ignored [limit]");
928 rte_free_quick(new);
929 return;
930 }
931 }
932
933 l = ah->in_limit;
934 if (l && !old_ok && new_ok)
935 {
936 if (stats->imp_routes >= l->limit)
937 proto_notify_limit(ah, l, PLD_IN, stats->imp_routes);
938
939 if (l->state == PLS_BLOCKED)
940 {
941 /* In import limit the situation is more complicated. We
942 shouldn't just drop the route, we should handle it like
943 it was filtered. We also have to continue the route
944 processing if old or new is non-NULL, but we should exit
945 if both are NULL as this case is probably assumed to be
946 already handled. */
947
948 stats->imp_updates_ignored++;
949 rte_trace_in(D_FILTERS, p, new, "ignored [limit]");
950
951 if (ah->in_keep_filtered)
952 new->flags |= REF_FILTERED;
953 else
954 { rte_free_quick(new); new = NULL; }
955
956 /* Note that old && !new could be possible when
957 ah->in_keep_filtered changed in the recent past. */
958
959 if (!old && !new)
960 return;
961
962 new_ok = 0;
963 goto skip_stats1;
964 }
965 }
966
967 if (new_ok)
968 stats->imp_updates_accepted++;
969 else if (old_ok)
970 stats->imp_withdraws_accepted++;
971 else
972 stats->imp_withdraws_ignored++;
973
974 skip_stats1:
975
976 if (new)
977 rte_is_filtered(new) ? stats->filt_routes++ : stats->imp_routes++;
978 if (old)
979 rte_is_filtered(old) ? stats->filt_routes-- : stats->imp_routes--;
980
981 if (table->config->sorted)
982 {
983 /* If routes are sorted, just insert new route to appropriate position */
984 if (new)
985 {
986 if (before_old && !rte_better(new, before_old))
987 k = &before_old->next;
988 else
989 k = &net->routes;
990
991 for (; *k; k=&(*k)->next)
992 if (rte_better(new, *k))
993 break;
994
995 new->next = *k;
996 *k = new;
997 }
998 }
999 else
1000 {
1001 /* If routes are not sorted, find the best route and move it on
1002 the first position. There are several optimized cases. */
1003
1004 if (src->proto->rte_recalculate && src->proto->rte_recalculate(table, net, new, old, old_best))
1005 goto do_recalculate;
1006
1007 if (new && rte_better(new, old_best))
1008 {
1009 /* The first case - the new route is cleary optimal,
1010 we link it at the first position */
1011
1012 new->next = net->routes;
1013 net->routes = new;
1014 }
1015 else if (old == old_best)
1016 {
1017 /* The second case - the old best route disappeared, we add the
1018 new route (if we have any) to the list (we don't care about
1019 position) and then we elect the new optimal route and relink
1020 that route at the first position and announce it. New optimal
1021 route might be NULL if there is no more routes */
1022
1023 do_recalculate:
1024 /* Add the new route to the list */
1025 if (new)
1026 {
1027 new->next = net->routes;
1028 net->routes = new;
1029 }
1030
1031 /* Find a new optimal route (if there is any) */
1032 if (net->routes)
1033 {
1034 rte **bp = &net->routes;
1035 for (k=&(*bp)->next; *k; k=&(*k)->next)
1036 if (rte_better(*k, *bp))
1037 bp = k;
1038
1039 /* And relink it */
1040 rte *best = *bp;
1041 *bp = best->next;
1042 best->next = net->routes;
1043 net->routes = best;
1044 }
1045 }
1046 else if (new)
1047 {
1048 /* The third case - the new route is not better than the old
1049 best route (therefore old_best != NULL) and the old best
1050 route was not removed (therefore old_best == net->routes).
1051 We just link the new route after the old best route. */
1052
1053 ASSERT(net->routes != NULL);
1054 new->next = net->routes->next;
1055 net->routes->next = new;
1056 }
1057 /* The fourth (empty) case - suboptimal route was removed, nothing to do */
1058 }
1059
1060 if (new)
1061 new->lastmod = now;
1062
1063 /* Log the route change */
1064 if (p->debug & D_ROUTES)
1065 {
1066 if (new_ok)
1067 rte_trace(p, new, '>', new == net->routes ? "added [best]" : "added");
1068 else if (old_ok)
1069 {
1070 if (old != old_best)
1071 rte_trace(p, old, '>', "removed");
1072 else if (rte_is_ok(net->routes))
1073 rte_trace(p, old, '>', "removed [replaced]");
1074 else
1075 rte_trace(p, old, '>', "removed [sole]");
1076 }
1077 }
1078
1079 /* Propagate the route change */
1080 rte_announce(table, RA_ANY, net, new, old, NULL, NULL, NULL);
1081 if (net->routes != old_best)
1082 rte_announce(table, RA_OPTIMAL, net, net->routes, old_best, NULL, NULL, NULL);
1083 if (table->config->sorted)
1084 rte_announce(table, RA_ACCEPTED, net, new, old, NULL, NULL, before_old);
1085 rte_announce(table, RA_MERGED, net, new, old, net->routes, old_best, NULL);
1086
1087 if (!net->routes &&
1088 (table->gc_counter++ >= table->config->gc_max_ops) &&
1089 (table->gc_time + table->config->gc_min_time <= now))
1090 rt_schedule_gc(table);
1091
1092 if (old_ok && p->rte_remove)
1093 p->rte_remove(net, old);
1094 if (new_ok && p->rte_insert)
1095 p->rte_insert(net, new);
1096
1097 if (old)
1098 rte_free_quick(old);
1099 }
1100
1101 static int rte_update_nest_cnt; /* Nesting counter to allow recursive updates */
1102
1103 static inline void
1104 rte_update_lock(void)
1105 {
1106 rte_update_nest_cnt++;
1107 }
1108
1109 static inline void
1110 rte_update_unlock(void)
1111 {
1112 if (!--rte_update_nest_cnt)
1113 lp_flush(rte_update_pool);
1114 }
1115
1116 static inline void
1117 rte_hide_dummy_routes(net *net, rte **dummy)
1118 {
1119 if (net->routes && net->routes->attrs->source == RTS_DUMMY)
1120 {
1121 *dummy = net->routes;
1122 net->routes = (*dummy)->next;
1123 }
1124 }
1125
1126 static inline void
1127 rte_unhide_dummy_routes(net *net, rte **dummy)
1128 {
1129 if (*dummy)
1130 {
1131 (*dummy)->next = net->routes;
1132 net->routes = *dummy;
1133 }
1134 }
1135
1136 /**
1137 * rte_update - enter a new update to a routing table
1138 * @table: table to be updated
1139 * @ah: pointer to table announce hook
1140 * @net: network node
1141 * @p: protocol submitting the update
1142 * @src: protocol originating the update
1143 * @new: a &rte representing the new route or %NULL for route removal.
1144 *
1145 * This function is called by the routing protocols whenever they discover
1146 * a new route or wish to update/remove an existing route. The right announcement
1147 * sequence is to build route attributes first (either un-cached with @aflags set
1148 * to zero or a cached one using rta_lookup(); in this case please note that
1149 * you need to increase the use count of the attributes yourself by calling
1150 * rta_clone()), call rte_get_temp() to obtain a temporary &rte, fill in all
1151 * the appropriate data and finally submit the new &rte by calling rte_update().
1152 *
1153 * @src specifies the protocol that originally created the route and the meaning
1154 * of protocol-dependent data of @new. If @new is not %NULL, @src have to be the
1155 * same value as @new->attrs->proto. @p specifies the protocol that called
1156 * rte_update(). In most cases it is the same protocol as @src. rte_update()
1157 * stores @p in @new->sender;
1158 *
1159 * When rte_update() gets any route, it automatically validates it (checks,
1160 * whether the network and next hop address are valid IP addresses and also
1161 * whether a normal routing protocol doesn't try to smuggle a host or link
1162 * scope route to the table), converts all protocol dependent attributes stored
1163 * in the &rte to temporary extended attributes, consults import filters of the
1164 * protocol to see if the route should be accepted and/or its attributes modified,
1165 * stores the temporary attributes back to the &rte.
1166 *
1167 * Now, having a "public" version of the route, we
1168 * automatically find any old route defined by the protocol @src
1169 * for network @n, replace it by the new one (or removing it if @new is %NULL),
1170 * recalculate the optimal route for this destination and finally broadcast
1171 * the change (if any) to all routing protocols by calling rte_announce().
1172 *
1173 * All memory used for attribute lists and other temporary allocations is taken
1174 * from a special linear pool @rte_update_pool and freed when rte_update()
1175 * finishes.
1176 */
1177
1178 void
1179 rte_update2(struct announce_hook *ah, net *net, rte *new, struct rte_src *src)
1180 {
1181 struct proto *p = ah->proto;
1182 struct proto_stats *stats = ah->stats;
1183 struct filter *filter = ah->in_filter;
1184 ea_list *tmpa = NULL;
1185 rte *dummy = NULL;
1186
1187 rte_update_lock();
1188 if (new)
1189 {
1190 new->sender = ah;
1191
1192 stats->imp_updates_received++;
1193 if (!rte_validate(new))
1194 {
1195 rte_trace_in(D_FILTERS, p, new, "invalid");
1196 stats->imp_updates_invalid++;
1197 goto drop;
1198 }
1199
1200 if (filter == FILTER_REJECT)
1201 {
1202 stats->imp_updates_filtered++;
1203 rte_trace_in(D_FILTERS, p, new, "filtered out");
1204
1205 if (! ah->in_keep_filtered)
1206 goto drop;
1207
1208 /* new is a private copy, i could modify it */
1209 new->flags |= REF_FILTERED;
1210 }
1211 else
1212 {
1213 tmpa = make_tmp_attrs(new, rte_update_pool);
1214 if (filter && (filter != FILTER_REJECT))
1215 {
1216 ea_list *old_tmpa = tmpa;
1217 int fr = f_run(filter, &new, &tmpa, rte_update_pool, 0);
1218 if (fr > F_ACCEPT)
1219 {
1220 stats->imp_updates_filtered++;
1221 rte_trace_in(D_FILTERS, p, new, "filtered out");
1222
1223 if (! ah->in_keep_filtered)
1224 goto drop;
1225
1226 new->flags |= REF_FILTERED;
1227 }
1228 if (tmpa != old_tmpa && src->proto->store_tmp_attrs)
1229 src->proto->store_tmp_attrs(new, tmpa);
1230 }
1231 }
1232 if (!rta_is_cached(new->attrs)) /* Need to copy attributes */
1233 new->attrs = rta_lookup(new->attrs);
1234 new->flags |= REF_COW;
1235 }
1236 else
1237 {
1238 stats->imp_withdraws_received++;
1239
1240 if (!net || !src)
1241 {
1242 stats->imp_withdraws_ignored++;
1243 rte_update_unlock();
1244 return;
1245 }
1246 }
1247
1248 recalc:
1249 rte_hide_dummy_routes(net, &dummy);
1250 rte_recalculate(ah, net, new, src);
1251 rte_unhide_dummy_routes(net, &dummy);
1252 rte_update_unlock();
1253 return;
1254
1255 drop:
1256 rte_free(new);
1257 new = NULL;
1258 goto recalc;
1259 }
1260
1261 /* Independent call to rte_announce(), used from next hop
1262 recalculation, outside of rte_update(). new must be non-NULL */
1263 static inline void
1264 rte_announce_i(rtable *tab, unsigned type, net *net, rte *new, rte *old,
1265 rte *new_best, rte *old_best)
1266 {
1267 rte_update_lock();
1268 rte_announce(tab, type, net, new, old, new_best, old_best, NULL);
1269 rte_update_unlock();
1270 }
1271
1272 void
1273 rte_discard(rtable *t, rte *old) /* Non-filtered route deletion, used during garbage collection */
1274 {
1275 rte_update_lock();
1276 rte_recalculate(old->sender, old->net, NULL, old->attrs->src);
1277 rte_update_unlock();
1278 }
1279
1280 /* Check rtable for best route to given net whether it would be exported do p */
1281 int
1282 rt_examine(rtable *t, ip_addr prefix, int pxlen, struct proto *p, struct filter *filter)
1283 {
1284 net *n = net_find(t, prefix, pxlen);
1285 rte *rt = n ? n->routes : NULL;
1286
1287 if (!rte_is_valid(rt))
1288 return 0;
1289
1290 rte_update_lock();
1291
1292 /* Rest is stripped down export_filter() */
1293 ea_list *tmpa = make_tmp_attrs(rt, rte_update_pool);
1294 int v = p->import_control ? p->import_control(p, &rt, &tmpa, rte_update_pool) : 0;
1295 if (v == RIC_PROCESS)
1296 v = (f_run(filter, &rt, &tmpa, rte_update_pool, FF_FORCE_TMPATTR) <= F_ACCEPT);
1297
1298 /* Discard temporary rte */
1299 if (rt != n->routes)
1300 rte_free(rt);
1301
1302 rte_update_unlock();
1303
1304 return v > 0;
1305 }
1306
1307
1308 /**
1309 * rt_refresh_begin - start a refresh cycle
1310 * @t: related routing table
1311 * @ah: related announce hook
1312 *
1313 * This function starts a refresh cycle for given routing table and announce
1314 * hook. The refresh cycle is a sequence where the protocol sends all its valid
1315 * routes to the routing table (by rte_update()). After that, all protocol
1316 * routes (more precisely routes with @ah as @sender) not sent during the
1317 * refresh cycle but still in the table from the past are pruned. This is
1318 * implemented by marking all related routes as stale by REF_STALE flag in
1319 * rt_refresh_begin(), then marking all related stale routes with REF_DISCARD
1320 * flag in rt_refresh_end() and then removing such routes in the prune loop.
1321 */
1322 void
1323 rt_refresh_begin(rtable *t, struct announce_hook *ah)
1324 {
1325 net *n;
1326 rte *e;
1327
1328 FIB_WALK(&t->fib, fn)
1329 {
1330 n = (net *) fn;
1331 for (e = n->routes; e; e = e->next)
1332 if (e->sender == ah)
1333 e->flags |= REF_STALE;
1334 }
1335 FIB_WALK_END;
1336 }
1337
1338 /**
1339 * rt_refresh_end - end a refresh cycle
1340 * @t: related routing table
1341 * @ah: related announce hook
1342 *
1343 * This function starts a refresh cycle for given routing table and announce
1344 * hook. See rt_refresh_begin() for description of refresh cycles.
1345 */
1346 void
1347 rt_refresh_end(rtable *t, struct announce_hook *ah)
1348 {
1349 int prune = 0;
1350 net *n;
1351 rte *e;
1352
1353 FIB_WALK(&t->fib, fn)
1354 {
1355 n = (net *) fn;
1356 for (e = n->routes; e; e = e->next)
1357 if ((e->sender == ah) && (e->flags & REF_STALE))
1358 {
1359 e->flags |= REF_DISCARD;
1360 prune = 1;
1361 }
1362 }
1363 FIB_WALK_END;
1364
1365 if (prune)
1366 rt_schedule_prune(t);
1367 }
1368
1369
1370 /**
1371 * rte_dump - dump a route
1372 * @e: &rte to be dumped
1373 *
1374 * This functions dumps contents of a &rte to debug output.
1375 */
1376 void
1377 rte_dump(rte *e)
1378 {
1379 net *n = e->net;
1380 debug("%-1I/%2d ", n->n.prefix, n->n.pxlen);
1381 debug("KF=%02x PF=%02x pref=%d lm=%d ", n->n.flags, e->pflags, e->pref, now-e->lastmod);
1382 rta_dump(e->attrs);
1383 if (e->attrs->src->proto->proto->dump_attrs)
1384 e->attrs->src->proto->proto->dump_attrs(e);
1385 debug("\n");
1386 }
1387
1388 /**
1389 * rt_dump - dump a routing table
1390 * @t: routing table to be dumped
1391 *
1392 * This function dumps contents of a given routing table to debug output.
1393 */
1394 void
1395 rt_dump(rtable *t)
1396 {
1397 rte *e;
1398 net *n;
1399 struct announce_hook *a;
1400
1401 debug("Dump of routing table <%s>\n", t->name);
1402 #ifdef DEBUGGING
1403 fib_check(&t->fib);
1404 #endif
1405 FIB_WALK(&t->fib, fn)
1406 {
1407 n = (net *) fn;
1408 for(e=n->routes; e; e=e->next)
1409 rte_dump(e);
1410 }
1411 FIB_WALK_END;
1412 WALK_LIST(a, t->hooks)
1413 debug("\tAnnounces routes to protocol %s\n", a->proto->name);
1414 debug("\n");
1415 }
1416
1417 /**
1418 * rt_dump_all - dump all routing tables
1419 *
1420 * This function dumps contents of all routing tables to debug output.
1421 */
1422 void
1423 rt_dump_all(void)
1424 {
1425 rtable *t;
1426
1427 WALK_LIST(t, routing_tables)
1428 rt_dump(t);
1429 }
1430
1431 static inline void
1432 rt_schedule_prune(rtable *tab)
1433 {
1434 rt_mark_for_prune(tab);
1435 ev_schedule(tab->rt_event);
1436 }
1437
1438 static inline void
1439 rt_schedule_gc(rtable *tab)
1440 {
1441 if (tab->gc_scheduled)
1442 return;
1443
1444 tab->gc_scheduled = 1;
1445 ev_schedule(tab->rt_event);
1446 }
1447
1448 static inline void
1449 rt_schedule_hcu(rtable *tab)
1450 {
1451 if (tab->hcu_scheduled)
1452 return;
1453
1454 tab->hcu_scheduled = 1;
1455 ev_schedule(tab->rt_event);
1456 }
1457
1458 static inline void
1459 rt_schedule_nhu(rtable *tab)
1460 {
1461 if (tab->nhu_state == 0)
1462 ev_schedule(tab->rt_event);
1463
1464 /* state change 0->1, 2->3 */
1465 tab->nhu_state |= 1;
1466 }
1467
1468
1469 static void
1470 rt_prune_nets(rtable *tab)
1471 {
1472 struct fib_iterator fit;
1473 int ncnt = 0, ndel = 0;
1474
1475 #ifdef DEBUGGING
1476 fib_check(&tab->fib);
1477 #endif
1478
1479 FIB_ITERATE_INIT(&fit, &tab->fib);
1480 again:
1481 FIB_ITERATE_START(&tab->fib, &fit, f)
1482 {
1483 net *n = (net *) f;
1484 ncnt++;
1485 if (!n->routes) /* Orphaned FIB entry */
1486 {
1487 FIB_ITERATE_PUT(&fit, f);
1488 fib_delete(&tab->fib, f);
1489 ndel++;
1490 goto again;
1491 }
1492 }
1493 FIB_ITERATE_END(f);
1494 DBG("Pruned %d of %d networks\n", ndel, ncnt);
1495
1496 tab->gc_counter = 0;
1497 tab->gc_time = now;
1498 tab->gc_scheduled = 0;
1499 }
1500
1501 static void
1502 rt_event(void *ptr)
1503 {
1504 rtable *tab = ptr;
1505
1506 if (tab->hcu_scheduled)
1507 rt_update_hostcache(tab);
1508
1509 if (tab->nhu_state)
1510 rt_next_hop_update(tab);
1511
1512 if (tab->prune_state)
1513 if (!rt_prune_table(tab))
1514 {
1515 /* Table prune unfinished */
1516 ev_schedule(tab->rt_event);
1517 return;
1518 }
1519
1520 if (tab->gc_scheduled)
1521 {
1522 rt_prune_nets(tab);
1523 rt_prune_sources(); // FIXME this should be moved to independent event
1524 }
1525 }
1526
1527 void
1528 rt_setup(pool *p, rtable *t, char *name, struct rtable_config *cf)
1529 {
1530 bzero(t, sizeof(*t));
1531 fib_init(&t->fib, p, sizeof(net), 0, rte_init);
1532 t->name = name;
1533 t->config = cf;
1534 init_list(&t->hooks);
1535 if (cf)
1536 {
1537 t->rt_event = ev_new(p);
1538 t->rt_event->hook = rt_event;
1539 t->rt_event->data = t;
1540 t->gc_time = now;
1541 }
1542 }
1543
1544 /**
1545 * rt_init - initialize routing tables
1546 *
1547 * This function is called during BIRD startup. It initializes the
1548 * routing table module.
1549 */
1550 void
1551 rt_init(void)
1552 {
1553 rta_init();
1554 rt_table_pool = rp_new(&root_pool, "Routing tables");
1555 rte_update_pool = lp_new(rt_table_pool, 4080);
1556 rte_slab = sl_new(rt_table_pool, sizeof(rte));
1557 init_list(&routing_tables);
1558 }
1559
1560
1561 static int
1562 rt_prune_step(rtable *tab, int *limit)
1563 {
1564 struct fib_iterator *fit = &tab->prune_fit;
1565
1566 DBG("Pruning route table %s\n", tab->name);
1567 #ifdef DEBUGGING
1568 fib_check(&tab->fib);
1569 #endif
1570
1571 if (tab->prune_state == RPS_NONE)
1572 return 1;
1573
1574 if (tab->prune_state == RPS_SCHEDULED)
1575 {
1576 FIB_ITERATE_INIT(fit, &tab->fib);
1577 tab->prune_state = RPS_RUNNING;
1578 }
1579
1580 again:
1581 FIB_ITERATE_START(&tab->fib, fit, fn)
1582 {
1583 net *n = (net *) fn;
1584 rte *e;
1585
1586 rescan:
1587 for (e=n->routes; e; e=e->next)
1588 if (e->sender->proto->flushing || (e->flags & REF_DISCARD))
1589 {
1590 if (*limit <= 0)
1591 {
1592 FIB_ITERATE_PUT(fit, fn);
1593 return 0;
1594 }
1595
1596 rte_discard(tab, e);
1597 (*limit)--;
1598
1599 goto rescan;
1600 }
1601 if (!n->routes) /* Orphaned FIB entry */
1602 {
1603 FIB_ITERATE_PUT(fit, fn);
1604 fib_delete(&tab->fib, fn);
1605 goto again;
1606 }
1607 }
1608 FIB_ITERATE_END(fn);
1609
1610 #ifdef DEBUGGING
1611 fib_check(&tab->fib);
1612 #endif
1613
1614 tab->prune_state = RPS_NONE;
1615 return 1;
1616 }
1617
1618 /**
1619 * rt_prune_table - prune a routing table
1620 * @tab: a routing table for pruning
1621 *
1622 * This function scans the routing table @tab and removes routes belonging to
1623 * flushing protocols, discarded routes and also stale network entries, in a
1624 * similar fashion like rt_prune_loop(). Returns 1 when all such routes are
1625 * pruned. Contrary to rt_prune_loop(), this function is not a part of the
1626 * protocol flushing loop, but it is called from rt_event() for just one routing
1627 * table.
1628 *
1629 * Note that rt_prune_table() and rt_prune_loop() share (for each table) the
1630 * prune state (@prune_state) and also the pruning iterator (@prune_fit).
1631 */
1632 static inline int
1633 rt_prune_table(rtable *tab)
1634 {
1635 int limit = 512;
1636 return rt_prune_step(tab, &limit);
1637 }
1638
1639 /**
1640 * rt_prune_loop - prune routing tables
1641 *
1642 * The prune loop scans routing tables and removes routes belonging to flushing
1643 * protocols, discarded routes and also stale network entries. Returns 1 when
1644 * all such routes are pruned. It is a part of the protocol flushing loop.
1645 */
1646 int
1647 rt_prune_loop(void)
1648 {
1649 int limit = 512;
1650 rtable *t;
1651
1652 WALK_LIST(t, routing_tables)
1653 if (! rt_prune_step(t, &limit))
1654 return 0;
1655
1656 return 1;
1657 }
1658
1659 void
1660 rt_preconfig(struct config *c)
1661 {
1662 struct symbol *s = cf_get_symbol("master");
1663
1664 init_list(&c->tables);
1665 c->master_rtc = rt_new_table(s);
1666 }
1667
1668
1669 /*
1670 * Some functions for handing internal next hop updates
1671 * triggered by rt_schedule_nhu().
1672 */
1673
1674 static inline int
1675 rta_next_hop_outdated(rta *a)
1676 {
1677 struct hostentry *he = a->hostentry;
1678
1679 if (!he)
1680 return 0;
1681
1682 if (!he->src)
1683 return a->dest != RTD_UNREACHABLE;
1684
1685 return (a->iface != he->src->iface) || !ipa_equal(a->gw, he->gw) ||
1686 (a->dest != he->dest) || (a->igp_metric != he->igp_metric) ||
1687 !mpnh_same(a->nexthops, he->src->nexthops);
1688 }
1689
1690 static inline void
1691 rta_apply_hostentry(rta *a, struct hostentry *he)
1692 {
1693 a->hostentry = he;
1694 a->iface = he->src ? he->src->iface : NULL;
1695 a->gw = he->gw;
1696 a->dest = he->dest;
1697 a->igp_metric = he->igp_metric;
1698 a->nexthops = he->src ? he->src->nexthops : NULL;
1699 }
1700
1701 static inline rte *
1702 rt_next_hop_update_rte(rtable *tab, rte *old)
1703 {
1704 rta a;
1705 memcpy(&a, old->attrs, sizeof(rta));
1706 rta_apply_hostentry(&a, old->attrs->hostentry);
1707 a.aflags = 0;
1708
1709 rte *e = sl_alloc(rte_slab);
1710 memcpy(e, old, sizeof(rte));
1711 e->attrs = rta_lookup(&a);
1712
1713 return e;
1714 }
1715
1716 static inline int
1717 rt_next_hop_update_net(rtable *tab, net *n)
1718 {
1719 rte **k, *e, *new, *old_best, **new_best;
1720 int count = 0;
1721 int free_old_best = 0;
1722
1723 old_best = n->routes;
1724 if (!old_best)
1725 return 0;
1726
1727 for (k = &n->routes; e = *k; k = &e->next)
1728 if (rta_next_hop_outdated(e->attrs))
1729 {
1730 new = rt_next_hop_update_rte(tab, e);
1731 *k = new;
1732
1733 rte_announce_i(tab, RA_ANY, n, new, e, NULL, NULL);
1734 rte_trace_in(D_ROUTES, new->sender->proto, new, "updated");
1735
1736 /* Call a pre-comparison hook */
1737 /* Not really an efficient way to compute this */
1738 if (e->attrs->src->proto->rte_recalculate)
1739 e->attrs->src->proto->rte_recalculate(tab, n, new, e, NULL);
1740
1741 if (e != old_best)
1742 rte_free_quick(e);
1743 else /* Freeing of the old best rte is postponed */
1744 free_old_best = 1;
1745
1746 e = new;
1747 count++;
1748 }
1749
1750 if (!count)
1751 return 0;
1752
1753 /* Find the new best route */
1754 new_best = NULL;
1755 for (k = &n->routes; e = *k; k = &e->next)
1756 {
1757 if (!new_best || rte_better(e, *new_best))
1758 new_best = k;
1759 }
1760
1761 /* Relink the new best route to the first position */
1762 new = *new_best;
1763 if (new != n->routes)
1764 {
1765 *new_best = new->next;
1766 new->next = n->routes;
1767 n->routes = new;
1768 }
1769
1770 /* Announce the new best route */
1771 if (new != old_best)
1772 {
1773 rte_announce_i(tab, RA_OPTIMAL, n, new, old_best, NULL, NULL);
1774 rte_trace_in(D_ROUTES, new->sender->proto, new, "updated [best]");
1775 }
1776
1777 /* FIXME: Better announcement of merged routes */
1778 rte_announce_i(tab, RA_MERGED, n, new, old_best, new, old_best);
1779
1780 if (free_old_best)
1781 rte_free_quick(old_best);
1782
1783 return count;
1784 }
1785
1786 static void
1787 rt_next_hop_update(rtable *tab)
1788 {
1789 struct fib_iterator *fit = &tab->nhu_fit;
1790 int max_feed = 32;
1791
1792 if (tab->nhu_state == 0)
1793 return;
1794
1795 if (tab->nhu_state == 1)
1796 {
1797 FIB_ITERATE_INIT(fit, &tab->fib);
1798 tab->nhu_state = 2;
1799 }
1800
1801 FIB_ITERATE_START(&tab->fib, fit, fn)
1802 {
1803 if (max_feed <= 0)
1804 {
1805 FIB_ITERATE_PUT(fit, fn);
1806 ev_schedule(tab->rt_event);
1807 return;
1808 }
1809 max_feed -= rt_next_hop_update_net(tab, (net *) fn);
1810 }
1811 FIB_ITERATE_END(fn);
1812
1813 /* state change 2->0, 3->1 */
1814 tab->nhu_state &= 1;
1815
1816 if (tab->nhu_state > 0)
1817 ev_schedule(tab->rt_event);
1818 }
1819
1820
1821 struct rtable_config *
1822 rt_new_table(struct symbol *s)
1823 {
1824 /* Hack that allows to 'redefine' the master table */
1825 if ((s->class == SYM_TABLE) && (s->def == new_config->master_rtc))
1826 return s->def;
1827
1828 struct rtable_config *c = cfg_allocz(sizeof(struct rtable_config));
1829
1830 cf_define_symbol(s, SYM_TABLE, c);
1831 c->name = s->name;
1832 add_tail(&new_config->tables, &c->n);
1833 c->gc_max_ops = 1000;
1834 c->gc_min_time = 5;
1835 return c;
1836 }
1837
1838 /**
1839 * rt_lock_table - lock a routing table
1840 * @r: routing table to be locked
1841 *
1842 * Lock a routing table, because it's in use by a protocol,
1843 * preventing it from being freed when it gets undefined in a new
1844 * configuration.
1845 */
1846 void
1847 rt_lock_table(rtable *r)
1848 {
1849 r->use_count++;
1850 }
1851
1852 /**
1853 * rt_unlock_table - unlock a routing table
1854 * @r: routing table to be unlocked
1855 *
1856 * Unlock a routing table formerly locked by rt_lock_table(),
1857 * that is decrease its use count and delete it if it's scheduled
1858 * for deletion by configuration changes.
1859 */
1860 void
1861 rt_unlock_table(rtable *r)
1862 {
1863 if (!--r->use_count && r->deleted)
1864 {
1865 struct config *conf = r->deleted;
1866 DBG("Deleting routing table %s\n", r->name);
1867 r->config->table = NULL;
1868 if (r->hostcache)
1869 rt_free_hostcache(r);
1870 rem_node(&r->n);
1871 fib_free(&r->fib);
1872 rfree(r->rt_event);
1873 mb_free(r);
1874 config_del_obstacle(conf);
1875 }
1876 }
1877
1878 /**
1879 * rt_commit - commit new routing table configuration
1880 * @new: new configuration
1881 * @old: original configuration or %NULL if it's boot time config
1882 *
1883 * Scan differences between @old and @new configuration and modify
1884 * the routing tables according to these changes. If @new defines a
1885 * previously unknown table, create it, if it omits a table existing
1886 * in @old, schedule it for deletion (it gets deleted when all protocols
1887 * disconnect from it by calling rt_unlock_table()), if it exists
1888 * in both configurations, leave it unchanged.
1889 */
1890 void
1891 rt_commit(struct config *new, struct config *old)
1892 {
1893 struct rtable_config *o, *r;
1894
1895 DBG("rt_commit:\n");
1896 if (old)
1897 {
1898 WALK_LIST(o, old->tables)
1899 {
1900 rtable *ot = o->table;
1901 if (!ot->deleted)
1902 {
1903 struct symbol *sym = cf_find_symbol(new, o->name);
1904 if (sym && sym->class == SYM_TABLE && !new->shutdown)
1905 {
1906 DBG("\t%s: same\n", o->name);
1907 r = sym->def;
1908 r->table = ot;
1909 ot->name = r->name;
1910 ot->config = r;
1911 if (o->sorted != r->sorted)
1912 log(L_WARN "Reconfiguration of rtable sorted flag not implemented");
1913 }
1914 else
1915 {
1916 DBG("\t%s: deleted\n", o->name);
1917 ot->deleted = old;
1918 config_add_obstacle(old);
1919 rt_lock_table(ot);
1920 rt_unlock_table(ot);
1921 }
1922 }
1923 }
1924 }
1925
1926 WALK_LIST(r, new->tables)
1927 if (!r->table)
1928 {
1929 rtable *t = mb_alloc(rt_table_pool, sizeof(struct rtable));
1930 DBG("\t%s: created\n", r->name);
1931 rt_setup(rt_table_pool, t, r->name, r);
1932 add_tail(&routing_tables, &t->n);
1933 r->table = t;
1934 }
1935 DBG("\tdone\n");
1936 }
1937
1938 static inline void
1939 do_feed_baby(struct proto *p, int type, struct announce_hook *h, net *n, rte *e)
1940 {
1941 rte_update_lock();
1942 if (type == RA_ACCEPTED)
1943 rt_notify_accepted(h, n, e, NULL, NULL, p->refeeding ? 2 : 1);
1944 else if (type == RA_MERGED)
1945 rt_notify_merged(h, n, NULL, NULL, e, p->refeeding ? e : NULL, p->refeeding);
1946 else
1947 rt_notify_basic(h, n, e, p->refeeding ? e : NULL, p->refeeding);
1948 rte_update_unlock();
1949 }
1950
1951 /**
1952 * rt_feed_baby - advertise routes to a new protocol
1953 * @p: protocol to be fed
1954 *
1955 * This function performs one pass of advertisement of routes to a newly
1956 * initialized protocol. It's called by the protocol code as long as it
1957 * has something to do. (We avoid transferring all the routes in single
1958 * pass in order not to monopolize CPU time.)
1959 */
1960 int
1961 rt_feed_baby(struct proto *p)
1962 {
1963 struct announce_hook *h;
1964 struct fib_iterator *fit;
1965 int max_feed = 256;
1966
1967 if (!p->feed_ahook) /* Need to initialize first */
1968 {
1969 if (!p->ahooks)
1970 return 1;
1971 DBG("Announcing routes to new protocol %s\n", p->name);
1972 p->feed_ahook = p->ahooks;
1973 fit = p->feed_iterator = mb_alloc(p->pool, sizeof(struct fib_iterator));
1974 goto next_hook;
1975 }
1976 fit = p->feed_iterator;
1977
1978 again:
1979 h = p->feed_ahook;
1980 FIB_ITERATE_START(&h->table->fib, fit, fn)
1981 {
1982 net *n = (net *) fn;
1983 rte *e = n->routes;
1984 if (max_feed <= 0)
1985 {
1986 FIB_ITERATE_PUT(fit, fn);
1987 return 0;
1988 }
1989
1990 /* XXXX perhaps we should change feed for RA_ACCEPTED to not use 'new' */
1991
1992 if ((p->accept_ra_types == RA_OPTIMAL) ||
1993 (p->accept_ra_types == RA_ACCEPTED) ||
1994 (p->accept_ra_types == RA_MERGED))
1995 if (rte_is_valid(e))
1996 {
1997 if (p->export_state != ES_FEEDING)
1998 return 1; /* In the meantime, the protocol fell down. */
1999
2000 do_feed_baby(p, p->accept_ra_types, h, n, e);
2001 max_feed--;
2002 }
2003
2004 if (p->accept_ra_types == RA_ANY)
2005 for(e = n->routes; e; e = e->next)
2006 {
2007 if (p->export_state != ES_FEEDING)
2008 return 1; /* In the meantime, the protocol fell down. */
2009
2010 if (!rte_is_valid(e))
2011 continue;
2012
2013 do_feed_baby(p, RA_ANY, h, n, e);
2014 max_feed--;
2015 }
2016 }
2017 FIB_ITERATE_END(fn);
2018 p->feed_ahook = h->next;
2019 if (!p->feed_ahook)
2020 {
2021 mb_free(p->feed_iterator);
2022 p->feed_iterator = NULL;
2023 return 1;
2024 }
2025
2026 next_hook:
2027 h = p->feed_ahook;
2028 FIB_ITERATE_INIT(fit, &h->table->fib);
2029 goto again;
2030 }
2031
2032 /**
2033 * rt_feed_baby_abort - abort protocol feeding
2034 * @p: protocol
2035 *
2036 * This function is called by the protocol code when the protocol
2037 * stops or ceases to exist before the last iteration of rt_feed_baby()
2038 * has finished.
2039 */
2040 void
2041 rt_feed_baby_abort(struct proto *p)
2042 {
2043 if (p->feed_ahook)
2044 {
2045 /* Unlink the iterator and exit */
2046 fit_get(&p->feed_ahook->table->fib, p->feed_iterator);
2047 p->feed_ahook = NULL;
2048 }
2049 }
2050
2051
2052 static inline unsigned
2053 ptr_hash(void *ptr)
2054 {
2055 uintptr_t p = (uintptr_t) ptr;
2056 return p ^ (p << 8) ^ (p >> 16);
2057 }
2058
2059 static inline unsigned
2060 hc_hash(ip_addr a, rtable *dep)
2061 {
2062 return (ipa_hash(a) ^ ptr_hash(dep)) & 0xffff;
2063 }
2064
2065 static inline void
2066 hc_insert(struct hostcache *hc, struct hostentry *he)
2067 {
2068 uint k = he->hash_key >> hc->hash_shift;
2069 he->next = hc->hash_table[k];
2070 hc->hash_table[k] = he;
2071 }
2072
2073 static inline void
2074 hc_remove(struct hostcache *hc, struct hostentry *he)
2075 {
2076 struct hostentry **hep;
2077 uint k = he->hash_key >> hc->hash_shift;
2078
2079 for (hep = &hc->hash_table[k]; *hep != he; hep = &(*hep)->next);
2080 *hep = he->next;
2081 }
2082
2083 #define HC_DEF_ORDER 10
2084 #define HC_HI_MARK *4
2085 #define HC_HI_STEP 2
2086 #define HC_HI_ORDER 16 /* Must be at most 16 */
2087 #define HC_LO_MARK /5
2088 #define HC_LO_STEP 2
2089 #define HC_LO_ORDER 10
2090
2091 static void
2092 hc_alloc_table(struct hostcache *hc, unsigned order)
2093 {
2094 unsigned hsize = 1 << order;
2095 hc->hash_order = order;
2096 hc->hash_shift = 16 - order;
2097 hc->hash_max = (order >= HC_HI_ORDER) ? ~0 : (hsize HC_HI_MARK);
2098 hc->hash_min = (order <= HC_LO_ORDER) ? 0 : (hsize HC_LO_MARK);
2099
2100 hc->hash_table = mb_allocz(rt_table_pool, hsize * sizeof(struct hostentry *));
2101 }
2102
2103 static void
2104 hc_resize(struct hostcache *hc, unsigned new_order)
2105 {
2106 unsigned old_size = 1 << hc->hash_order;
2107 struct hostentry **old_table = hc->hash_table;
2108 struct hostentry *he, *hen;
2109 int i;
2110
2111 hc_alloc_table(hc, new_order);
2112 for (i = 0; i < old_size; i++)
2113 for (he = old_table[i]; he != NULL; he=hen)
2114 {
2115 hen = he->next;
2116 hc_insert(hc, he);
2117 }
2118 mb_free(old_table);
2119 }
2120
2121 static struct hostentry *
2122 hc_new_hostentry(struct hostcache *hc, ip_addr a, ip_addr ll, rtable *dep, unsigned k)
2123 {
2124 struct hostentry *he = sl_alloc(hc->slab);
2125
2126 he->addr = a;
2127 he->link = ll;
2128 he->tab = dep;
2129 he->hash_key = k;
2130 he->uc = 0;
2131 he->src = NULL;
2132
2133 add_tail(&hc->hostentries, &he->ln);
2134 hc_insert(hc, he);
2135
2136 hc->hash_items++;
2137 if (hc->hash_items > hc->hash_max)
2138 hc_resize(hc, hc->hash_order + HC_HI_STEP);
2139
2140 return he;
2141 }
2142
2143 static void
2144 hc_delete_hostentry(struct hostcache *hc, struct hostentry *he)
2145 {
2146 rta_free(he->src);
2147
2148 rem_node(&he->ln);
2149 hc_remove(hc, he);
2150 sl_free(hc->slab, he);
2151
2152 hc->hash_items--;
2153 if (hc->hash_items < hc->hash_min)
2154 hc_resize(hc, hc->hash_order - HC_LO_STEP);
2155 }
2156
2157 static void
2158 rt_init_hostcache(rtable *tab)
2159 {
2160 struct hostcache *hc = mb_allocz(rt_table_pool, sizeof(struct hostcache));
2161 init_list(&hc->hostentries);
2162
2163 hc->hash_items = 0;
2164 hc_alloc_table(hc, HC_DEF_ORDER);
2165 hc->slab = sl_new(rt_table_pool, sizeof(struct hostentry));
2166
2167 hc->lp = lp_new(rt_table_pool, 1008);
2168 hc->trie = f_new_trie(hc->lp, sizeof(struct f_trie_node));
2169
2170 tab->hostcache = hc;
2171 }
2172
2173 static void
2174 rt_free_hostcache(rtable *tab)
2175 {
2176 struct hostcache *hc = tab->hostcache;
2177
2178 node *n;
2179 WALK_LIST(n, hc->hostentries)
2180 {
2181 struct hostentry *he = SKIP_BACK(struct hostentry, ln, n);
2182 rta_free(he->src);
2183
2184 if (he->uc)
2185 log(L_ERR "Hostcache is not empty in table %s", tab->name);
2186 }
2187
2188 rfree(hc->slab);
2189 rfree(hc->lp);
2190 mb_free(hc->hash_table);
2191 mb_free(hc);
2192 }
2193
2194 static void
2195 rt_notify_hostcache(rtable *tab, net *net)
2196 {
2197 struct hostcache *hc = tab->hostcache;
2198
2199 if (tab->hcu_scheduled)
2200 return;
2201
2202 if (trie_match_prefix(hc->trie, net->n.prefix, net->n.pxlen))
2203 rt_schedule_hcu(tab);
2204 }
2205
2206 static int
2207 if_local_addr(ip_addr a, struct iface *i)
2208 {
2209 struct ifa *b;
2210
2211 WALK_LIST(b, i->addrs)
2212 if (ipa_equal(a, b->ip))
2213 return 1;
2214
2215 return 0;
2216 }
2217
2218 static u32
2219 rt_get_igp_metric(rte *rt)
2220 {
2221 eattr *ea = ea_find(rt->attrs->eattrs, EA_GEN_IGP_METRIC);
2222
2223 if (ea)
2224 return ea->u.data;
2225
2226 rta *a = rt->attrs;
2227
2228 #ifdef CONFIG_OSPF
2229 if ((a->source == RTS_OSPF) ||
2230 (a->source == RTS_OSPF_IA) ||
2231 (a->source == RTS_OSPF_EXT1))
2232 return rt->u.ospf.metric1;
2233 #endif
2234
2235 #ifdef CONFIG_RIP
2236 if (a->source == RTS_RIP)
2237 return rt->u.rip.metric;
2238 #endif
2239
2240 /* Device routes */
2241 if ((a->dest != RTD_ROUTER) && (a->dest != RTD_MULTIPATH))
2242 return 0;
2243
2244 return IGP_METRIC_UNKNOWN;
2245 }
2246
2247 static int
2248 rt_update_hostentry(rtable *tab, struct hostentry *he)
2249 {
2250 rta *old_src = he->src;
2251 int pxlen = 0;
2252
2253 /* Reset the hostentry */
2254 he->src = NULL;
2255 he->gw = IPA_NONE;
2256 he->dest = RTD_UNREACHABLE;
2257 he->igp_metric = 0;
2258
2259 net *n = net_route(tab, he->addr, MAX_PREFIX_LENGTH);
2260 if (n)
2261 {
2262 rte *e = n->routes;
2263 rta *a = e->attrs;
2264 pxlen = n->n.pxlen;
2265
2266 if (a->hostentry)
2267 {
2268 /* Recursive route should not depend on another recursive route */
2269 log(L_WARN "Next hop address %I resolvable through recursive route for %I/%d",
2270 he->addr, n->n.prefix, pxlen);
2271 goto done;
2272 }
2273
2274 if (a->dest == RTD_DEVICE)
2275 {
2276 if (if_local_addr(he->addr, a->iface))
2277 {
2278 /* The host address is a local address, this is not valid */
2279 log(L_WARN "Next hop address %I is a local address of iface %s",
2280 he->addr, a->iface->name);
2281 goto done;
2282 }
2283
2284 /* The host is directly reachable, use link as a gateway */
2285 he->gw = he->link;
2286 he->dest = RTD_ROUTER;
2287 }
2288 else
2289 {
2290 /* The host is reachable through some route entry */
2291 he->gw = a->gw;
2292 he->dest = a->dest;
2293 }
2294
2295 he->src = rta_clone(a);
2296 he->igp_metric = rt_get_igp_metric(e);
2297 }
2298
2299 done:
2300 /* Add a prefix range to the trie */
2301 trie_add_prefix(tab->hostcache->trie, he->addr, MAX_PREFIX_LENGTH, pxlen, MAX_PREFIX_LENGTH);
2302
2303 rta_free(old_src);
2304 return old_src != he->src;
2305 }
2306
2307 static void
2308 rt_update_hostcache(rtable *tab)
2309 {
2310 struct hostcache *hc = tab->hostcache;
2311 struct hostentry *he;
2312 node *n, *x;
2313
2314 /* Reset the trie */
2315 lp_flush(hc->lp);
2316 hc->trie = f_new_trie(hc->lp, sizeof(struct f_trie_node));
2317
2318 WALK_LIST_DELSAFE(n, x, hc->hostentries)
2319 {
2320 he = SKIP_BACK(struct hostentry, ln, n);
2321 if (!he->uc)
2322 {
2323 hc_delete_hostentry(hc, he);
2324 continue;
2325 }
2326
2327 if (rt_update_hostentry(tab, he))
2328 rt_schedule_nhu(he->tab);
2329 }
2330
2331 tab->hcu_scheduled = 0;
2332 }
2333
2334 static struct hostentry *
2335 rt_get_hostentry(rtable *tab, ip_addr a, ip_addr ll, rtable *dep)
2336 {
2337 struct hostentry *he;
2338
2339 if (!tab->hostcache)
2340 rt_init_hostcache(tab);
2341
2342 uint k = hc_hash(a, dep);
2343 struct hostcache *hc = tab->hostcache;
2344 for (he = hc->hash_table[k >> hc->hash_shift]; he != NULL; he = he->next)
2345 if (ipa_equal(he->addr, a) && (he->tab == dep))
2346 return he;
2347
2348 he = hc_new_hostentry(hc, a, ll, dep, k);
2349 rt_update_hostentry(tab, he);
2350 return he;
2351 }
2352
2353 void
2354 rta_set_recursive_next_hop(rtable *dep, rta *a, rtable *tab, ip_addr *gw, ip_addr *ll)
2355 {
2356 rta_apply_hostentry(a, rt_get_hostentry(tab, *gw, *ll, dep));
2357 }
2358
2359
2360 /*
2361 * CLI commands
2362 */
2363
2364 static byte *
2365 rt_format_via(rte *e)
2366 {
2367 rta *a = e->attrs;
2368
2369 /* Max text length w/o IP addr and interface name is 16 */
2370 static byte via[STD_ADDRESS_P_LENGTH+sizeof(a->iface->name)+16];
2371
2372 switch (a->dest)
2373 {
2374 case RTD_ROUTER: bsprintf(via, "via %I on %s", a->gw, a->iface->name); break;
2375 case RTD_DEVICE: bsprintf(via, "dev %s", a->iface->name); break;
2376 case RTD_BLACKHOLE: bsprintf(via, "blackhole"); break;
2377 case RTD_UNREACHABLE: bsprintf(via, "unreachable"); break;
2378 case RTD_PROHIBIT: bsprintf(via, "prohibited"); break;
2379 case RTD_MULTIPATH: bsprintf(via, "multipath"); break;
2380 default: bsprintf(via, "???");
2381 }
2382 return via;
2383 }
2384
2385 static void
2386 rt_show_rte(struct cli *c, byte *ia, rte *e, struct rt_show_data *d, ea_list *tmpa)
2387 {
2388 byte from[STD_ADDRESS_P_LENGTH+8];
2389 byte tm[TM_DATETIME_BUFFER_SIZE], info[256];
2390 rta *a = e->attrs;
2391 int primary = (e->net->routes == e);
2392 int sync_error = (e->net->n.flags & KRF_SYNC_ERROR);
2393 void (*get_route_info)(struct rte *, byte *buf, struct ea_list *attrs);
2394 struct mpnh *nh;
2395
2396 tm_format_datetime(tm, &config->tf_route, e->lastmod);
2397 if (ipa_nonzero(a->from) && !ipa_equal(a->from, a->gw))
2398 bsprintf(from, " from %I", a->from);
2399 else
2400 from[0] = 0;
2401
2402 get_route_info = a->src->proto->proto->get_route_info;
2403 if (get_route_info || d->verbose)
2404 {
2405 /* Need to normalize the extended attributes */
2406 ea_list *t = tmpa;
2407 t = ea_append(t, a->eattrs);
2408 tmpa = alloca(ea_scan(t));
2409 ea_merge(t, tmpa);
2410 ea_sort(tmpa);
2411 }
2412 if (get_route_info)
2413 get_route_info(e, info, tmpa);
2414 else
2415 bsprintf(info, " (%d)", e->pref);
2416 cli_printf(c, -1007, "%-18s %s [%s %s%s]%s%s", ia, rt_format_via(e), a->src->proto->name,
2417 tm, from, primary ? (sync_error ? " !" : " *") : "", info);
2418 for (nh = a->nexthops; nh; nh = nh->next)
2419 cli_printf(c, -1007, "\tvia %I on %s weight %d", nh->gw, nh->iface->name, nh->weight + 1);
2420 if (d->verbose)
2421 rta_show(c, a, tmpa);
2422 }
2423
2424 static void
2425 rt_show_net(struct cli *c, net *n, struct rt_show_data *d)
2426 {
2427 rte *e, *ee;
2428 byte ia[STD_ADDRESS_P_LENGTH+8];
2429 struct ea_list *tmpa;
2430 struct announce_hook *a = NULL;
2431 int first = 1;
2432 int pass = 0;
2433
2434 bsprintf(ia, "%I/%d", n->n.prefix, n->n.pxlen);
2435
2436 if (d->export_mode)
2437 {
2438 if (! d->export_protocol->rt_notify)
2439 return;
2440
2441 a = proto_find_announce_hook(d->export_protocol, d->table);
2442 if (!a)
2443 return;
2444 }
2445
2446 for (e = n->routes; e; e = e->next)
2447 {
2448 if (rte_is_filtered(e) != d->filtered)
2449 continue;
2450
2451 d->rt_counter++;
2452 d->net_counter += first;
2453 first = 0;
2454
2455 if (pass)
2456 continue;
2457
2458 ee = e;
2459 rte_update_lock(); /* We use the update buffer for filtering */
2460 tmpa = make_tmp_attrs(e, rte_update_pool);
2461
2462 /* Special case for merged export */
2463 if ((d->export_mode == RSEM_EXPORT) && (d->export_protocol->accept_ra_types == RA_MERGED))
2464 {
2465 rte *rt_free;
2466 e = rt_export_merged(a, n, &rt_free, &tmpa, 1);
2467 pass = 1;
2468
2469 if (!e)
2470 { e = ee; goto skip; }
2471 }
2472 else if (d->export_mode)
2473 {
2474 struct proto *ep = d->export_protocol;
2475 int ic = ep->import_control ? ep->import_control(ep, &e, &tmpa, rte_update_pool) : 0;
2476
2477 if (ep->accept_ra_types == RA_OPTIMAL || ep->accept_ra_types == RA_MERGED)
2478 pass = 1;
2479
2480 if (ic < 0)
2481 goto skip;
2482
2483 if (d->export_mode > RSEM_PREEXPORT)
2484 {
2485 /*
2486 * FIXME - This shows what should be exported according to current
2487 * filters, but not what was really exported. 'configure soft'
2488 * command may change the export filter and do not update routes.
2489 */
2490 int do_export = (ic > 0) ||
2491 (f_run(a->out_filter, &e, &tmpa, rte_update_pool, FF_FORCE_TMPATTR) <= F_ACCEPT);
2492
2493 if (do_export != (d->export_mode == RSEM_EXPORT))
2494 goto skip;
2495
2496 if ((d->export_mode == RSEM_EXPORT) && (ep->accept_ra_types == RA_ACCEPTED))
2497 pass = 1;
2498 }
2499 }
2500
2501 if (d->show_protocol && (d->show_protocol != e->attrs->src->proto))
2502 goto skip;
2503
2504 if (f_run(d->filter, &e, &tmpa, rte_update_pool, FF_FORCE_TMPATTR) > F_ACCEPT)
2505 goto skip;
2506
2507 d->show_counter++;
2508 if (d->stats < 2)
2509 rt_show_rte(c, ia, e, d, tmpa);
2510 ia[0] = 0;
2511
2512 skip:
2513 if (e != ee)
2514 {
2515 rte_free(e);
2516 e = ee;
2517 }
2518 rte_update_unlock();
2519
2520 if (d->primary_only)
2521 break;
2522 }
2523 }
2524
2525 static void
2526 rt_show_cont(struct cli *c)
2527 {
2528 struct rt_show_data *d = c->rover;
2529 #ifdef DEBUGGING
2530 unsigned max = 4;
2531 #else
2532 unsigned max = 64;
2533 #endif
2534 struct fib *fib = &d->table->fib;
2535 struct fib_iterator *it = &d->fit;
2536
2537 FIB_ITERATE_START(fib, it, f)
2538 {
2539 net *n = (net *) f;
2540 if (d->running_on_config && d->running_on_config != config)
2541 {
2542 cli_printf(c, 8004, "Stopped due to reconfiguration");
2543 goto done;
2544 }
2545 if (d->export_protocol && (d->export_protocol->export_state == ES_DOWN))
2546 {
2547 cli_printf(c, 8005, "Protocol is down");
2548 goto done;
2549 }
2550 if (!max--)
2551 {
2552 FIB_ITERATE_PUT(it, f);
2553 return;
2554 }
2555 rt_show_net(c, n, d);
2556 }
2557 FIB_ITERATE_END(f);
2558 if (d->stats)
2559 cli_printf(c, 14, "%d of %d routes for %d networks", d->show_counter, d->rt_counter, d->net_counter);
2560 else
2561 cli_printf(c, 0, "");
2562 done:
2563 c->cont = c->cleanup = NULL;
2564 }
2565
2566 static void
2567 rt_show_cleanup(struct cli *c)
2568 {
2569 struct rt_show_data *d = c->rover;
2570
2571 /* Unlink the iterator */
2572 fit_get(&d->table->fib, &d->fit);
2573 }
2574
2575 void
2576 rt_show(struct rt_show_data *d)
2577 {
2578 net *n;
2579
2580 /* Default is either a master table or a table related to a respective protocol */
2581 if (!d->table && d->export_protocol) d->table = d->export_protocol->table;
2582 if (!d->table && d->show_protocol) d->table = d->show_protocol->table;
2583 if (!d->table) d->table = config->master_rtc->table;
2584
2585 /* Filtered routes are neither exported nor have sensible ordering */
2586 if (d->filtered && (d->export_mode || d->primary_only))
2587 cli_msg(0, "");
2588
2589 if (d->pxlen == 256)
2590 {
2591 FIB_ITERATE_INIT(&d->fit, &d->table->fib);
2592 this_cli->cont = rt_show_cont;
2593 this_cli->cleanup = rt_show_cleanup;
2594 this_cli->rover = d;
2595 }
2596 else
2597 {
2598 if (d->show_for)
2599 n = net_route(d->table, d->prefix, d->pxlen);
2600 else
2601 n = net_find(d->table, d->prefix, d->pxlen);
2602
2603 if (n)
2604 rt_show_net(this_cli, n, d);
2605
2606 if (d->rt_counter)
2607 cli_msg(0, "");
2608 else
2609 cli_msg(8001, "Network not in table");
2610 }
2611 }
2612
2613 /*
2614 * Documentation for functions declared inline in route.h
2615 */
2616 #if 0
2617
2618 /**
2619 * net_find - find a network entry
2620 * @tab: a routing table
2621 * @addr: address of the network
2622 * @len: length of the network prefix
2623 *
2624 * net_find() looks up the given network in routing table @tab and
2625 * returns a pointer to its &net entry or %NULL if no such network
2626 * exists.
2627 */
2628 static inline net *net_find(rtable *tab, ip_addr addr, unsigned len)
2629 { DUMMY; }
2630
2631 /**
2632 * net_get - obtain a network entry
2633 * @tab: a routing table
2634 * @addr: address of the network
2635 * @len: length of the network prefix
2636 *
2637 * net_get() looks up the given network in routing table @tab and
2638 * returns a pointer to its &net entry. If no such entry exists, it's
2639 * created.
2640 */
2641 static inline net *net_get(rtable *tab, ip_addr addr, unsigned len)
2642 { DUMMY; }
2643
2644 /**
2645 * rte_cow - copy a route for writing
2646 * @r: a route entry to be copied
2647 *
2648 * rte_cow() takes a &rte and prepares it for modification. The exact action
2649 * taken depends on the flags of the &rte -- if it's a temporary entry, it's
2650 * just returned unchanged, else a new temporary entry with the same contents
2651 * is created.
2652 *
2653 * The primary use of this function is inside the filter machinery -- when
2654 * a filter wants to modify &rte contents (to change the preference or to
2655 * attach another set of attributes), it must ensure that the &rte is not
2656 * shared with anyone else (and especially that it isn't stored in any routing
2657 * table).
2658 *
2659 * Result: a pointer to the new writable &rte.
2660 */
2661 static inline rte * rte_cow(rte *r)
2662 { DUMMY; }
2663
2664 #endif