/* * BIRD -- Routing Tables * * (c) 1998--2000 Martin Mares * * Can be freely distributed and used under the terms of the GNU GPL. */ /** * DOC: Routing tables * * Routing tables are probably the most important structures BIRD uses. They * hold all the information about known networks, the associated routes and * their attributes. * * There are multiple routing tables (a primary one together with any * number of secondary ones if requested by the configuration). Each table * is basically a FIB containing entries describing the individual * destination networks. For each network (represented by structure &net), * there is a one-way linked list of route entries (&rte), the first entry * on the list being the best one (i.e., the one we currently use * for routing), the order of the other ones is undetermined. * * The &rte contains information specific to the route (preference, protocol * metrics, time of last modification etc.) and a pointer to a &rta structure * (see the route attribute module for a precise explanation) holding the * remaining route attributes which are expected to be shared by multiple * routes in order to conserve memory. * * There are several mechanisms that allow automatic update of routes in one * routing table (dst) as a result of changes in another routing table (src). * They handle issues of recursive next hop resolving, flowspec validation and * RPKI validation. * * The first such mechanism is handling of recursive next hops. A route in the * dst table has an indirect next hop address, which is resolved through a route * in the src table (which may also be the same table) to get an immediate next * hop. This is implemented using structure &hostcache attached to the src * table, which contains &hostentry structures for each tracked next hop * address. These structures are linked from recursive routes in dst tables, * possibly multiple routes sharing one hostentry (as many routes may have the * same indirect next hop). There is also a trie in the hostcache, which matches * all prefixes that may influence resolving of tracked next hops. * * When a best route changes in the src table, the hostcache is notified using * rt_notify_hostcache(), which immediately checks using the trie whether the * change is relevant and if it is, then it schedules asynchronous hostcache * recomputation. The recomputation is done by rt_update_hostcache() (called * from rt_event() of src table), it walks through all hostentries and resolves * them (by rt_update_hostentry()). It also updates the trie. If a change in * hostentry resolution was found, then it schedules asynchronous nexthop * recomputation of associated dst table. That is done by rt_next_hop_update() * (called from rt_event() of dst table), it iterates over all routes in the dst * table and re-examines their hostentries for changes. Note that in contrast to * hostcache update, next hop update can be interrupted by main loop. These two * full-table walks (over hostcache and dst table) are necessary due to absence * of direct lookups (route -> affected nexthop, nexthop -> its route). * * The second mechanism is for flowspec validation, where validity of flowspec * routes depends of resolving their network prefixes in IP routing tables. This * is similar to the recursive next hop mechanism, but simpler as there are no * intermediate hostcache and hostentries (because flows are less likely to * share common net prefix than routes sharing a common next hop). In src table, * there is a list of dst tables (list flowspec_links), this list is updated by * flowpsec channels (by rt_flowspec_link() and rt_flowspec_unlink() during * channel start/stop). Each dst table has its own trie of prefixes that may * influence validation of flowspec routes in it (flowspec_trie). * * When a best route changes in the src table, rt_flowspec_notify() immediately * checks all dst tables from the list using their tries to see whether the * change is relevant for them. If it is, then an asynchronous re-validation of * flowspec routes in the dst table is scheduled. That is also done by function * rt_next_hop_update(), like nexthop recomputation above. It iterates over all * flowspec routes and re-validates them. It also recalculates the trie. * * Note that in contrast to the hostcache update, here the trie is recalculated * during the rt_next_hop_update(), which may be interleaved with IP route * updates. The trie is flushed at the beginning of recalculation, which means * that such updates may use partial trie to see if they are relevant. But it * works anyway! Either affected flowspec was already re-validated and added to * the trie, then IP route change would match the trie and trigger a next round * of re-validation, or it was not yet re-validated and added to the trie, but * will be re-validated later in this round anyway. * * The third mechanism is used for RPKI re-validation of IP routes and it is the * simplest. It is just a list of subscribers in src table, who are notified * when any change happened, but only after a settle time. Also, in RPKI case * the dst is not a table, but a channel, who refeeds routes through a filter. */ #undef LOCAL_DEBUG #include "nest/bird.h" #include "nest/route.h" #include "nest/protocol.h" #include "nest/iface.h" #include "lib/resource.h" #include "lib/event.h" #include "lib/timer.h" #include "lib/string.h" #include "conf/conf.h" #include "filter/filter.h" #include "filter/data.h" #include "lib/hash.h" #include "lib/string.h" #include "lib/alloca.h" #include "lib/flowspec.h" #ifdef CONFIG_BGP #include "proto/bgp/bgp.h" #endif pool *rt_table_pool; static linpool *rte_update_pool; list routing_tables; list deleted_routing_tables; static void rt_free_hostcache(rtable *tab); static void rt_notify_hostcache(rtable *tab, net *net); static void rt_update_hostcache(rtable *tab); static void rt_next_hop_update(rtable *tab); static inline void rt_prune_table(rtable *tab); static inline void rt_schedule_notify(rtable *tab); static void rt_flowspec_notify(rtable *tab, net *net); static void rt_feed_channel(void *); const char *rt_import_state_name_array[TIS_MAX] = { [TIS_DOWN] = "DOWN", [TIS_UP] = "UP", [TIS_STOP] = "STOP", [TIS_FLUSHING] = "FLUSHING", [TIS_WAITING] = "WAITING", [TIS_CLEARED] = "CLEARED", }; const char *rt_export_state_name_array[TES_MAX] = { [TES_DOWN] = "DOWN", [TES_HUNGRY] = "HUNGRY", [TES_FEEDING] = "FEEDING", [TES_READY] = "READY", [TES_STOP] = "STOP" }; const char *rt_import_state_name(u8 state) { if (state >= TIS_MAX) return "!! INVALID !!"; else return rt_import_state_name_array[state]; } const char *rt_export_state_name(u8 state) { if (state >= TES_MAX) return "!! INVALID !!"; else return rt_export_state_name_array[state]; } static void net_init_with_trie(struct fib *f, void *N) { rtable *tab = SKIP_BACK(rtable, fib, f); net *n = N; if (tab->trie) trie_add_prefix(tab->trie, n->n.addr, n->n.addr->pxlen, n->n.addr->pxlen); if (tab->trie_new) trie_add_prefix(tab->trie_new, n->n.addr, n->n.addr->pxlen, n->n.addr->pxlen); } static inline net * net_route_ip4_trie(rtable *t, const net_addr_ip4 *n0) { TRIE_WALK_TO_ROOT_IP4(t->trie, n0, n) { net *r; if (r = net_find_valid(t, (net_addr *) &n)) return r; } TRIE_WALK_TO_ROOT_END; return NULL; } static inline net * net_route_vpn4_trie(rtable *t, const net_addr_vpn4 *n0) { TRIE_WALK_TO_ROOT_IP4(t->trie, (const net_addr_ip4 *) n0, px) { net_addr_vpn4 n = NET_ADDR_VPN4(px.prefix, px.pxlen, n0->rd); net *r; if (r = net_find_valid(t, (net_addr *) &n)) return r; } TRIE_WALK_TO_ROOT_END; return NULL; } static inline net * net_route_ip6_trie(rtable *t, const net_addr_ip6 *n0) { TRIE_WALK_TO_ROOT_IP6(t->trie, n0, n) { net *r; if (r = net_find_valid(t, (net_addr *) &n)) return r; } TRIE_WALK_TO_ROOT_END; return NULL; } static inline net * net_route_vpn6_trie(rtable *t, const net_addr_vpn6 *n0) { TRIE_WALK_TO_ROOT_IP6(t->trie, (const net_addr_ip6 *) n0, px) { net_addr_vpn6 n = NET_ADDR_VPN6(px.prefix, px.pxlen, n0->rd); net *r; if (r = net_find_valid(t, (net_addr *) &n)) return r; } TRIE_WALK_TO_ROOT_END; return NULL; } static inline void * net_route_ip6_sadr_trie(rtable *t, const net_addr_ip6_sadr *n0) { TRIE_WALK_TO_ROOT_IP6(t->trie, (const net_addr_ip6 *) n0, px) { net_addr_ip6_sadr n = NET_ADDR_IP6_SADR(px.prefix, px.pxlen, n0->src_prefix, n0->src_pxlen); net *best = NULL; int best_pxlen = 0; /* We need to do dst first matching. Since sadr addresses are hashed on dst prefix only, find the hash table chain and go through it to find the match with the longest matching src prefix. */ for (struct fib_node *fn = fib_get_chain(&t->fib, (net_addr *) &n); fn; fn = fn->next) { net_addr_ip6_sadr *a = (void *) fn->addr; if (net_equal_dst_ip6_sadr(&n, a) && net_in_net_src_ip6_sadr(&n, a) && (a->src_pxlen >= best_pxlen)) { best = fib_node_to_user(&t->fib, fn); best_pxlen = a->src_pxlen; } } if (best) return best; } TRIE_WALK_TO_ROOT_END; return NULL; } static inline net * net_route_ip4_fib(rtable *t, const net_addr_ip4 *n0) { net_addr_ip4 n; net_copy_ip4(&n, n0); net *r; while (r = net_find_valid(t, (net_addr *) &n), (!r) && (n.pxlen > 0)) { n.pxlen--; ip4_clrbit(&n.prefix, n.pxlen); } return r; } static inline net * net_route_vpn4_fib(rtable *t, const net_addr_vpn4 *n0) { net_addr_vpn4 n; net_copy_vpn4(&n, n0); net *r; while (r = net_find_valid(t, (net_addr *) &n), (!r) && (n.pxlen > 0)) { n.pxlen--; ip4_clrbit(&n.prefix, n.pxlen); } return r; } static inline net * net_route_ip6_fib(rtable *t, const net_addr_ip6 *n0) { net_addr_ip6 n; net_copy_ip6(&n, n0); net *r; while (r = net_find_valid(t, (net_addr *) &n), (!r) && (n.pxlen > 0)) { n.pxlen--; ip6_clrbit(&n.prefix, n.pxlen); } return r; } static inline net * net_route_vpn6_fib(rtable *t, const net_addr_vpn6 *n0) { net_addr_vpn6 n; net_copy_vpn6(&n, n0); net *r; while (r = net_find_valid(t, (net_addr *) &n), (!r) && (n.pxlen > 0)) { n.pxlen--; ip6_clrbit(&n.prefix, n.pxlen); } return r; } static inline void * net_route_ip6_sadr_fib(rtable *t, const net_addr_ip6_sadr *n0) { net_addr_ip6_sadr n; net_copy_ip6_sadr(&n, n0); while (1) { net *best = NULL; int best_pxlen = 0; /* We need to do dst first matching. Since sadr addresses are hashed on dst prefix only, find the hash table chain and go through it to find the match with the longest matching src prefix. */ for (struct fib_node *fn = fib_get_chain(&t->fib, (net_addr *) &n); fn; fn = fn->next) { net_addr_ip6_sadr *a = (void *) fn->addr; if (net_equal_dst_ip6_sadr(&n, a) && net_in_net_src_ip6_sadr(&n, a) && (a->src_pxlen >= best_pxlen)) { best = fib_node_to_user(&t->fib, fn); best_pxlen = a->src_pxlen; } } if (best) return best; if (!n.dst_pxlen) break; n.dst_pxlen--; ip6_clrbit(&n.dst_prefix, n.dst_pxlen); } return NULL; } net * net_route(rtable *tab, const net_addr *n) { ASSERT(tab->addr_type == n->type); switch (n->type) { case NET_IP4: if (tab->trie) return net_route_ip4_trie(tab, (net_addr_ip4 *) n); else return net_route_ip4_fib (tab, (net_addr_ip4 *) n); case NET_VPN4: if (tab->trie) return net_route_vpn4_trie(tab, (net_addr_vpn4 *) n); else return net_route_vpn4_fib (tab, (net_addr_vpn4 *) n); case NET_IP6: if (tab->trie) return net_route_ip6_trie(tab, (net_addr_ip6 *) n); else return net_route_ip6_fib (tab, (net_addr_ip6 *) n); case NET_VPN6: if (tab->trie) return net_route_vpn6_trie(tab, (net_addr_vpn6 *) n); else return net_route_vpn6_fib (tab, (net_addr_vpn6 *) n); case NET_IP6_SADR: if (tab->trie) return net_route_ip6_sadr_trie(tab, (net_addr_ip6_sadr *) n); else return net_route_ip6_sadr_fib (tab, (net_addr_ip6_sadr *) n); default: return NULL; } } static int net_roa_check_ip4_trie(rtable *tab, const net_addr_ip4 *px, u32 asn) { int anything = 0; TRIE_WALK_TO_ROOT_IP4(tab->trie, px, px0) { net_addr_roa4 roa0 = NET_ADDR_ROA4(px0.prefix, px0.pxlen, 0, 0); struct fib_node *fn; for (fn = fib_get_chain(&tab->fib, (net_addr *) &roa0); fn; fn = fn->next) { net_addr_roa4 *roa = (void *) fn->addr; net *r = fib_node_to_user(&tab->fib, fn); if (net_equal_prefix_roa4(roa, &roa0) && rte_is_valid(r->routes)) { anything = 1; if (asn && (roa->asn == asn) && (roa->max_pxlen >= px->pxlen)) return ROA_VALID; } } } TRIE_WALK_TO_ROOT_END; return anything ? ROA_INVALID : ROA_UNKNOWN; } static int net_roa_check_ip4_fib(rtable *tab, const net_addr_ip4 *px, u32 asn) { struct net_addr_roa4 n = NET_ADDR_ROA4(px->prefix, px->pxlen, 0, 0); struct fib_node *fn; int anything = 0; while (1) { for (fn = fib_get_chain(&tab->fib, (net_addr *) &n); fn; fn = fn->next) { net_addr_roa4 *roa = (void *) fn->addr; net *r = fib_node_to_user(&tab->fib, fn); if (net_equal_prefix_roa4(roa, &n) && rte_is_valid(r->routes)) { anything = 1; if (asn && (roa->asn == asn) && (roa->max_pxlen >= px->pxlen)) return ROA_VALID; } } if (n.pxlen == 0) break; n.pxlen--; ip4_clrbit(&n.prefix, n.pxlen); } return anything ? ROA_INVALID : ROA_UNKNOWN; } static int net_roa_check_ip6_trie(rtable *tab, const net_addr_ip6 *px, u32 asn) { int anything = 0; TRIE_WALK_TO_ROOT_IP6(tab->trie, px, px0) { net_addr_roa6 roa0 = NET_ADDR_ROA6(px0.prefix, px0.pxlen, 0, 0); struct fib_node *fn; for (fn = fib_get_chain(&tab->fib, (net_addr *) &roa0); fn; fn = fn->next) { net_addr_roa6 *roa = (void *) fn->addr; net *r = fib_node_to_user(&tab->fib, fn); if (net_equal_prefix_roa6(roa, &roa0) && rte_is_valid(r->routes)) { anything = 1; if (asn && (roa->asn == asn) && (roa->max_pxlen >= px->pxlen)) return ROA_VALID; } } } TRIE_WALK_TO_ROOT_END; return anything ? ROA_INVALID : ROA_UNKNOWN; } static int net_roa_check_ip6_fib(rtable *tab, const net_addr_ip6 *px, u32 asn) { struct net_addr_roa6 n = NET_ADDR_ROA6(px->prefix, px->pxlen, 0, 0); struct fib_node *fn; int anything = 0; while (1) { for (fn = fib_get_chain(&tab->fib, (net_addr *) &n); fn; fn = fn->next) { net_addr_roa6 *roa = (void *) fn->addr; net *r = fib_node_to_user(&tab->fib, fn); if (net_equal_prefix_roa6(roa, &n) && rte_is_valid(r->routes)) { anything = 1; if (asn && (roa->asn == asn) && (roa->max_pxlen >= px->pxlen)) return ROA_VALID; } } if (n.pxlen == 0) break; n.pxlen--; ip6_clrbit(&n.prefix, n.pxlen); } return anything ? ROA_INVALID : ROA_UNKNOWN; } /** * roa_check - check validity of route origination in a ROA table * @tab: ROA table * @n: network prefix to check * @asn: AS number of network prefix * * Implements RFC 6483 route validation for the given network prefix. The * procedure is to find all candidate ROAs - ROAs whose prefixes cover the given * network prefix. If there is no candidate ROA, return ROA_UNKNOWN. If there is * a candidate ROA with matching ASN and maxlen field greater than or equal to * the given prefix length, return ROA_VALID. Otherwise, return ROA_INVALID. If * caller cannot determine origin AS, 0 could be used (in that case ROA_VALID * cannot happen). Table @tab must have type NET_ROA4 or NET_ROA6, network @n * must have type NET_IP4 or NET_IP6, respectively. */ int net_roa_check(rtable *tab, const net_addr *n, u32 asn) { if ((tab->addr_type == NET_ROA4) && (n->type == NET_IP4)) { if (tab->trie) return net_roa_check_ip4_trie(tab, (const net_addr_ip4 *) n, asn); else return net_roa_check_ip4_fib (tab, (const net_addr_ip4 *) n, asn); } else if ((tab->addr_type == NET_ROA6) && (n->type == NET_IP6)) { if (tab->trie) return net_roa_check_ip6_trie(tab, (const net_addr_ip6 *) n, asn); else return net_roa_check_ip6_fib (tab, (const net_addr_ip6 *) n, asn); } else return ROA_UNKNOWN; /* Should not happen */ } /** * rte_find - find a route * @net: network node * @src: route source * * The rte_find() function returns a pointer to a route for destination @net * which is from route source @src. List end pointer is returned if no route is found. */ static struct rte_storage ** rte_find(net *net, struct rte_src *src) { struct rte_storage **e = &net->routes; while ((*e) && (*e)->rte.src != src) e = &(*e)->next; return e; } struct rte_storage * rte_store(const rte *r, net *net, rtable *tab) { struct rte_storage *e = sl_alloc(tab->rte_slab); e->rte = *r; e->rte.net = net->n.addr; rt_lock_source(e->rte.src); if (e->rte.attrs->cached) e->rte.attrs = rta_clone(e->rte.attrs); else e->rte.attrs = rta_lookup(e->rte.attrs); return e; } /** * rte_free - delete a &rte * @e: &struct rte_storage to be deleted * @tab: the table which the rte belongs to * * rte_free() deletes the given &rte from the routing table it's linked to. */ void rte_free(struct rte_storage *e) { rt_unlock_source(e->rte.src); rta_free(e->rte.attrs); sl_free(e); } static int /* Actually better or at least as good as */ rte_better(rte *new, rte *old) { int (*better)(rte *, rte *); if (!rte_is_valid(old)) return 1; if (!rte_is_valid(new)) return 0; if (new->attrs->pref > old->attrs->pref) return 1; if (new->attrs->pref < old->attrs->pref) return 0; if (new->src->proto->proto != old->src->proto->proto) { /* * If the user has configured protocol preferences, so that two different protocols * have the same preference, try to break the tie by comparing addresses. Not too * useful, but keeps the ordering of routes unambiguous. */ return new->src->proto->proto > old->src->proto->proto; } if (better = new->src->proto->rte_better) return better(new, old); return 0; } static int rte_mergable(rte *pri, rte *sec) { int (*mergable)(rte *, rte *); if (!rte_is_valid(pri) || !rte_is_valid(sec)) return 0; if (pri->attrs->pref != sec->attrs->pref) return 0; if (pri->src->proto->proto != sec->src->proto->proto) return 0; if (mergable = pri->src->proto->rte_mergable) return mergable(pri, sec); return 0; } static void rte_trace(const char *name, const rte *e, int dir, const char *msg) { log(L_TRACE "%s %c %s %N %uL %uG %s", name, dir, msg, e->net, e->src->private_id, e->src->global_id, rta_dest_name(e->attrs->dest)); } static inline void channel_rte_trace_in(uint flag, struct channel *c, const rte *e, const char *msg) { if ((c->debug & flag) || (c->proto->debug & flag)) rte_trace(c->in_req.name, e, '>', msg); } static inline void channel_rte_trace_out(uint flag, struct channel *c, const rte *e, const char *msg) { if ((c->debug & flag) || (c->proto->debug & flag)) rte_trace(c->out_req.name, e, '<', msg); } static inline void rt_rte_trace_in(uint flag, struct rt_import_request *req, const rte *e, const char *msg) { if (req->trace_routes & flag) rte_trace(req->name, e, '>', msg); } #if 0 // seems to be unused at all static inline void rt_rte_trace_out(uint flag, struct rt_export_request *req, const rte *e, const char *msg) { if (req->trace_routes & flag) rte_trace(req->name, e, '<', msg); } #endif static uint rte_feed_count(net *n) { uint count = 0; for (struct rte_storage *e = n->routes; e; e = e->next) if (rte_is_valid(RTE_OR_NULL(e))) count++; return count; } static void rte_feed_obtain(net *n, struct rte **feed, uint count) { uint i = 0; for (struct rte_storage *e = n->routes; e; e = e->next) if (rte_is_valid(RTE_OR_NULL(e))) { ASSERT_DIE(i < count); feed[i++] = &e->rte; } ASSERT_DIE(i == count); } static rte * export_filter_(struct channel *c, rte *rt, linpool *pool, int silent) { struct proto *p = c->proto; const struct filter *filter = c->out_filter; struct channel_export_stats *stats = &c->export_stats; /* Do nothing if we have already rejected the route */ if (silent && bmap_test(&c->export_reject_map, rt->id)) goto reject_noset; int v = p->preexport ? p->preexport(c, rt) : 0; if (v < 0) { if (silent) goto reject_noset; stats->updates_rejected++; if (v == RIC_REJECT) channel_rte_trace_out(D_FILTERS, c, rt, "rejected by protocol"); goto reject_noset; } if (v > 0) { if (!silent) channel_rte_trace_out(D_FILTERS, c, rt, "forced accept by protocol"); goto accept; } v = filter && ((filter == FILTER_REJECT) || (f_run(filter, rt, pool, (silent ? FF_SILENT : 0)) > F_ACCEPT)); if (v) { if (silent) goto reject; stats->updates_filtered++; channel_rte_trace_out(D_FILTERS, c, rt, "filtered out"); goto reject; } accept: /* We have accepted the route */ bmap_clear(&c->export_reject_map, rt->id); return rt; reject: /* We have rejected the route by filter */ bmap_set(&c->export_reject_map, rt->id); reject_noset: /* Discard temporary rte */ return NULL; } static inline rte * export_filter(struct channel *c, rte *rt, int silent) { return export_filter_(c, rt, rte_update_pool, silent); } static void do_rt_notify(struct channel *c, const net_addr *net, rte *new, const rte *old) { struct proto *p = c->proto; struct channel_export_stats *stats = &c->export_stats; if (c->refeeding && new) c->refeed_count++; if (!old && new) if (CHANNEL_LIMIT_PUSH(c, OUT)) { stats->updates_rejected++; channel_rte_trace_out(D_FILTERS, c, new, "rejected [limit]"); return; } if (!new && old) CHANNEL_LIMIT_POP(c, OUT); /* Apply export table */ struct rte_storage *old_exported = NULL; if (c->out_table) { if (!rte_update_out(c, net, new, old, &old_exported)) { channel_rte_trace_out(D_ROUTES, c, new, "idempotent"); return; } } if (new) stats->updates_accepted++; else stats->withdraws_accepted++; if (old) bmap_clear(&c->export_map, old->id); if (new) bmap_set(&c->export_map, new->id); if (p->debug & D_ROUTES) { if (new && old) channel_rte_trace_out(D_ROUTES, c, new, "replaced"); else if (new) channel_rte_trace_out(D_ROUTES, c, new, "added"); else if (old) channel_rte_trace_out(D_ROUTES, c, old, "removed"); } p->rt_notify(p, c, net, new, old_exported ? &old_exported->rte : old); if (c->out_table && old_exported) rte_free(old_exported); } static void rt_notify_basic(struct channel *c, const net_addr *net, rte *new, rte *old) { if (new) new = export_filter(c, new, 0); if (old && !bmap_test(&c->export_map, old->id)) old = NULL; if (!new && !old) return; do_rt_notify(c, net, new, old); } void rt_notify_accepted(struct rt_export_request *req, const net_addr *n, struct rt_pending_export *rpe, struct rte **feed, uint count) { struct channel *c = SKIP_BACK(struct channel, out_req, req); rte nb0, *new_best = NULL; const rte *old_best = NULL; for (uint i = 0; i < count; i++) { if (!rte_is_valid(feed[i])) continue; /* Has been already rejected, won't bother with it */ if (!c->refeeding && bmap_test(&c->export_reject_map, feed[i]->id)) continue; /* Previously exported */ if (!old_best && bmap_test(&c->export_map, feed[i]->id)) { /* is still best */ if (!new_best) { DBG("rt_notify_accepted: idempotent\n"); goto done; } /* is superseded */ old_best = feed[i]; break; } /* Have no new best route yet */ if (!new_best) { /* Try this route not seen before */ nb0 = *feed[i]; new_best = export_filter(c, &nb0, 0); DBG("rt_notify_accepted: checking route id %u: %s\n", feed[i]->id, new_best ? "ok" : "no"); } } /* Check obsolete routes for previously exported */ if (!old_best) if (rpe && rpe->old && bmap_test(&c->export_map, rpe->old->rte.id)) old_best = &rpe->old->rte; /* for (; rpe; rpe = atomic_load_explicit(&rpe->next, memory_order_relaxed)) { if (rpe->old && bmap_test(&hook->accept_map, rpe->old->id)) { old_best = &rpe->old.rte; break; } if (rpe == rpe_last) break; } */ /* Nothing to export */ if (!new_best && !old_best) { DBG("rt_notify_accepted: nothing to export\n"); goto done; } do_rt_notify(c, n, new_best, old_best); done: /* Drop the old stored rejection if applicable. * new->id == old->id happens when updating hostentries. */ if (rpe && rpe->old && (!rpe->new || (rpe->new->rte.id != rpe->old->rte.id))) bmap_clear(&c->export_reject_map, rpe->old->rte.id); } static struct nexthop * nexthop_merge_rta(struct nexthop *nhs, rta *a, linpool *pool, int max) { return nexthop_merge(nhs, &(a->nh), 1, 0, max, pool); } rte * rt_export_merged(struct channel *c, struct rte **feed, uint count, linpool *pool, int silent) { _Thread_local static rte rloc; // struct proto *p = c->proto; struct nexthop *nhs = NULL; rte *best0 = feed[0]; rte *best = NULL; if (!rte_is_valid(best0)) return NULL; /* Already rejected, no need to re-run the filter */ if (!c->refeeding && bmap_test(&c->export_reject_map, best0->id)) return NULL; rloc = *best0; best = export_filter_(c, &rloc, pool, silent); if (!best) /* Best route doesn't pass the filter */ return NULL; if (!rte_is_reachable(best)) /* Unreachable routes can't be merged */ return best; for (uint i = 1; i < count; i++) { if (!rte_mergable(best0, feed[i])) continue; rte tmp0 = *feed[i]; rte *tmp = export_filter_(c, &tmp0, pool, 1); if (!tmp || !rte_is_reachable(tmp)) continue; nhs = nexthop_merge_rta(nhs, tmp->attrs, pool, c->merge_limit); } if (nhs) { nhs = nexthop_merge_rta(nhs, best->attrs, pool, c->merge_limit); if (nhs->next) { best->attrs = rta_cow(best->attrs, pool); nexthop_link(best->attrs, nhs); } } return best; } void rt_notify_merged(struct rt_export_request *req, const net_addr *n, struct rt_pending_export *rpe, struct rte **feed, uint count) { struct channel *c = SKIP_BACK(struct channel, out_req, req); // struct proto *p = c->proto; #if 0 /* TODO: Find whether this check is possible when processing multiple changes at once. */ /* Check whether the change is relevant to the merged route */ if ((new_best == old_best) && (new_changed != old_changed) && !rte_mergable(new_best, new_changed) && !rte_mergable(old_best, old_changed)) return; #endif rte *old_best = NULL; /* Find old best route */ for (uint i = 0; i < count; i++) if (bmap_test(&c->export_map, feed[i]->id)) { old_best = feed[i]; break; } /* Check obsolete routes for previously exported */ if (!old_best) if (rpe && rpe->old && bmap_test(&c->export_map, rpe->old->rte.id)) old_best = &rpe->old->rte; /* for (; rpe; rpe = atomic_load_explicit(&rpe->next, memory_order_relaxed)) { if (rpe->old && bmap_test(&hook->accept_map, rpe->old->id)) { old_best = &rpe->old.rte; break; } if (rpe == rpe_last) break; } */ /* Prepare new merged route */ rte *new_merged = count ? rt_export_merged(c, feed, count, rte_update_pool, 0) : NULL; if (new_merged || old_best) do_rt_notify(c, n, new_merged, old_best); /* Drop the old stored rejection if applicable. * new->id == old->id happens when updating hostentries. */ if (rpe && rpe->old && (!rpe->new || (rpe->new->rte.id != rpe->old->rte.id))) bmap_clear(&c->export_reject_map, rpe->old->rte.id); } void rt_notify_optimal(struct rt_export_request *req, const net_addr *net, struct rt_pending_export *rpe) { struct channel *c = SKIP_BACK(struct channel, out_req, req); rte n0; if (rpe->new_best != rpe->old_best) rt_notify_basic(c, net, RTE_COPY(rpe->new_best, &n0), RTE_OR_NULL(rpe->old_best)); /* Drop the old stored rejection if applicable. * new->id == old->id happens when updating hostentries. */ if (rpe->old && (!rpe->new || (rpe->new->rte.id != rpe->old->rte.id))) bmap_clear(&c->export_reject_map, rpe->old->rte.id); } void rt_notify_any(struct rt_export_request *req, const net_addr *net, struct rt_pending_export *rpe) { struct channel *c = SKIP_BACK(struct channel, out_req, req); rte n0; if (rpe->new != rpe->old) rt_notify_basic(c, net, RTE_COPY(rpe->new, &n0), RTE_OR_NULL(rpe->old)); /* Drop the old stored rejection if applicable. * new->id == old->id happens when updating hostentries. */ if (rpe->old && (!rpe->new || (rpe->new->rte.id != rpe->old->rte.id))) bmap_clear(&c->export_reject_map, rpe->old->rte.id); } void rt_feed_any(struct rt_export_request *req, const net_addr *net, struct rt_pending_export *rpe UNUSED, rte **feed, uint count) { struct channel *c = SKIP_BACK(struct channel, out_req, req); for (uint i=0; irte.sender->stats.pref++; if (old_best) old_best->rte.sender->stats.pref--; if (tab->hostcache) rt_notify_hostcache(tab, net); if (!EMPTY_LIST(tab->flowspec_links)) rt_flowspec_notify(tab, net); } rt_schedule_notify(tab); struct rt_pending_export rpe = { .new = new, .old = old, .new_best = new_best, .old_best = old_best }; uint count = rte_feed_count(net); rte **feed = NULL; if (count) { feed = alloca(count * sizeof(rte *)); rte_feed_obtain(net, feed, count); } struct rt_export_hook *eh; WALK_LIST(eh, tab->exports) { if (eh->export_state == TES_STOP) continue; if (new) eh->stats.updates_received++; else eh->stats.withdraws_received++; if (eh->req->export_one) eh->req->export_one(eh->req, net->n.addr, &rpe); else if (eh->req->export_bulk) eh->req->export_bulk(eh->req, net->n.addr, &rpe, feed, count); else bug("Export request must always provide an export method"); } } static inline int rte_validate(struct channel *ch, rte *e) { int c; const net_addr *n = e->net; if (!net_validate(n)) { log(L_WARN "Ignoring bogus prefix %N received via %s", n, ch->proto->name); return 0; } /* FIXME: better handling different nettypes */ c = !net_is_flow(n) ? net_classify(n): (IADDR_HOST | SCOPE_UNIVERSE); if ((c < 0) || !(c & IADDR_HOST) || ((c & IADDR_SCOPE_MASK) <= SCOPE_LINK)) { log(L_WARN "Ignoring bogus route %N received via %s", n, ch->proto->name); return 0; } if (net_type_match(n, NB_DEST) == !e->attrs->dest) { /* Exception for flowspec that failed validation */ if (net_is_flow(n) && (e->attrs->dest == RTD_UNREACHABLE)) return 1; log(L_WARN "Ignoring route %N with invalid dest %d received via %s", n, e->attrs->dest, ch->proto->name); return 0; } if ((e->attrs->dest == RTD_UNICAST) && !nexthop_is_sorted(&(e->attrs->nh))) { log(L_WARN "Ignoring unsorted multipath route %N received via %s", n, ch->proto->name); return 0; } return 1; } static int rte_same(rte *x, rte *y) { /* rte.flags are not checked, as they are mostly internal to rtable */ return x->attrs == y->attrs && x->pflags == y->pflags && x->src == y->src && rte_is_filtered(x) == rte_is_filtered(y); } static inline int rte_is_ok(rte *e) { return e && !rte_is_filtered(e); } static void rte_recalculate(struct rt_import_hook *c, net *net, rte *new, struct rte_src *src) { struct rt_import_request *req = c->req; struct rtable *table = c->table; struct rt_import_stats *stats = &c->stats; struct rte_storage *old_best_stored = net->routes, *old_stored = NULL; rte *old_best = old_best_stored ? &old_best_stored->rte : NULL; rte *old = NULL; /* Find and remove original route from the same protocol */ struct rte_storage **before_old = rte_find(net, src); if (*before_old) { old = &(old_stored = (*before_old))->rte; /* If there is the same route in the routing table but from * a different sender, then there are two paths from the * source protocol to this routing table through transparent * pipes, which is not allowed. * We log that and ignore the route. */ if (old->sender != c) { if (!old->generation && !new->generation) bug("Two protocols claim to author a route with the same rte_src in table %s: %N %s/%u:%u", c->table->name, net->n.addr, old->src->proto->name, old->src->private_id, old->src->global_id); log_rl(&table->rl_pipe, L_ERR "Route source collision in table %s: %N %s/%u:%u", c->table->name, net->n.addr, old->src->proto->name, old->src->private_id, old->src->global_id); } if (new && rte_same(old, new)) { /* No changes, ignore the new route and refresh the old one */ old->flags &= ~(REF_STALE | REF_DISCARD | REF_MODIFY); if (!rte_is_filtered(new)) { stats->updates_ignored++; rt_rte_trace_in(D_ROUTES, req, new, "ignored"); } } *before_old = (*before_old)->next; table->rt_count--; } if (!old && !new) { stats->withdraws_ignored++; return; } if (req->preimport) new = req->preimport(req, new, old); int new_ok = rte_is_ok(new); int old_ok = rte_is_ok(old); if (new_ok) stats->updates_accepted++; else if (old_ok) stats->withdraws_accepted++; else stats->withdraws_ignored++; if (old_ok || new_ok) table->last_rt_change = current_time(); struct rte_storage *new_stored = new ? rte_store(new, net, table) : NULL; if (table->config->sorted) { /* If routes are sorted, just insert new route to appropriate position */ if (new_stored) { struct rte_storage **k; if ((before_old != &net->routes) && !rte_better(new, &SKIP_BACK(struct rte_storage, next, before_old)->rte)) k = before_old; else k = &net->routes; for (; *k; k=&(*k)->next) if (rte_better(new, &(*k)->rte)) break; new_stored->next = *k; *k = new_stored; table->rt_count++; } } else { /* If routes are not sorted, find the best route and move it on the first position. There are several optimized cases. */ if (src->proto->rte_recalculate && src->proto->rte_recalculate(table, net, new_stored ? &new_stored->rte : NULL, old, old_best)) goto do_recalculate; if (new_stored && rte_better(&new_stored->rte, old_best)) { /* The first case - the new route is cleary optimal, we link it at the first position */ new_stored->next = net->routes; net->routes = new_stored; table->rt_count++; } else if (old == old_best) { /* The second case - the old best route disappeared, we add the new route (if we have any) to the list (we don't care about position) and then we elect the new optimal route and relink that route at the first position and announce it. New optimal route might be NULL if there is no more routes */ do_recalculate: /* Add the new route to the list */ if (new_stored) { new_stored->next = *before_old; *before_old = new_stored; table->rt_count++; } /* Find a new optimal route (if there is any) */ if (net->routes) { struct rte_storage **bp = &net->routes; for (struct rte_storage **k=&(*bp)->next; *k; k=&(*k)->next) if (rte_better(&(*k)->rte, &(*bp)->rte)) bp = k; /* And relink it */ struct rte_storage *best = *bp; *bp = best->next; best->next = net->routes; net->routes = best; } } else if (new_stored) { /* The third case - the new route is not better than the old best route (therefore old_best != NULL) and the old best route was not removed (therefore old_best == net->routes). We just link the new route to the old/last position. */ new_stored->next = *before_old; *before_old = new_stored; table->rt_count++; } /* The fourth (empty) case - suboptimal route was removed, nothing to do */ } if (new_stored) { new_stored->rte.lastmod = current_time(); if (!old) { new_stored->rte.id = hmap_first_zero(&table->id_map); hmap_set(&table->id_map, new_stored->rte.id); } else new_stored->rte.id = old->id; } /* Log the route change */ if (new_ok) rt_rte_trace_in(D_ROUTES, req, &new_stored->rte, new_stored == net->routes ? "added [best]" : "added"); else if (old_ok) { if (old != old_best) rt_rte_trace_in(D_ROUTES, req, old, "removed"); else if (net->routes && rte_is_ok(&net->routes->rte)) rt_rte_trace_in(D_ROUTES, req, old, "removed [replaced]"); else rt_rte_trace_in(D_ROUTES, req, old, "removed [sole]"); } /* Propagate the route change */ rte_announce(table, net, new_stored, old_stored, net->routes, old_best_stored); if (!net->routes && (table->gc_counter++ >= table->config->gc_max_ops) && (table->gc_time + table->config->gc_min_time <= current_time())) rt_schedule_prune(table); #if 0 /* Enable and reimplement these callbacks if anybody wants to use them */ if (old_ok && p->rte_remove) p->rte_remove(net, old); if (new_ok && p->rte_insert) p->rte_insert(net, &new_stored->rte); #endif if (old) { if (!new_stored) hmap_clear(&table->id_map, old->id); rte_free(old_stored); } } static int rte_update_nest_cnt; /* Nesting counter to allow recursive updates */ static inline void rte_update_lock(void) { rte_update_nest_cnt++; } static inline void rte_update_unlock(void) { if (!--rte_update_nest_cnt) lp_flush(rte_update_pool); } rte * channel_preimport(struct rt_import_request *req, rte *new, rte *old) { struct channel *c = SKIP_BACK(struct channel, in_req, req); if (new && !old) if (CHANNEL_LIMIT_PUSH(c, RX)) return NULL; if (!new && old) CHANNEL_LIMIT_POP(c, RX); int new_in = new && !rte_is_filtered(new); int old_in = old && !rte_is_filtered(old); if (new_in && !old_in) if (CHANNEL_LIMIT_PUSH(c, IN)) if (c->in_keep_filtered) { new->flags |= REF_FILTERED; return new; } else return NULL; if (!new_in && old_in) CHANNEL_LIMIT_POP(c, IN); return new; } static void rte_update_direct(struct channel *c, const net_addr *n, rte *new, struct rte_src *src); void rte_update(struct channel *c, const net_addr *n, rte *new, struct rte_src *src) { if (!c->in_req.hook) return; ASSERT(c->channel_state == CS_UP); if (c->in_table && !rte_update_in(c, n, new, src)) return; return rte_update_direct(c, n, new, src); } static void rte_update_direct(struct channel *c, const net_addr *n, rte *new, struct rte_src *src) { const struct filter *filter = c->in_filter; struct channel_import_stats *stats = &c->import_stats; rte_update_lock(); if (new) { new->net = n; int fr; stats->updates_received++; if (!rte_validate(c, new)) { channel_rte_trace_in(D_FILTERS, c, new, "invalid"); stats->updates_invalid++; new = NULL; } else if ((filter == FILTER_REJECT) || ((fr = f_run(filter, new, rte_update_pool, 0)) > F_ACCEPT)) { stats->updates_filtered++; channel_rte_trace_in(D_FILTERS, c, new, "filtered out"); if (c->in_keep_filtered) new->flags |= REF_FILTERED; else new = NULL; } } else stats->withdraws_received++; rte_import(&c->in_req, n, new, src); rte_update_unlock(); } void rte_import(struct rt_import_request *req, const net_addr *n, rte *new, struct rte_src *src) { struct rt_import_hook *hook = req->hook; if (!hook) return; net *nn; if (new) { /* Use the actual struct network, not the dummy one */ nn = net_get(hook->table, n); new->net = nn->n.addr; new->sender = hook; } else if (!(nn = net_find(hook->table, n))) { req->hook->stats.withdraws_ignored++; return; } /* And recalculate the best route */ rte_recalculate(hook, nn, new, src); } /* Independent call to rte_announce(), used from next hop recalculation, outside of rte_update(). new must be non-NULL */ static inline void rte_announce_i(rtable *tab, net *net, struct rte_storage *new, struct rte_storage *old, struct rte_storage *new_best, struct rte_storage *old_best) { rte_update_lock(); rte_announce(tab, net, new, old, new_best, old_best); rte_update_unlock(); } static inline void rte_discard(net *net, rte *old) /* Non-filtered route deletion, used during garbage collection */ { rte_update_lock(); rte_recalculate(old->sender, net, NULL, old->src); rte_update_unlock(); } /* Modify existing route by protocol hook, used for long-lived graceful restart */ static inline void rte_modify(net *net, rte *old) { rte_update_lock(); rte *new = old->sender->req->rte_modify(old, rte_update_pool); if (new != old) { if (new) new->flags = old->flags & ~REF_MODIFY; rte_recalculate(old->sender, net, new, old->src); } rte_update_unlock(); } /* Check rtable for best route to given net whether it would be exported do p */ int rt_examine(rtable *t, net_addr *a, struct channel *c, const struct filter *filter) { net *n = net_find(t, a); if (!n || !rte_is_valid(n->routes)) return 0; rte rt = n->routes->rte; rte_update_lock(); /* Rest is stripped down export_filter() */ int v = c->proto->preexport ? c->proto->preexport(c, &rt) : 0; if (v == RIC_PROCESS) v = (f_run(filter, &rt, rte_update_pool, FF_SILENT) <= F_ACCEPT); rte_update_unlock(); return v > 0; } static void rt_export_stopped(void *data) { struct rt_export_hook *hook = data; rtable *tab = hook->table; /* Unlist */ rem_node(&hook->n); /* Reporting the channel as stopped. */ hook->stopped(hook->req); /* Freeing the hook together with its coroutine. */ rfree(hook->pool); rt_unlock_table(tab); DBG("Export hook %p in table %s finished uc=%u\n", hook, tab->name, tab->use_count); } static inline void rt_set_import_state(struct rt_import_hook *hook, u8 state) { hook->last_state_change = current_time(); hook->import_state = state; if (hook->req->log_state_change) hook->req->log_state_change(hook->req, state); } static inline void rt_set_export_state(struct rt_export_hook *hook, u8 state) { hook->last_state_change = current_time(); hook->export_state = state; if (hook->req->log_state_change) hook->req->log_state_change(hook->req, state); } void rt_request_import(rtable *tab, struct rt_import_request *req) { rt_lock_table(tab); struct rt_import_hook *hook = req->hook = mb_allocz(tab->rp, sizeof(struct rt_import_hook)); DBG("Lock table %s for import %p req=%p uc=%u\n", tab->name, hook, req, tab->use_count); hook->req = req; hook->table = tab; rt_set_import_state(hook, TIS_UP); hook->n = (node) {}; add_tail(&tab->imports, &hook->n); } void rt_stop_import(struct rt_import_request *req, void (*stopped)(struct rt_import_request *)) { ASSERT_DIE(req->hook); struct rt_import_hook *hook = req->hook; rt_schedule_prune(hook->table); rt_set_import_state(hook, TIS_STOP); hook->stopped = stopped; } void rt_request_export(rtable *tab, struct rt_export_request *req) { rt_lock_table(tab); pool *p = rp_new(tab->rp, "Export hook"); struct rt_export_hook *hook = req->hook = mb_allocz(p, sizeof(struct rt_export_hook)); hook->pool = p; hook->lp = lp_new_default(p); hook->req = req; hook->table = tab; /* stats zeroed by mb_allocz */ rt_set_export_state(hook, TES_HUNGRY); hook->n = (node) {}; add_tail(&tab->exports, &hook->n); FIB_ITERATE_INIT(&hook->feed_fit, &tab->fib); DBG("New export hook %p req %p in table %s uc=%u\n", hook, req, tab->name, tab->use_count); rt_set_export_state(hook, TES_FEEDING); hook->event = ev_new_init(p, rt_feed_channel, hook); ev_schedule_work(hook->event); } void rt_stop_export(struct rt_export_request *req, void (*stopped)(struct rt_export_request *)) { ASSERT_DIE(req->hook); struct rt_export_hook *hook = req->hook; rtable *tab = hook->table; /* Stop feeding */ ev_postpone(hook->event); if (hook->export_state == TES_FEEDING) fit_get(&tab->fib, &hook->feed_fit); hook->event->hook = rt_export_stopped; hook->stopped = stopped; rt_set_export_state(hook, TES_STOP); ev_schedule(hook->event); } /** * rt_refresh_begin - start a refresh cycle * @t: related routing table * @c related channel * * This function starts a refresh cycle for given routing table and announce * hook. The refresh cycle is a sequence where the protocol sends all its valid * routes to the routing table (by rte_update()). After that, all protocol * routes (more precisely routes with @c as @sender) not sent during the * refresh cycle but still in the table from the past are pruned. This is * implemented by marking all related routes as stale by REF_STALE flag in * rt_refresh_begin(), then marking all related stale routes with REF_DISCARD * flag in rt_refresh_end() and then removing such routes in the prune loop. */ void rt_refresh_begin(rtable *t, struct rt_import_request *req) { FIB_WALK(&t->fib, net, n) { for (struct rte_storage *e = n->routes; e; e = e->next) if (e->rte.sender == req->hook) e->rte.flags |= REF_STALE; } FIB_WALK_END; } /** * rt_refresh_end - end a refresh cycle * @t: related routing table * @c: related channel * * This function ends a refresh cycle for given routing table and announce * hook. See rt_refresh_begin() for description of refresh cycles. */ void rt_refresh_end(rtable *t, struct rt_import_request *req) { int prune = 0; FIB_WALK(&t->fib, net, n) { for (struct rte_storage *e = n->routes; e; e = e->next) if ((e->rte.sender == req->hook) && (e->rte.flags & REF_STALE)) { e->rte.flags |= REF_DISCARD; prune = 1; } } FIB_WALK_END; if (prune) rt_schedule_prune(t); } void rt_modify_stale(rtable *t, struct rt_import_request *req) { int prune = 0; FIB_WALK(&t->fib, net, n) { for (struct rte_storage *e = n->routes; e; e = e->next) if ((e->rte.sender == req->hook) && (e->rte.flags & REF_STALE) && !(e->rte.flags & REF_FILTERED)) { e->rte.flags |= REF_MODIFY; prune = 1; } } FIB_WALK_END; if (prune) rt_schedule_prune(t); } /** * rte_dump - dump a route * @e: &rte to be dumped * * This functions dumps contents of a &rte to debug output. */ void rte_dump(struct rte_storage *e) { debug("%-1N ", e->rte.net); debug("PF=%02x ", e->rte.pflags); rta_dump(e->rte.attrs); debug("\n"); } /** * rt_dump - dump a routing table * @t: routing table to be dumped * * This function dumps contents of a given routing table to debug output. */ void rt_dump(rtable *t) { debug("Dump of routing table <%s>%s\n", t->name, t->deleted ? " (deleted)" : ""); #ifdef DEBUGGING fib_check(&t->fib); #endif FIB_WALK(&t->fib, net, n) { for(struct rte_storage *e=n->routes; e; e=e->next) rte_dump(e); } FIB_WALK_END; debug("\n"); } /** * rt_dump_all - dump all routing tables * * This function dumps contents of all routing tables to debug output. */ void rt_dump_all(void) { rtable *t; node *n; WALK_LIST2(t, n, routing_tables, n) rt_dump(t); WALK_LIST2(t, n, deleted_routing_tables, n) rt_dump(t); } void rt_dump_hooks(rtable *tab) { debug("Dump of hooks in routing table <%s>%s\n", tab->name, tab->deleted ? " (deleted)" : ""); debug(" nhu_state=%u hcu_scheduled=%u use_count=%d rt_count=%u\n", tab->nhu_state, tab->hcu_scheduled, tab->use_count, tab->rt_count); debug(" last_rt_change=%t gc_time=%t gc_counter=%d prune_state=%u\n", tab->last_rt_change, tab->gc_time, tab->gc_counter, tab->prune_state); struct rt_import_hook *ih; WALK_LIST(ih, tab->imports) { ih->req->dump_req(ih->req); debug(" Import hook %p requested by %p: pref=%u" " last_state_change=%t import_state=%u stopped=%p\n", ih, ih->req, ih->stats.pref, ih->last_state_change, ih->import_state, ih->stopped); } struct rt_export_hook *eh; WALK_LIST(eh, tab->exports) { eh->req->dump_req(eh->req); debug(" Export hook %p requested by %p:" " refeed_pending=%u last_state_change=%t export_state=%u stopped=%p\n", eh, eh->req, eh->refeed_pending, eh->last_state_change, eh->export_state, eh->stopped); } debug("\n"); } void rt_dump_hooks_all(void) { rtable *t; node *n; debug("Dump of all table hooks\n"); WALK_LIST2(t, n, routing_tables, n) rt_dump_hooks(t); WALK_LIST2(t, n, deleted_routing_tables, n) rt_dump_hooks(t); } static inline void rt_schedule_hcu(rtable *tab) { if (tab->hcu_scheduled) return; tab->hcu_scheduled = 1; ev_schedule(tab->rt_event); } static inline void rt_schedule_nhu(rtable *tab) { if (tab->nhu_state == NHU_CLEAN) ev_schedule(tab->rt_event); /* state change: * NHU_CLEAN -> NHU_SCHEDULED * NHU_RUNNING -> NHU_DIRTY */ tab->nhu_state |= NHU_SCHEDULED; } void rt_schedule_prune(rtable *tab) { if (tab->prune_state == 0) ev_schedule(tab->rt_event); /* state change 0->1, 2->3 */ tab->prune_state |= 1; } static void rt_event(void *ptr) { rtable *tab = ptr; rt_lock_table(tab); if (tab->hcu_scheduled) rt_update_hostcache(tab); if (tab->nhu_state) rt_next_hop_update(tab); if (tab->prune_state) rt_prune_table(tab); rt_unlock_table(tab); } static inline btime rt_settled_time(rtable *tab) { ASSUME(tab->base_settle_time != 0); return MIN(tab->last_rt_change + tab->config->min_settle_time, tab->base_settle_time + tab->config->max_settle_time); } static void rt_settle_timer(timer *t) { rtable *tab = t->data; if (!tab->base_settle_time) return; btime settled_time = rt_settled_time(tab); if (current_time() < settled_time) { tm_set(tab->settle_timer, settled_time); return; } /* Settled */ tab->base_settle_time = 0; struct rt_subscription *s; WALK_LIST(s, tab->subscribers) s->hook(s); } static void rt_kick_settle_timer(rtable *tab) { tab->base_settle_time = current_time(); if (!tab->settle_timer) tab->settle_timer = tm_new_init(tab->rp, rt_settle_timer, tab, 0, 0); if (!tm_active(tab->settle_timer)) tm_set(tab->settle_timer, rt_settled_time(tab)); } static inline void rt_schedule_notify(rtable *tab) { if (EMPTY_LIST(tab->subscribers)) return; if (tab->base_settle_time) return; rt_kick_settle_timer(tab); } void rt_subscribe(rtable *tab, struct rt_subscription *s) { s->tab = tab; rt_lock_table(tab); DBG("rt_subscribe(%s)\n", tab->name); add_tail(&tab->subscribers, &s->n); } void rt_unsubscribe(struct rt_subscription *s) { rem_node(&s->n); rt_unlock_table(s->tab); } static struct rt_flowspec_link * rt_flowspec_find_link(rtable *src, rtable *dst) { struct rt_flowspec_link *ln; WALK_LIST(ln, src->flowspec_links) if ((ln->src == src) && (ln->dst == dst)) return ln; return NULL; } void rt_flowspec_link(rtable *src, rtable *dst) { ASSERT(rt_is_ip(src)); ASSERT(rt_is_flow(dst)); struct rt_flowspec_link *ln = rt_flowspec_find_link(src, dst); if (!ln) { rt_lock_table(src); rt_lock_table(dst); ln = mb_allocz(src->rp, sizeof(struct rt_flowspec_link)); ln->src = src; ln->dst = dst; add_tail(&src->flowspec_links, &ln->n); } ln->uc++; } void rt_flowspec_unlink(rtable *src, rtable *dst) { struct rt_flowspec_link *ln = rt_flowspec_find_link(src, dst); ASSERT(ln && (ln->uc > 0)); ln->uc--; if (!ln->uc) { rem_node(&ln->n); mb_free(ln); rt_unlock_table(src); rt_unlock_table(dst); } } static void rt_flowspec_notify(rtable *src, net *net) { /* Only IP tables are src links */ ASSERT(rt_is_ip(src)); struct rt_flowspec_link *ln; WALK_LIST(ln, src->flowspec_links) { rtable *dst = ln->dst; ASSERT(rt_is_flow(dst)); /* No need to inspect it further if recalculation is already active */ if ((dst->nhu_state == NHU_SCHEDULED) || (dst->nhu_state == NHU_DIRTY)) continue; if (trie_match_net(dst->flowspec_trie, net->n.addr)) rt_schedule_nhu(dst); } } static void rt_flowspec_reset_trie(rtable *tab) { linpool *lp = tab->flowspec_trie->lp; int ipv4 = tab->flowspec_trie->ipv4; lp_flush(lp); tab->flowspec_trie = f_new_trie(lp, 0); tab->flowspec_trie->ipv4 = ipv4; } static void rt_free(resource *_r) { rtable *r = (rtable *) _r; DBG("Deleting routing table %s\n", r->name); ASSERT_DIE(r->use_count == 0); if (r->internal) return; r->config->table = NULL; rem_node(&r->n); if (r->hostcache) rt_free_hostcache(r); /* Freed automagically by the resource pool fib_free(&r->fib); hmap_free(&r->id_map); rfree(r->rt_event); rfree(r->settle_timer); mb_free(r); */ } static void rt_res_dump(resource *_r) { rtable *r = (rtable *) _r; debug("name \"%s\", addr_type=%s, rt_count=%u, use_count=%d\n", r->name, net_label[r->addr_type], r->rt_count, r->use_count); } static struct resclass rt_class = { .name = "Routing table", .size = sizeof(struct rtable), .free = rt_free, .dump = rt_res_dump, .lookup = NULL, .memsize = NULL, }; rtable * rt_setup(pool *pp, struct rtable_config *cf) { pool *p = rp_newf(pp, "Routing table %s", cf->name); rtable *t = ralloc(p, &rt_class); t->rp = p; t->rte_slab = sl_new(p, sizeof(struct rte_storage)); t->name = cf->name; t->config = cf; t->addr_type = cf->addr_type; fib_init(&t->fib, p, t->addr_type, sizeof(net), OFFSETOF(net, n), 0, NULL); if (cf->trie_used) { t->trie = f_new_trie(lp_new_default(p), 0); t->trie->ipv4 = net_val_match(t->addr_type, NB_IP4 | NB_VPN4 | NB_ROA4); t->fib.init = net_init_with_trie; } init_list(&t->flowspec_links); if (!(t->internal = cf->internal)) { init_list(&t->imports); init_list(&t->exports); hmap_init(&t->id_map, p, 1024); hmap_set(&t->id_map, 0); init_list(&t->subscribers); t->rt_event = ev_new_init(p, rt_event, t); t->last_rt_change = t->gc_time = current_time(); t->rl_pipe = (struct tbf) TBF_DEFAULT_LOG_LIMITS; if (rt_is_flow(t)) { t->flowspec_trie = f_new_trie(lp_new_default(p), 0); t->flowspec_trie->ipv4 = (t->addr_type == NET_FLOW4); } } return t; } /** * rt_init - initialize routing tables * * This function is called during BIRD startup. It initializes the * routing table module. */ void rt_init(void) { rta_init(); rt_table_pool = rp_new(&root_pool, "Routing tables"); rte_update_pool = lp_new_default(rt_table_pool); init_list(&routing_tables); init_list(&deleted_routing_tables); } /** * rt_prune_table - prune a routing table * * The prune loop scans routing tables and removes routes belonging to flushing * protocols, discarded routes and also stale network entries. It is called from * rt_event(). The event is rescheduled if the current iteration do not finish * the table. The pruning is directed by the prune state (@prune_state), * specifying whether the prune cycle is scheduled or running, and there * is also a persistent pruning iterator (@prune_fit). * * The prune loop is used also for channel flushing. For this purpose, the * channels to flush are marked before the iteration and notified after the * iteration. */ static void rt_prune_table(rtable *tab) { struct fib_iterator *fit = &tab->prune_fit; int limit = 2000; struct rt_import_hook *ih; node *n, *x; DBG("Pruning route table %s\n", tab->name); #ifdef DEBUGGING fib_check(&tab->fib); #endif if (tab->prune_state == 0) return; if (tab->prune_state == 1) { /* Mark channels to flush */ WALK_LIST2(ih, n, tab->imports, n) if (ih->import_state == TIS_STOP) rt_set_import_state(ih, TIS_FLUSHING); FIB_ITERATE_INIT(fit, &tab->fib); tab->prune_state = 2; if (tab->prune_trie) { /* Init prefix trie pruning */ tab->trie_new = f_new_trie(lp_new_default(tab->rp), 0); tab->trie_new->ipv4 = tab->trie->ipv4; } } again: FIB_ITERATE_START(&tab->fib, fit, net, n) { rescan: if (limit <= 0) { FIB_ITERATE_PUT(fit); ev_schedule(tab->rt_event); return; } for (struct rte_storage *e=n->routes; e; e=e->next) { if ((e->rte.sender->import_state == TIS_FLUSHING) || (e->rte.flags & REF_DISCARD)) { rte_discard(n, &e->rte); limit--; goto rescan; } if (e->rte.flags & REF_MODIFY) { rte_modify(n, &e->rte); limit--; goto rescan; } } if (!n->routes) /* Orphaned FIB entry */ { FIB_ITERATE_PUT(fit); fib_delete(&tab->fib, n); goto again; } if (tab->trie_new) { trie_add_prefix(tab->trie_new, n->n.addr, n->n.addr->pxlen, n->n.addr->pxlen); limit--; } } FIB_ITERATE_END; #ifdef DEBUGGING fib_check(&tab->fib); #endif tab->gc_counter = 0; tab->gc_time = current_time(); /* state change 2->0, 3->1 */ tab->prune_state &= 1; if (tab->trie_new) { /* Finish prefix trie pruning */ if (!tab->trie_lock_count) { rfree(tab->trie->lp); } else { ASSERT(!tab->trie_old); tab->trie_old = tab->trie; tab->trie_old_lock_count = tab->trie_lock_count; tab->trie_lock_count = 0; } tab->trie = tab->trie_new; tab->trie_new = NULL; tab->prune_trie = 0; } else { /* Schedule prefix trie pruning */ if (tab->trie && !tab->trie_old && (tab->trie->prefix_count > (2 * tab->fib.entries))) { /* state change 0->1, 2->3 */ tab->prune_state |= 1; tab->prune_trie = 1; } } rt_prune_sources(); /* Close flushed channels */ WALK_LIST2_DELSAFE(ih, n, x, tab->imports, n) if (ih->import_state == TIS_FLUSHING) { rt_set_import_state(ih, TIS_CLEARED); ih->stopped(ih->req); rem_node(&ih->n); mb_free(ih); rt_unlock_table(tab); } } /** * rt_lock_trie - lock a prefix trie of a routing table * @tab: routing table with prefix trie to be locked * * The prune loop may rebuild the prefix trie and invalidate f_trie_walk_state * structures. Therefore, asynchronous walks should lock the prefix trie using * this function. That allows the prune loop to rebuild the trie, but postpones * its freeing until all walks are done (unlocked by rt_unlock_trie()). * * Return a current trie that will be locked, the value should be passed back to * rt_unlock_trie() for unlocking. * */ struct f_trie * rt_lock_trie(rtable *tab) { ASSERT(tab->trie); tab->trie_lock_count++; return tab->trie; } /** * rt_unlock_trie - unlock a prefix trie of a routing table * @tab: routing table with prefix trie to be locked * @trie: value returned by matching rt_lock_trie() * * Done for trie locked by rt_lock_trie() after walk over the trie is done. * It may free the trie and schedule next trie pruning. */ void rt_unlock_trie(rtable *tab, struct f_trie *trie) { ASSERT(trie); if (trie == tab->trie) { /* Unlock the current prefix trie */ ASSERT(tab->trie_lock_count); tab->trie_lock_count--; } else if (trie == tab->trie_old) { /* Unlock the old prefix trie */ ASSERT(tab->trie_old_lock_count); tab->trie_old_lock_count--; /* Free old prefix trie that is no longer needed */ if (!tab->trie_old_lock_count) { rfree(tab->trie_old->lp); tab->trie_old = NULL; /* Kick prefix trie pruning that was postponed */ if (tab->trie && (tab->trie->prefix_count > (2 * tab->fib.entries))) { tab->prune_trie = 1; rt_schedule_prune(tab); } } } else log(L_BUG "Invalid arg to rt_unlock_trie()"); } void rt_preconfig(struct config *c) { init_list(&c->tables); rt_new_table(cf_get_symbol("master4"), NET_IP4); rt_new_table(cf_get_symbol("master6"), NET_IP6); } /* * Some functions for handing internal next hop updates * triggered by rt_schedule_nhu(). */ void rta_apply_hostentry(rta *a, struct hostentry *he, mpls_label_stack *mls) { a->hostentry = he; a->dest = he->dest; a->igp_metric = he->igp_metric; if (a->dest != RTD_UNICAST) { /* No nexthop */ no_nexthop: a->nh = (struct nexthop) {}; if (mls) { /* Store the label stack for later changes */ a->nh.labels_orig = a->nh.labels = mls->len; memcpy(a->nh.label, mls->stack, mls->len * sizeof(u32)); } return; } if (((!mls) || (!mls->len)) && he->nexthop_linkable) { /* Just link the nexthop chain, no label append happens. */ memcpy(&(a->nh), &(he->src->nh), nexthop_size(&(he->src->nh))); return; } struct nexthop *nhp = NULL, *nhr = NULL; int skip_nexthop = 0; for (struct nexthop *nh = &(he->src->nh); nh; nh = nh->next) { if (skip_nexthop) skip_nexthop--; else { nhr = nhp; nhp = (nhp ? (nhp->next = lp_alloc(rte_update_pool, NEXTHOP_MAX_SIZE)) : &(a->nh)); } memset(nhp, 0, NEXTHOP_MAX_SIZE); nhp->iface = nh->iface; nhp->weight = nh->weight; if (mls) { nhp->labels = nh->labels + mls->len; nhp->labels_orig = mls->len; if (nhp->labels <= MPLS_MAX_LABEL_STACK) { memcpy(nhp->label, nh->label, nh->labels * sizeof(u32)); /* First the hostentry labels */ memcpy(&(nhp->label[nh->labels]), mls->stack, mls->len * sizeof(u32)); /* Then the bottom labels */ } else { log(L_WARN "Sum of label stack sizes %d + %d = %d exceedes allowed maximum (%d)", nh->labels, mls->len, nhp->labels, MPLS_MAX_LABEL_STACK); skip_nexthop++; continue; } } else if (nh->labels) { nhp->labels = nh->labels; nhp->labels_orig = 0; memcpy(nhp->label, nh->label, nh->labels * sizeof(u32)); } if (ipa_nonzero(nh->gw)) { nhp->gw = nh->gw; /* Router nexthop */ nhp->flags |= (nh->flags & RNF_ONLINK); } else if (!(nh->iface->flags & IF_MULTIACCESS) || (nh->iface->flags & IF_LOOPBACK)) nhp->gw = IPA_NONE; /* PtP link - no need for nexthop */ else if (ipa_nonzero(he->link)) nhp->gw = he->link; /* Device nexthop with link-local address known */ else nhp->gw = he->addr; /* Device nexthop with link-local address unknown */ } if (skip_nexthop) if (nhr) nhr->next = NULL; else { a->dest = RTD_UNREACHABLE; log(L_WARN "No valid nexthop remaining, setting route unreachable"); goto no_nexthop; } } static inline int rta_next_hop_outdated(rta *a) { struct hostentry *he = a->hostentry; if (!he) return 0; if (!he->src) return a->dest != RTD_UNREACHABLE; return (a->dest != he->dest) || (a->igp_metric != he->igp_metric) || (!he->nexthop_linkable) || !nexthop_same(&(a->nh), &(he->src->nh)); } static inline struct rte_storage * rt_next_hop_update_rte(rtable *tab, net *n, rte *old) { if (!rta_next_hop_outdated(old->attrs)) return NULL; rta *a = alloca(RTA_MAX_SIZE); memcpy(a, old->attrs, rta_size(old->attrs)); mpls_label_stack mls = { .len = a->nh.labels_orig }; memcpy(mls.stack, &a->nh.label[a->nh.labels - mls.len], mls.len * sizeof(u32)); rta_apply_hostentry(a, old->attrs->hostentry, &mls); a->cached = 0; rte e0 = *old; e0.attrs = a; return rte_store(&e0, n, tab); } #ifdef CONFIG_BGP static inline int net_flow_has_dst_prefix(const net_addr *n) { ASSUME(net_is_flow(n)); if (n->pxlen) return 1; if (n->type == NET_FLOW4) { const net_addr_flow4 *n4 = (void *) n; return (n4->length > sizeof(net_addr_flow4)) && (n4->data[0] == FLOW_TYPE_DST_PREFIX); } else { const net_addr_flow6 *n6 = (void *) n; return (n6->length > sizeof(net_addr_flow6)) && (n6->data[0] == FLOW_TYPE_DST_PREFIX); } } static inline int rta_as_path_is_empty(rta *a) { eattr *e = ea_find(a->eattrs, EA_CODE(PROTOCOL_BGP, BA_AS_PATH)); return !e || (as_path_getlen(e->u.ptr) == 0); } static inline u32 rta_get_first_asn(rta *a) { eattr *e = ea_find(a->eattrs, EA_CODE(PROTOCOL_BGP, BA_AS_PATH)); u32 asn; return (e && as_path_get_first_regular(e->u.ptr, &asn)) ? asn : 0; } int rt_flowspec_check(rtable *tab_ip, rtable *tab_flow, const net_addr *n, rta *a, int interior) { ASSERT(rt_is_ip(tab_ip)); ASSERT(rt_is_flow(tab_flow)); ASSERT(tab_ip->trie); /* RFC 8955 6. a) Flowspec has defined dst prefix */ if (!net_flow_has_dst_prefix(n)) return 0; /* RFC 9117 4.1. Accept AS_PATH is empty (fr */ if (interior && rta_as_path_is_empty(a)) return 1; /* RFC 8955 6. b) Flowspec and its best-match route have the same originator */ /* Find flowspec dst prefix */ net_addr dst; if (n->type == NET_FLOW4) net_fill_ip4(&dst, net4_prefix(n), net4_pxlen(n)); else net_fill_ip6(&dst, net6_prefix(n), net6_pxlen(n)); /* Find best-match BGP unicast route for flowspec dst prefix */ net *nb = net_route(tab_ip, &dst); const rte *rb = nb ? &nb->routes->rte : NULL; /* Register prefix to trie for tracking further changes */ int max_pxlen = (n->type == NET_FLOW4) ? IP4_MAX_PREFIX_LENGTH : IP6_MAX_PREFIX_LENGTH; trie_add_prefix(tab_flow->flowspec_trie, &dst, (nb ? nb->n.addr->pxlen : 0), max_pxlen); /* No best-match BGP route -> no flowspec */ if (!rb || (rb->attrs->source != RTS_BGP)) return 0; /* Find ORIGINATOR_ID values */ u32 orig_a = ea_get_int(a->eattrs, EA_CODE(PROTOCOL_BGP, BA_ORIGINATOR_ID), 0); u32 orig_b = ea_get_int(rb->attrs->eattrs, EA_CODE(PROTOCOL_BGP, BA_ORIGINATOR_ID), 0); /* Originator is either ORIGINATOR_ID (if present), or BGP neighbor address (if not) */ if ((orig_a != orig_b) || (!orig_a && !orig_b && !ipa_equal(a->from, rb->attrs->from))) return 0; /* Find ASN of the best-match route, for use in next checks */ u32 asn_b = rta_get_first_asn(rb->attrs); if (!asn_b) return 0; /* RFC 9117 4.2. For EBGP, flowspec and its best-match route are from the same AS */ if (!interior && (rta_get_first_asn(a) != asn_b)) return 0; /* RFC 8955 6. c) More-specific routes are from the same AS as the best-match route */ TRIE_WALK(tab_ip->trie, subnet, &dst) { net *nc = net_find_valid(tab_ip, &subnet); if (!nc) continue; const rte *rc = &nc->routes->rte; if (rc->attrs->source != RTS_BGP) return 0; if (rta_get_first_asn(rc->attrs) != asn_b) return 0; } TRIE_WALK_END; return 1; } #endif /* CONFIG_BGP */ static struct rte_storage * rt_flowspec_update_rte(rtable *tab, net *n, rte *r) { #ifdef CONFIG_BGP if (r->attrs->source != RTS_BGP) return NULL; struct bgp_channel *bc = (struct bgp_channel *) r->sender; if (!bc->base_table) return NULL; struct bgp_proto *p = (void *) r->src->proto; int valid = rt_flowspec_check(bc->base_table, tab, n->n.addr, r->attrs, p->is_interior); int dest = valid ? RTD_NONE : RTD_UNREACHABLE; if (dest == r->attrs->dest) return NULL; rta *a = alloca(RTA_MAX_SIZE); memcpy(a, r->attrs, rta_size(r->attrs)); a->dest = dest; a->cached = 0; rte new; memcpy(&new, r, sizeof(rte)); new.attrs = a; return rte_store(&new, n, tab); #else return NULL; #endif } static inline int rt_next_hop_update_net(rtable *tab, net *n) { struct rte_storage *new; int count = 0; int is_flow = net_is_flow(n->n.addr); struct rte_storage *old_best = n->routes; if (!old_best) return 0; for (struct rte_storage *e, **k = &n->routes; e = *k; k = &e->next) if (is_flow || rta_next_hop_outdated(e->rte.attrs)) count++; if (!count) return 0; struct rte_multiupdate { struct rte_storage *old, *new; } *updates = alloca(sizeof(struct rte_multiupdate) * count); int pos = 0; for (struct rte_storage *e, **k = &n->routes; e = *k; k = &e->next) if (is_flow || rta_next_hop_outdated(e->rte.attrs)) { struct rte_storage *new = is_flow ? rt_flowspec_update_rte(tab, n, &e->rte) : rt_next_hop_update_rte(tab, n, &e->rte); /* Call a pre-comparison hook */ /* Not really an efficient way to compute this */ if (e->rte.src->proto->rte_recalculate) e->rte.src->proto->rte_recalculate(tab, n, &new->rte, &e->rte, &old_best->rte); updates[pos++] = (struct rte_multiupdate) { .old = e, .new = new, }; /* Replace the route in the list */ new->next = e->next; *k = e = new; } ASSERT_DIE(pos == count); /* Find the new best route */ struct rte_storage **new_best = NULL; for (struct rte_storage *e, **k = &n->routes; e = *k; k = &e->next) { if (!new_best || rte_better(&e->rte, &(*new_best)->rte)) new_best = k; } /* Relink the new best route to the first position */ new = *new_best; if (new != n->routes) { *new_best = new->next; new->next = n->routes; n->routes = new; } /* Announce the changes */ for (int i=0; irte.sender->req, &updates[i].new->rte, best_indicator[nb][ob]); rte_announce_i(tab, n, updates[i].new, updates[i].old, new, old_best); } for (int i=0; inhu_fit; int max_feed = 32; if (tab->nhu_state == NHU_CLEAN) return; if (tab->nhu_state == NHU_SCHEDULED) { FIB_ITERATE_INIT(fit, &tab->fib); tab->nhu_state = NHU_RUNNING; if (tab->flowspec_trie) rt_flowspec_reset_trie(tab); } FIB_ITERATE_START(&tab->fib, fit, net, n) { if (max_feed <= 0) { FIB_ITERATE_PUT(fit); ev_schedule(tab->rt_event); return; } max_feed -= rt_next_hop_update_net(tab, n); } FIB_ITERATE_END; /* State change: * NHU_DIRTY -> NHU_SCHEDULED * NHU_RUNNING -> NHU_CLEAN */ tab->nhu_state &= 1; if (tab->nhu_state != NHU_CLEAN) ev_schedule(tab->rt_event); } struct rtable_config * rt_new_table(struct symbol *s, uint addr_type) { /* Hack that allows to 'redefine' the master table */ if ((s->class == SYM_TABLE) && (s->table == new_config->def_tables[addr_type]) && ((addr_type == NET_IP4) || (addr_type == NET_IP6))) return s->table; struct rtable_config *c = cfg_allocz(sizeof(struct rtable_config)); cf_define_symbol(s, SYM_TABLE, table, c); c->name = s->name; c->addr_type = addr_type; c->gc_max_ops = 1000; c->gc_min_time = 5; c->min_settle_time = 1 S; c->max_settle_time = 20 S; add_tail(&new_config->tables, &c->n); /* First table of each type is kept as default */ if (! new_config->def_tables[addr_type]) new_config->def_tables[addr_type] = c; return c; } /** * rt_lock_table - lock a routing table * @r: routing table to be locked * * Lock a routing table, because it's in use by a protocol, * preventing it from being freed when it gets undefined in a new * configuration. */ void rt_lock_table(rtable *r) { r->use_count++; } /** * rt_unlock_table - unlock a routing table * @r: routing table to be unlocked * * Unlock a routing table formerly locked by rt_lock_table(), * that is decrease its use count and delete it if it's scheduled * for deletion by configuration changes. */ void rt_unlock_table(rtable *r) { if (!--r->use_count && r->deleted) { struct config *conf = r->deleted; /* Delete the routing table by freeing its pool */ rt_shutdown(r); config_del_obstacle(conf); } } static int rt_reconfigure(rtable *tab, struct rtable_config *new, struct rtable_config *old) { if ((new->addr_type != old->addr_type) || (new->sorted != old->sorted) || (new->trie_used != old->trie_used)) return 0; DBG("\t%s: same\n", new->name); new->table = tab; tab->name = new->name; tab->config = new; return 1; } static struct rtable_config * rt_find_table_config(struct config *cf, char *name) { struct symbol *sym = cf_find_symbol(cf, name); return (sym && (sym->class == SYM_TABLE)) ? sym->table : NULL; } /** * rt_commit - commit new routing table configuration * @new: new configuration * @old: original configuration or %NULL if it's boot time config * * Scan differences between @old and @new configuration and modify * the routing tables according to these changes. If @new defines a * previously unknown table, create it, if it omits a table existing * in @old, schedule it for deletion (it gets deleted when all protocols * disconnect from it by calling rt_unlock_table()), if it exists * in both configurations, leave it unchanged. */ void rt_commit(struct config *new, struct config *old) { struct rtable_config *o, *r; DBG("rt_commit:\n"); if (old) { WALK_LIST(o, old->tables) { rtable *tab = o->table; if (tab->deleted) continue; r = rt_find_table_config(new, o->name); if (r && !new->shutdown && rt_reconfigure(tab, r, o)) continue; DBG("\t%s: deleted\n", o->name); tab->deleted = old; config_add_obstacle(old); rt_lock_table(tab); rt_unlock_table(tab); } } WALK_LIST(r, new->tables) if (!r->table) { r->table = rt_setup(rt_table_pool, r); DBG("\t%s: created\n", r->name); add_tail(&routing_tables, &r->table->n); } DBG("\tdone\n"); } /** * rt_feed_channel - advertise all routes to a channel * @c: channel to be fed * * This function performs one pass of advertisement of routes to a channel that * is in the ES_FEEDING state. It is called by the protocol code as long as it * has something to do. (We avoid transferring all the routes in single pass in * order not to monopolize CPU time.) */ static void rt_feed_channel(void *data) { struct rt_export_hook *c = data; struct fib_iterator *fit = &c->feed_fit; int max_feed = 256; ASSERT(c->export_state == TES_FEEDING); FIB_ITERATE_START(&c->table->fib, fit, net, n) { if (max_feed <= 0) { FIB_ITERATE_PUT(fit); ev_schedule_work(c->event); return; } if (c->export_state != TES_FEEDING) goto done; if (c->req->export_bulk) { uint count = rte_feed_count(n); if (count) { rte_update_lock(); rte **feed = alloca(count * sizeof(rte *)); rte_feed_obtain(n, feed, count); struct rt_pending_export rpe = { .new_best = n->routes }; c->req->export_bulk(c->req, n->n.addr, &rpe, feed, count); max_feed -= count; rte_update_unlock(); } } else if (n->routes && rte_is_valid(&n->routes->rte)) { rte_update_lock(); struct rt_pending_export rpe = { .new = n->routes, .new_best = n->routes }; c->req->export_one(c->req, n->n.addr, &rpe); max_feed--; rte_update_unlock(); } } FIB_ITERATE_END; done: rt_set_export_state(c, TES_READY); } /* * Import table */ int rte_update_in(struct channel *c, const net_addr *n, rte *new, struct rte_src *src) { struct rtable *tab = c->in_table; net *net; if (new) net = net_get(tab, n); else { net = net_find(tab, n); if (!net) goto drop_withdraw; } /* Find the old rte */ struct rte_storage **pos = rte_find(net, src); if (*pos) { rte *old = &(*pos)->rte; if (new && rte_same(old, new)) { /* Refresh the old rte, continue with update to main rtable */ if (old->flags & (REF_STALE | REF_DISCARD | REF_MODIFY)) { old->flags &= ~(REF_STALE | REF_DISCARD | REF_MODIFY); return 1; } goto drop_update; } if (!new) CHANNEL_LIMIT_POP(c, RX); /* Move iterator if needed */ if (*pos == c->reload_next_rte) c->reload_next_rte = (*pos)->next; /* Remove the old rte */ struct rte_storage *del = *pos; *pos = (*pos)->next; rte_free(del); tab->rt_count--; } else if (new) { if (CHANNEL_LIMIT_PUSH(c, RX)) { /* Required by rte_trace_in() */ new->net = n; channel_rte_trace_in(D_FILTERS, c, new, "ignored [limit]"); goto drop_update; } } else goto drop_withdraw; if (!new) { if (!net->routes) fib_delete(&tab->fib, net); return 1; } /* Insert the new rte */ struct rte_storage *e = rte_store(new, net, tab); e->rte.lastmod = current_time(); e->next = *pos; *pos = e; tab->rt_count++; return 1; drop_update: c->import_stats.updates_received++; c->in_req.hook->stats.updates_ignored++; if (!net->routes) fib_delete(&tab->fib, net); return 0; drop_withdraw: c->import_stats.withdraws_received++; c->in_req.hook->stats.withdraws_ignored++; return 0; } int rt_reload_channel(struct channel *c) { struct rtable *tab = c->in_table; struct fib_iterator *fit = &c->reload_fit; int max_feed = 64; ASSERT(c->channel_state == CS_UP); if (!c->reload_active) { FIB_ITERATE_INIT(fit, &tab->fib); c->reload_active = 1; } do { for (struct rte_storage *e = c->reload_next_rte; e; e = e->next) { if (max_feed-- <= 0) { c->reload_next_rte = e; debug("%s channel reload burst split (max_feed=%d)", c->proto->name, max_feed); return 0; } rte r = e->rte; rte_update_direct(c, r.net, &r, r.src); } c->reload_next_rte = NULL; FIB_ITERATE_START(&tab->fib, fit, net, n) { if (c->reload_next_rte = n->routes) { FIB_ITERATE_PUT_NEXT(fit, &tab->fib); break; } } FIB_ITERATE_END; } while (c->reload_next_rte); c->reload_active = 0; return 1; } void rt_reload_channel_abort(struct channel *c) { if (c->reload_active) { /* Unlink the iterator */ fit_get(&c->in_table->fib, &c->reload_fit); c->reload_next_rte = NULL; c->reload_active = 0; } } void rt_prune_sync(rtable *t, int all) { struct fib_iterator fit; FIB_ITERATE_INIT(&fit, &t->fib); again: FIB_ITERATE_START(&t->fib, &fit, net, n) { struct rte_storage *e, **ee = &n->routes; while (e = *ee) { if (all || (e->rte.flags & (REF_STALE | REF_DISCARD))) { *ee = e->next; rte_free(e); t->rt_count--; } else ee = &e->next; } if (all || !n->routes) { FIB_ITERATE_PUT(&fit); fib_delete(&t->fib, n); goto again; } } FIB_ITERATE_END; } /* * Export table */ int rte_update_out(struct channel *c, const net_addr *n, rte *new, const rte *old0, struct rte_storage **old_exported) { struct rtable *tab = c->out_table; struct rte_src *src; net *net; if (new) { net = net_get(tab, n); src = new->src; } else { net = net_find(tab, n); src = old0->src; if (!net) goto drop; } /* Find the old rte */ struct rte_storage **pos = (c->ra_mode == RA_ANY) ? rte_find(net, src) : &net->routes; struct rte_storage *old = NULL; if (old = *pos) { if (new && rte_same(&(*pos)->rte, new)) goto drop; /* Remove the old rte */ *pos = old->next; *old_exported = old; tab->rt_count--; } if (!new) { if (!old) goto drop; if (!net->routes) fib_delete(&tab->fib, net); return 1; } /* Insert the new rte */ struct rte_storage *e = rte_store(new, net, tab); e->rte.lastmod = current_time(); e->next = *pos; *pos = e; tab->rt_count++; return 1; drop: return 0; } void rt_refeed_channel(struct channel *c) { if (!c->out_table) { channel_request_feeding(c); return; } ASSERT_DIE(c->ra_mode != RA_ANY); c->proto->feed_begin(c, 0); FIB_WALK(&c->out_table->fib, net, n) { if (!n->routes) continue; rte e = n->routes->rte; c->proto->rt_notify(c->proto, c, n->n.addr, &e, NULL); } FIB_WALK_END; c->proto->feed_end(c); } /* * Hostcache */ static inline u32 hc_hash(ip_addr a, rtable *dep) { return ipa_hash(a) ^ ptr_hash(dep); } static inline void hc_insert(struct hostcache *hc, struct hostentry *he) { uint k = he->hash_key >> hc->hash_shift; he->next = hc->hash_table[k]; hc->hash_table[k] = he; } static inline void hc_remove(struct hostcache *hc, struct hostentry *he) { struct hostentry **hep; uint k = he->hash_key >> hc->hash_shift; for (hep = &hc->hash_table[k]; *hep != he; hep = &(*hep)->next); *hep = he->next; } #define HC_DEF_ORDER 10 #define HC_HI_MARK *4 #define HC_HI_STEP 2 #define HC_HI_ORDER 16 /* Must be at most 16 */ #define HC_LO_MARK /5 #define HC_LO_STEP 2 #define HC_LO_ORDER 10 static void hc_alloc_table(struct hostcache *hc, pool *p, unsigned order) { uint hsize = 1 << order; hc->hash_order = order; hc->hash_shift = 32 - order; hc->hash_max = (order >= HC_HI_ORDER) ? ~0U : (hsize HC_HI_MARK); hc->hash_min = (order <= HC_LO_ORDER) ? 0U : (hsize HC_LO_MARK); hc->hash_table = mb_allocz(p, hsize * sizeof(struct hostentry *)); } static void hc_resize(struct hostcache *hc, pool *p, unsigned new_order) { struct hostentry **old_table = hc->hash_table; struct hostentry *he, *hen; uint old_size = 1 << hc->hash_order; uint i; hc_alloc_table(hc, p, new_order); for (i = 0; i < old_size; i++) for (he = old_table[i]; he != NULL; he=hen) { hen = he->next; hc_insert(hc, he); } mb_free(old_table); } static struct hostentry * hc_new_hostentry(struct hostcache *hc, pool *p, ip_addr a, ip_addr ll, rtable *dep, unsigned k) { struct hostentry *he = sl_alloc(hc->slab); *he = (struct hostentry) { .addr = a, .link = ll, .tab = dep, .hash_key = k, }; add_tail(&hc->hostentries, &he->ln); hc_insert(hc, he); hc->hash_items++; if (hc->hash_items > hc->hash_max) hc_resize(hc, p, hc->hash_order + HC_HI_STEP); return he; } static void hc_delete_hostentry(struct hostcache *hc, pool *p, struct hostentry *he) { rta_free(he->src); rem_node(&he->ln); hc_remove(hc, he); sl_free(he); hc->hash_items--; if (hc->hash_items < hc->hash_min) hc_resize(hc, p, hc->hash_order - HC_LO_STEP); } static void rt_init_hostcache(rtable *tab) { struct hostcache *hc = mb_allocz(tab->rp, sizeof(struct hostcache)); init_list(&hc->hostentries); hc->hash_items = 0; hc_alloc_table(hc, tab->rp, HC_DEF_ORDER); hc->slab = sl_new(tab->rp, sizeof(struct hostentry)); hc->lp = lp_new(tab->rp); hc->trie = f_new_trie(hc->lp, 0); tab->hostcache = hc; } static void rt_free_hostcache(rtable *tab) { struct hostcache *hc = tab->hostcache; node *n; WALK_LIST(n, hc->hostentries) { struct hostentry *he = SKIP_BACK(struct hostentry, ln, n); rta_free(he->src); if (he->uc) log(L_ERR "Hostcache is not empty in table %s", tab->name); } /* Freed automagically by the resource pool rfree(hc->slab); rfree(hc->lp); mb_free(hc->hash_table); mb_free(hc); */ } static void rt_notify_hostcache(rtable *tab, net *net) { if (tab->hcu_scheduled) return; if (trie_match_net(tab->hostcache->trie, net->n.addr)) rt_schedule_hcu(tab); } static int if_local_addr(ip_addr a, struct iface *i) { struct ifa *b; WALK_LIST(b, i->addrs) if (ipa_equal(a, b->ip)) return 1; return 0; } u32 rt_get_igp_metric(rte *rt) { eattr *ea = ea_find(rt->attrs->eattrs, EA_GEN_IGP_METRIC); if (ea) return ea->u.data; if (rt->attrs->source == RTS_DEVICE) return 0; if (rt->src->proto->rte_igp_metric) return rt->src->proto->rte_igp_metric(rt); return IGP_METRIC_UNKNOWN; } static int rt_update_hostentry(rtable *tab, struct hostentry *he) { rta *old_src = he->src; int direct = 0; int pxlen = 0; /* Reset the hostentry */ he->src = NULL; he->dest = RTD_UNREACHABLE; he->nexthop_linkable = 0; he->igp_metric = 0; net_addr he_addr; net_fill_ip_host(&he_addr, he->addr); net *n = net_route(tab, &he_addr); if (n) { struct rte_storage *e = n->routes; rta *a = e->rte.attrs; pxlen = n->n.addr->pxlen; if (a->hostentry) { /* Recursive route should not depend on another recursive route */ log(L_WARN "Next hop address %I resolvable through recursive route for %N", he->addr, n->n.addr); goto done; } if (a->dest == RTD_UNICAST) { for (struct nexthop *nh = &(a->nh); nh; nh = nh->next) if (ipa_zero(nh->gw)) { if (if_local_addr(he->addr, nh->iface)) { /* The host address is a local address, this is not valid */ log(L_WARN "Next hop address %I is a local address of iface %s", he->addr, nh->iface->name); goto done; } direct++; } } he->src = rta_clone(a); he->dest = a->dest; he->nexthop_linkable = !direct; he->igp_metric = rt_get_igp_metric(&e->rte); } done: /* Add a prefix range to the trie */ trie_add_prefix(tab->hostcache->trie, &he_addr, pxlen, he_addr.pxlen); rta_free(old_src); return old_src != he->src; } static void rt_update_hostcache(rtable *tab) { struct hostcache *hc = tab->hostcache; struct hostentry *he; node *n, *x; /* Reset the trie */ lp_flush(hc->lp); hc->trie = f_new_trie(hc->lp, 0); WALK_LIST_DELSAFE(n, x, hc->hostentries) { he = SKIP_BACK(struct hostentry, ln, n); if (!he->uc) { hc_delete_hostentry(hc, tab->rp, he); continue; } if (rt_update_hostentry(tab, he)) rt_schedule_nhu(he->tab); } tab->hcu_scheduled = 0; } struct hostentry * rt_get_hostentry(rtable *tab, ip_addr a, ip_addr ll, rtable *dep) { struct hostentry *he; if (!tab->hostcache) rt_init_hostcache(tab); u32 k = hc_hash(a, dep); struct hostcache *hc = tab->hostcache; for (he = hc->hash_table[k >> hc->hash_shift]; he != NULL; he = he->next) if (ipa_equal(he->addr, a) && (he->tab == dep)) return he; he = hc_new_hostentry(hc, tab->rp, a, ipa_zero(ll) ? a : ll, dep, k); rt_update_hostentry(tab, he); return he; } /* * Documentation for functions declared inline in route.h */ #if 0 /** * net_find - find a network entry * @tab: a routing table * @addr: address of the network * * net_find() looks up the given network in routing table @tab and * returns a pointer to its &net entry or %NULL if no such network * exists. */ static inline net *net_find(rtable *tab, net_addr *addr) { DUMMY; } /** * net_get - obtain a network entry * @tab: a routing table * @addr: address of the network * * net_get() looks up the given network in routing table @tab and * returns a pointer to its &net entry. If no such entry exists, it's * created. */ static inline net *net_get(rtable *tab, net_addr *addr) { DUMMY; } /** * rte_cow - copy a route for writing * @r: a route entry to be copied * * rte_cow() takes a &rte and prepares it for modification. The exact action * taken depends on the flags of the &rte -- if it's a temporary entry, it's * just returned unchanged, else a new temporary entry with the same contents * is created. * * The primary use of this function is inside the filter machinery -- when * a filter wants to modify &rte contents (to change the preference or to * attach another set of attributes), it must ensure that the &rte is not * shared with anyone else (and especially that it isn't stored in any routing * table). * * Result: a pointer to the new writable &rte. */ static inline rte * rte_cow(rte *r) { DUMMY; } #endif