0
0
mirror of https://gitlab.nic.cz/labs/bird.git synced 2024-09-19 20:05:21 +00:00
bird/proto/perf/perf.c
Maria Matejka 950775f6fa Route destination field merged with nexthop attribute; splitting flowspec validation result out.
As there is either a nexthop or another destination specification
(or othing in case of ROAs and Flowspec), it may be merged together.
This code is somehow quirky and should be replaced in future by better
implementation of nexthop.

Also flowspec validation result has its own attribute now as it doesn't
have anything to do with route nexthop.
2022-05-30 14:39:09 +02:00

327 lines
7.1 KiB
C

/*
* BIRD -- Table-to-Table Routing Protocol a.k.a Pipe
*
* (c) 1999--2000 Martin Mares <mj@ucw.cz>
*
* Can be freely distributed and used under the terms of the GNU GPL.
*/
/**
* DOC: Perf
*
* Run this protocol to measure route import and export times.
* Generates a load of dummy routes and measures time to import.
*/
#undef LOCAL_DEBUG
#include "nest/bird.h"
#include "nest/iface.h"
#include "nest/protocol.h"
#include "nest/rt.h"
#include "nest/cli.h"
#include "conf/conf.h"
#include "filter/filter.h"
#include "lib/string.h"
#include "perf.h"
#include <stdlib.h>
#include <time.h>
#define PLOG(msg, ...) log(L_INFO "Perf %s %s " msg, BIRD_VERSION, p->p.name, ##__VA_ARGS__)
static inline void
random_data(void *p, uint len)
{
uint ints = (len + sizeof(int) - 1) / sizeof(int);
int *d = alloca(sizeof(uint) * ints);
for (uint i=0; i<ints; i++)
d[i] = random();
memcpy(p, d, len);
}
static ip_addr
random_gw(net_addr *prefix)
{
ASSERT(net_is_ip(prefix));
ip_addr px = net_prefix(prefix);
ip_addr mask = net_pxmask(prefix);
ip_addr out;
random_data(&out, sizeof(ip_addr));
if (ipa_is_ip4(px))
out = ipa_and(out, ipa_from_ip4(ip4_mkmask(32)));
return ipa_or(ipa_and(px, mask), ipa_and(out, ipa_not(mask)));
}
static net_addr_ip4
random_net_ip4(void)
{
u32 x; random_data(&x, sizeof(u32));
x &= ((1 << 20) - 1);
uint pxlen = u32_log2(x) + 5;
ip4_addr px; random_data(&px, sizeof(ip4_addr));
net_addr_ip4 out = {
.type = NET_IP4,
.pxlen = pxlen,
.length = sizeof(net_addr_ip4),
.prefix = ip4_and(ip4_mkmask(pxlen), px),
};
if (!net_validate((net_addr *) &out))
return random_net_ip4();
int c = net_classify((net_addr *) &out);
if ((c < 0) || !(c & IADDR_HOST) || ((c & IADDR_SCOPE_MASK) <= SCOPE_LINK))
return random_net_ip4();
return out;
}
struct perf_random_routes {
struct rta *a;
net_addr net;
};
//static const uint perf_random_routes_size = sizeof(struct perf_random_routes) + (RTA_MAX_SIZE - sizeof(struct rta));
static inline s64 timediff(struct timespec *begin, struct timespec *end)
{ return (end->tv_sec - begin->tv_sec) * (s64) 1000000000 + end->tv_nsec - begin->tv_nsec; }
static void
perf_ifa_notify(struct proto *P, uint flags, struct ifa *ad)
{
struct perf_proto *p = (struct perf_proto *) P;
if (ad->flags & IA_SECONDARY)
return;
if (p->ifa && p->ifa == ad && (flags & IF_CHANGE_DOWN)) {
p->ifa = NULL;
if (ev_active(p->loop))
ev_postpone(p->loop);
return;
}
if (!p->ifa && (flags & IF_CHANGE_UP)) {
p->ifa = ad;
ev_schedule(p->loop);
PLOG("starting");
return;
}
}
static void
perf_loop(void *data)
{
struct proto *P = data;
struct perf_proto *p = data;
const uint N = 1U << p->exp;
if (!p->run) {
ASSERT(p->data == NULL);
p->data = xmalloc(sizeof(struct perf_random_routes) * N);
p->stop = 1;
}
ip_addr gw = random_gw(&p->ifa->prefix);
struct timespec ts_begin, ts_generated, ts_update, ts_withdraw;
clock_gettime(CLOCK_MONOTONIC, &ts_begin);
for (uint i=0; i<N; i++) {
*((net_addr_ip4 *) &(p->data[i].net)) = random_net_ip4();
if (!p->attrs_per_rte || !(i % p->attrs_per_rte)) {
struct rta a0 = {};
ea_set_attr_u32(&a0.eattrs, &ea_gen_preference, 0, p->p.main_channel->preference);
ea_set_attr_u32(&a0.eattrs, &ea_gen_source, 0, RTS_PERF);
struct nexthop_adata nhad = {
.nh.iface = p->ifa->iface,
.nh.gw = gw,
.nh.weight = 1,
};
ea_set_attr_data(&a0.eattrs, &ea_gen_nexthop, 0,
&nhad.ad.data, sizeof nhad - sizeof nhad.ad);
p->data[i].a = rta_lookup(&a0);
}
else
p->data[i].a = rta_clone(p->data[i-1].a);
}
clock_gettime(CLOCK_MONOTONIC, &ts_generated);
for (uint i=0; i<N; i++) {
rte *e = rte_get_temp(p->data[i].a, p->p.main_source);
e->pflags = 0;
rte_update(P, &(p->data[i].net), e);
}
clock_gettime(CLOCK_MONOTONIC, &ts_update);
if (!p->keep)
for (uint i=0; i<N; i++)
rte_update(P, &(p->data[i].net), NULL);
clock_gettime(CLOCK_MONOTONIC, &ts_withdraw);
s64 gentime = timediff(&ts_begin, &ts_generated);
s64 updatetime = timediff(&ts_generated, &ts_update);
s64 withdrawtime = timediff(&ts_update, &ts_withdraw);
if (updatetime NS >= p->threshold_min)
PLOG("exp=%u times: gen=%ld update=%ld withdraw=%ld",
p->exp, gentime, updatetime, withdrawtime);
if (updatetime NS < p->threshold_max)
p->stop = 0;
if ((updatetime NS < p->threshold_min) || (++p->run == p->repeat)) {
xfree(p->data);
p->data = NULL;
if (p->stop || (p->exp == p->to)) {
PLOG("done with exp=%u", p->exp);
return;
}
p->run = 0;
p->exp++;
}
rt_schedule_prune(P->main_channel->table);
ev_schedule(p->loop);
}
static void
perf_rt_notify(struct proto *P, struct channel *c UNUSED, struct network *net UNUSED, struct rte *new UNUSED, struct rte *old UNUSED)
{
struct perf_proto *p = (struct perf_proto *) P;
p->exp++;
return;
}
static void
perf_feed_begin(struct channel *c, int initial UNUSED)
{
struct perf_proto *p = (struct perf_proto *) c->proto;
p->run++;
p->feed_begin = xmalloc(sizeof(struct timespec));
p->exp = 0;
clock_gettime(CLOCK_MONOTONIC, p->feed_begin);
}
static void
perf_feed_end(struct channel *c)
{
struct perf_proto *p = (struct perf_proto *) c->proto;
struct timespec ts_end;
clock_gettime(CLOCK_MONOTONIC, &ts_end);
s64 feedtime = timediff(p->feed_begin, &ts_end);
PLOG("feed n=%lu time=%lu", p->exp, feedtime);
xfree(p->feed_begin);
p->feed_begin = NULL;
if (p->run < p->repeat)
channel_request_feeding(c);
else
PLOG("feed done");
}
static struct proto *
perf_init(struct proto_config *CF)
{
struct proto *P = proto_new(CF);
P->main_channel = proto_add_channel(P, proto_cf_main_channel(CF));
struct perf_proto *p = (struct perf_proto *) P;
p->loop = ev_new_init(P->pool, perf_loop, p);
struct perf_config *cf = (struct perf_config *) CF;
p->threshold_min = cf->threshold_min;
p->threshold_max = cf->threshold_max;
p->from = cf->from;
p->to = cf->to;
p->repeat = cf->repeat;
p->keep = cf->keep;
p->mode = cf->mode;
p->attrs_per_rte = cf->attrs_per_rte;
switch (p->mode) {
case PERF_MODE_IMPORT:
P->ifa_notify = perf_ifa_notify;
break;
case PERF_MODE_EXPORT:
P->rt_notify = perf_rt_notify;
P->feed_begin = perf_feed_begin;
P->feed_end = perf_feed_end;
break;
}
return P;
}
static int
perf_start(struct proto *P)
{
struct perf_proto *p = (struct perf_proto *) P;
p->ifa = NULL;
p->run = 0;
p->exp = p->from;
ASSERT(p->data == NULL);
return PS_UP;
}
static int
perf_reconfigure(struct proto *P UNUSED, struct proto_config *CF UNUSED)
{
return 0;
}
static void
perf_copy_config(struct proto_config *dest UNUSED, struct proto_config *src UNUSED)
{
}
struct protocol proto_perf = {
.name = "Perf",
.template = "perf%d",
.channel_mask = NB_IP,
.proto_size = sizeof(struct perf_proto),
.config_size = sizeof(struct perf_config),
.init = perf_init,
.start = perf_start,
.reconfigure = perf_reconfigure,
.copy_config = perf_copy_config,
};
void
perf_build(void)
{
proto_build(&proto_perf);
}