0
0
mirror of https://gitlab.nic.cz/labs/bird.git synced 2024-12-23 10:11:53 +00:00
bird/proto/snmp/subagent.c

1630 lines
43 KiB
C
Raw Normal View History

/*
* BIRD -- Simple Network Management Protocol (SNMP)
*
* (c) 2022 Vojtech Vilimek <vojtech.vilimek@nic.cz>
* (c) 2022 CZ.NIC z.s.p.o
*
* Can be freely distributed and used under the terms of the GNU GPL.
*
*/
2022-08-10 15:31:32 +00:00
#include "lib/unaligned.h"
#include "subagent.h"
#include "snmp_utils.h"
2022-09-20 12:28:57 +00:00
#include "bgp_mib.h"
2022-09-06 16:04:29 +00:00
/* =============================================================
* Problems
* ------------------------------------------------------------
*
2022-09-30 07:36:09 +00:00
* change of remote ip -> no notification, no update
2022-09-06 16:04:29 +00:00
* same ip, different ports
* distinct VRF (two interfaces with overlapping private addrs)
2022-09-20 12:28:57 +00:00
* posible link-local addresses in LOCAL_IP
2022-09-06 16:04:29 +00:00
*
2023-07-26 12:30:34 +00:00
* context is allocated as copied, is it approach really needed? wouldn't it
* sufficient just use the context in rx-buffer?
*
2022-09-06 16:04:29 +00:00
*/
2023-07-26 12:30:34 +00:00
static void snmp_mib_fill2(struct snmp_proto *p, struct oid *oid, struct snmp_pdu_context *c);
2023-03-14 13:10:08 +00:00
static uint parse_response(struct snmp_proto *p, byte *buf, uint size);
static void do_response(struct snmp_proto *p, byte *buf, uint size);
2023-07-26 12:30:34 +00:00
static uint parse_gets2_pdu(struct snmp_proto *p, byte *buf, uint size, uint *skip);
2023-03-14 13:10:08 +00:00
static uint parse_close_pdu(struct snmp_proto *p, byte *buf, uint size);
2023-07-26 12:30:34 +00:00
static struct agentx_response *prepare_response(struct snmp_proto *p, struct snmp_pdu_context *c);
static void response_err_ind(struct agentx_response *res, uint err, uint ind);
static uint update_packet_size(struct snmp_proto *p, byte *start, byte *end);
2023-07-26 12:30:34 +00:00
static struct oid *search_mib(struct snmp_proto *p, const struct oid *o_start, const struct oid *o_end, struct oid *o_curr, struct snmp_pdu_context *c, enum snmp_search_res *result);
2023-09-04 07:25:51 +00:00
static void notify_pdu(struct snmp_proto *p, struct oid *oid, void *opaque, uint size, int include_uptime);
2022-11-29 15:30:20 +00:00
2022-08-10 15:31:32 +00:00
static const char * const snmp_errs[] = {
#define SNMP_ERR_SHIFT 256
2022-12-17 17:16:19 +00:00
[AGENTX_RES_OPEN_FAILED - SNMP_ERR_SHIFT] = "Open failed",
[AGENTX_RES_NOT_OPEN - SNMP_ERR_SHIFT] = "Not open",
[AGENTX_RES_INDEX_WRONG_TYPE - SNMP_ERR_SHIFT] = "Index wrong type",
2022-08-10 15:31:32 +00:00
[AGENTX_RES_INDEX_ALREADY_ALLOC - SNMP_ERR_SHIFT] = "Index already allocated",
2022-12-17 17:16:19 +00:00
[AGENTX_RES_INDEX_NONE_AVAIL - SNMP_ERR_SHIFT] = "Index none availlable",
[AGENTX_RES_NOT_ALLOCATED - SNMP_ERR_SHIFT] = "Not allocated",
2022-08-10 15:31:32 +00:00
[AGENTX_RES_UNSUPPORTED_CONTEXT - SNMP_ERR_SHIFT] = "Unsupported contex",
2022-12-17 17:16:19 +00:00
[AGENTX_RES_DUPLICATE_REGISTER - SNMP_ERR_SHIFT] = "Duplicate registration",
[AGENTX_RES_UNKNOWN_REGISTER - SNMP_ERR_SHIFT] = "Unknown registration",
[AGENTX_RES_UNKNOWN_AGENT_CAPS - SNMP_ERR_SHIFT] = "Unknown agent caps",
[AGENTX_RES_PARSE_ERROR - SNMP_ERR_SHIFT] = "Parse error",
[AGENTX_RES_REQUEST_DENIED - SNMP_ERR_SHIFT] = "Request denied",
[AGENTX_RES_PROCESSING_ERR - SNMP_ERR_SHIFT] = "Processing error",
2022-08-10 15:31:32 +00:00
};
2022-12-10 17:08:00 +00:00
static const char * const snmp_pkt_type[] = {
[AGENTX_OPEN_PDU] = "Open-PDU",
[AGENTX_CLOSE_PDU] = "Close-PDU",
[AGENTX_REGISTER_PDU] = "Register-PDU",
[AGENTX_UNREGISTER_PDU] = "Unregister-PDU",
[AGENTX_GET_PDU] = "Get-PDU",
[AGENTX_GET_NEXT_PDU] = "GetNext-PDU",
[AGENTX_GET_BULK_PDU] = "GetBulk-PDU",
[AGENTX_TEST_SET_PDU] = "TestSet-PDU",
[AGENTX_COMMIT_SET_PDU] = "CommitSet-PDU",
[AGENTX_UNDO_SET_PDU] = "UndoSet-PDU",
[AGENTX_CLEANUP_SET_PDU] = "CleanupSet-PDU",
[AGENTX_NOTIFY_PDU] = "Notify-PDU",
[AGENTX_PING_PDU] = "Ping-PDU",
[AGENTX_INDEX_ALLOCATE_PDU] = "IndexAllocate-PDU",
[AGENTX_INDEX_DEALLOCATE_PDU] = "IndexDeallocate-PDU",
[AGENTX_ADD_AGENT_CAPS_PDU] = "AddAgentCaps-PDU",
[AGENTX_REMOVE_AGENT_CAPS_PDU] = "RemoveAgentCaps-PDU",
[AGENTX_RESPONSE_PDU] = "Response-PDU",
};
2022-09-06 16:04:29 +00:00
static void
open_pdu(struct snmp_proto *p, struct oid *oid)
2022-08-10 15:31:32 +00:00
{
2022-09-06 16:04:29 +00:00
sock *sk = p->sock;
2023-09-04 07:25:51 +00:00
2023-09-04 11:46:02 +00:00
struct snmp_pdu_context c = SNMP_PDU_CONTEXT(sk);
2023-07-26 12:30:34 +00:00
byte *buf = c.buffer;
2022-09-06 16:04:29 +00:00
2023-07-26 12:30:34 +00:00
// TODO should be configurable; with check on string length
2022-09-06 16:04:29 +00:00
const char *str = "bird";
2023-07-26 12:30:34 +00:00
/* +4 for timeout (1B with 4B alignment) */
2023-09-04 07:25:51 +00:00
if (c.size < AGENTX_HEADER_SIZE + 4 + snmp_oid_size(oid) + snmp_str_size(str))
2022-08-10 15:31:32 +00:00
{
2023-09-04 07:25:51 +00:00
return;
// TODO create and add message info into message queue
2023-07-26 12:30:34 +00:00
snmp_manage_tbuf(p, &c);
buf = c.buffer;
}
2022-09-06 16:04:29 +00:00
2023-07-26 12:30:34 +00:00
struct agentx_header *h = (struct agentx_header *) c.buffer;
ADVANCE(c.buffer, c.size, AGENTX_HEADER_SIZE);
2023-07-26 12:30:34 +00:00
SNMP_BLANK_HEADER(h, AGENTX_OPEN_PDU);
2023-09-04 11:46:02 +00:00
c.byte_ord = h->flags & AGENTX_NETWORK_BYTE_ORDER;
2022-09-06 16:04:29 +00:00
2023-07-26 12:30:34 +00:00
STORE_U32(h->session_id, 1);
STORE_U32(h->transaction_id, 1);
STORE_U32(h->packet_id, 1);
2022-09-06 16:04:29 +00:00
c.size -= (4 + snmp_oid_size(oid) + snmp_str_size(str));
2023-07-26 12:30:34 +00:00
c.buffer = snmp_put_fbyte(c.buffer, p->timeout);
c.buffer = snmp_put_oid(c.buffer, oid);
c.buffer = snmp_put_str(c.buffer, str);
2022-09-06 16:04:29 +00:00
uint s = update_packet_size(p, buf, c.buffer);
int ret = sk_send(sk, s);
if (ret > 0)
snmp_log("sk_send OK!");
else if (ret == 0)
snmp_log("sk_send sleep");
else
snmp_log("sk_send error");
2023-07-26 12:30:34 +00:00
}
2023-09-04 07:25:51 +00:00
static void
notify_pdu(struct snmp_proto *p, struct oid *oid, void *opaque, uint size, int include_uptime)
2023-07-26 12:30:34 +00:00
{
sock *sk = p->sock;
2023-09-04 11:46:02 +00:00
struct snmp_pdu_context c = SNMP_PDU_CONTEXT(sk);
2023-07-26 12:30:34 +00:00
2023-09-04 07:25:51 +00:00
#define UPTIME_SIZE \
(6 * sizeof(u32)) /* sizeof( { u32 vb_type, u32 oid_hdr, u32 ids[4] } )*/
#define TRAP0_HEADER_SIZE \
(7 * sizeof(u32)) /* sizeof( { u32 vb_type, u32 oid_hdr, u32 ids[6] } ) */
2023-07-26 12:30:34 +00:00
2023-09-04 07:25:51 +00:00
uint sz = AGENTX_HEADER_SIZE + TRAP0_HEADER_SIZE + snmp_oid_size(oid) \
2023-09-04 11:46:02 +00:00
+ size;
2023-09-04 07:25:51 +00:00
if (include_uptime)
sz += UPTIME_SIZE;
2023-09-04 11:46:02 +00:00
2023-09-04 07:25:51 +00:00
if (c.size < sz)
snmp_manage_tbuf(p, &c);
2023-09-04 11:46:02 +00:00
struct agentx_header *h = (struct agentx_header *) c.buffer;
ADVANCE(c.buffer, c.size, AGENTX_HEADER_SIZE);
SNMP_BLANK_HEADER(h, AGENTX_NOTIFY_PDU);
SNMP_SESSION(h, p);
c.byte_ord = h->flags & AGENTX_NETWORK_BYTE_ORDER;
2023-09-04 07:25:51 +00:00
if (include_uptime)
2023-07-26 12:30:34 +00:00
{
2023-09-04 07:25:51 +00:00
/* sysUpTime.0 oid */
struct oid uptime = {
.n_subid = 4,
.prefix = 2,
.include = 0,
.pad = 0,
};
u32 uptime_ids[] = { 1, 1, 3, 0 };
struct agentx_varbind *vb = snmp_create_varbind(c.buffer, &uptime);
for (uint i = 0; i < uptime.n_subid; i++)
STORE_U32(vb->name.ids[i], uptime_ids[i]);
ADVANCE(c.buffer, c.size, snmp_varbind_header_size(vb));
snmp_varbind_ticks(vb, c.size, current_time() TO_S);
ADVANCE(c.buffer, c.size, agentx_type_size(AGENTX_TIME_TICKS));
2023-07-26 12:30:34 +00:00
}
2023-09-04 07:25:51 +00:00
/* snmpTrapOID.0 oid */
struct oid trap0 = {
.n_subid = 6,
.prefix = 6,
.include = 0,
.pad = 0,
};
u32 trap0_ids[] = { 3, 1, 1, 4, 1, 0 };
struct agentx_varbind *trap_vb = snmp_create_varbind(c.buffer, &trap0);
for (uint i = 0; i < trap0.n_subid; i++)
STORE_U32(trap_vb->name.ids[i], trap0_ids[i]);
trap_vb->type = AGENTX_OBJECT_ID;
snmp_put_oid(SNMP_VB_DATA(trap_vb), oid);
ADVANCE(c.buffer, c.size, snmp_varbind_size(trap_vb, c.byte_ord));
2023-07-26 12:30:34 +00:00
2023-09-04 07:25:51 +00:00
// TODO fix the endianess
memcpy(c.buffer, opaque, size);
ADVANCE(c.buffer, c.size, (size + snmp_varbind_hdr_size_from_oid(oid)));
uint s = update_packet_size(p, sk->tbuf, c.buffer);
int ret = sk_send(sk, s);
if (ret > 0)
snmp_log("sk_send OK!");
else if (ret == 0)
2023-07-26 12:30:34 +00:00
snmp_log("sk_send sleep");
2022-08-10 15:31:32 +00:00
else
2023-09-04 07:25:51 +00:00
snmp_log("sk_send error");
2022-08-10 15:31:32 +00:00
2023-09-04 07:25:51 +00:00
#undef TRAP0_HEADER_SIZE
#undef UPTIME_SIZE
2023-07-26 12:30:34 +00:00
}
2023-09-04 07:25:51 +00:00
/* index allocate / deallocate pdu * /
static void
de_allocate_pdu(struct snmp_proto *p, struct oid *oid, u8 type)
2023-07-26 12:30:34 +00:00
{
2023-09-04 07:25:51 +00:00
sock *sk = p->sock;
byte *buf, *pkt;
buf = pkt = sk->tbuf;
uint size = sk->tbsize;
if (size > AGENTX_HEADER_SIZE + )
{
snmp_log("de_allocate_pdu()");
struct agentx_header *h;
SNMP_CREATE(pkt, struct agentx_header, h);
SNMP_BLANK_HEADER(h, type);
SNMP_SESSION(h,p);
struct agentx_varbind *vb = (struct agentx_varbind *) pkt;
STORE_16(vb->type, AGENTX_OBJECT_ID);
STORE(vb->oid,
}
else
snmp_log("de_allocate_pdu(): insufficient size");
2023-07-26 12:30:34 +00:00
}
2023-09-04 07:25:51 +00:00
*/
2023-07-26 12:30:34 +00:00
2023-09-04 07:25:51 +00:00
/* Register-PDU / Unregister-PDU */
static void
un_register_pdu(struct snmp_proto *p, struct oid *oid, uint index, uint len, u8 type, u8 is_instance)
2022-09-06 16:04:29 +00:00
{
sock *sk = p->sock;
2023-09-04 11:46:02 +00:00
struct snmp_pdu_context c = SNMP_PDU_CONTEXT(sk);
2023-07-26 12:30:34 +00:00
byte *buf = c.buffer;
2022-09-06 16:04:29 +00:00
2023-09-04 07:25:51 +00:00
/* conditional +4 for upper-bound (optinal field) */
2023-07-26 12:30:34 +00:00
if (c.size < AGENTX_HEADER_SIZE + snmp_oid_size(oid) + ((len > 1) ? 4 : 0))
2022-08-10 15:31:32 +00:00
{
2023-07-26 12:30:34 +00:00
snmp_log("un_register_pdu() insufficient size");
snmp_manage_tbuf(p, &c);
buf = c.buffer;
}
2022-09-06 16:04:29 +00:00
2023-07-26 12:30:34 +00:00
snmp_log("un_register_pdu()");
struct agentx_un_register_pdu *ur = (struct agentx_un_register_pdu *)c.buffer;
ADVANCE(c.buffer, c.size, sizeof(struct agentx_un_register_pdu));
struct agentx_header *h = &ur->h;
2023-09-04 07:25:51 +00:00
// FIXME correctly set INSTANCE REGISTRATION bit
SNMP_HEADER(h, type, is_instance ? AGENTX_FLAG_INSTANCE_REGISTRATION : 0);
2023-09-04 07:25:51 +00:00
/* use new transactionID, reset packetID */
p->packet_id++;
2023-07-26 12:30:34 +00:00
SNMP_SESSION(h, p);
2023-09-04 11:46:02 +00:00
c.byte_ord = h->flags & AGENTX_NETWORK_BYTE_ORDER;
2023-07-26 12:30:34 +00:00
/* do not override timeout */
STORE_U32(ur->timeout, 15);
2023-09-04 07:25:51 +00:00
/* default priority */
STORE_U32(ur->priority, AGENTX_PRIORITY);
2023-07-26 12:30:34 +00:00
STORE_U32(ur->range_subid, (len > 1) ? index : 0);
snmp_put_oid(c.buffer, oid);
ADVANCE(c.buffer, c.size, snmp_oid_size(oid));
2023-09-04 07:25:51 +00:00
/* place upper-bound if needed */
2023-07-26 12:30:34 +00:00
if (len > 1)
{
STORE_PTR(c.buffer, len);
ADVANCE(c.buffer, c.size, 4);
}
2022-09-06 16:04:29 +00:00
uint s = update_packet_size(p, buf, c.buffer);
2022-08-10 15:31:32 +00:00
2023-07-26 12:30:34 +00:00
snmp_log("sending (un)register %s", snmp_pkt_type[type]);
int ret = sk_send(sk, s);
if (ret > 0)
snmp_log("sk_send OK!");
else if (ret == 0)
snmp_log("sk_send sleep");
else
snmp_log("sk_send error");
2022-09-06 16:04:29 +00:00
}
2023-09-04 07:25:51 +00:00
/* register pdu */
2022-12-10 17:08:00 +00:00
void
snmp_register(struct snmp_proto *p, struct oid *oid, uint index, uint len, u8 is_instance)
2022-09-06 16:04:29 +00:00
{
un_register_pdu(p, oid, index, len, AGENTX_REGISTER_PDU, is_instance);
2022-09-06 16:04:29 +00:00
}
2023-09-04 07:25:51 +00:00
/* unregister pdu */
2022-12-10 17:08:00 +00:00
void UNUSED
2022-09-06 16:04:29 +00:00
snmp_unregister(struct snmp_proto *p, struct oid *oid, uint index, uint len)
{
un_register_pdu(p, oid, index, len, AGENTX_UNREGISTER_PDU, 0);
2022-08-10 15:31:32 +00:00
}
static void
close_pdu(struct snmp_proto *p, u8 reason)
{
2022-08-10 15:31:32 +00:00
sock *sk = p->sock;
2023-09-04 11:46:02 +00:00
struct snmp_pdu_context c = SNMP_PDU_CONTEXT(sk);
2023-07-26 12:30:34 +00:00
byte *buf = c.buffer;
snmp_log("close_pdu() size: %u %c %u", c.size, (c.size > AGENTX_HEADER_SIZE + 4)
2022-08-10 15:31:32 +00:00
? '>':'<', AGENTX_HEADER_SIZE);
2022-08-10 15:31:32 +00:00
/* +4B for reason */
2023-07-26 12:30:34 +00:00
if (c.size < AGENTX_HEADER_SIZE + 4)
2022-08-10 15:31:32 +00:00
{
2023-07-26 12:30:34 +00:00
snmp_manage_tbuf(p, &c);
buf = c.buffer;
}
struct agentx_header *h = (struct agentx_header *) c.buffer;
ADVANCE(c.buffer, c.size, AGENTX_HEADER_SIZE);
SNMP_BLANK_HEADER(h, AGENTX_CLOSE_PDU);
p->packet_id++;
2023-07-26 12:30:34 +00:00
SNMP_SESSION(h, p);
2023-09-04 11:46:02 +00:00
c.byte_ord = h->flags & AGENTX_NETWORK_BYTE_ORDER;
2023-07-26 12:30:34 +00:00
snmp_put_fbyte(c.buffer, reason);
ADVANCE(c.buffer, c.size, 4);
uint s = update_packet_size(p, buf, c.buffer);
2023-07-26 12:30:34 +00:00
snmp_log("preparing to sk_send() (close)");
int ret = sk_send(sk, s);
if (ret > 0)
snmp_log("sk_send OK!");
else if (ret == 0)
snmp_log("sk_send sleep");
else
snmp_log("sk_send error");
2023-07-26 12:30:34 +00:00
}
#if 0
static void UNUSED
parse_testset_pdu(struct snmp_proto *p)
{
sock *sk = p->sock;
sk_send(sk, 0);
}
static void UNUSED
parse_commitset_pdu(struct snmp_proto *p)
{
sock *sk = p->sock;
sk_send(sk, 0);
}
static void UNUSED
parse_undoset_pdu(struct snmp_proto *p)
{
sock *sk = p->sock;
sk_send(sk, 0);
}
static void UNUSED
parse_cleanupset_pdu(struct snmp_proto *p)
{
sock *sk = p->sock;
sk_send(sk, 0);
}
static void UNUSED
addagentcaps_pdu(struct snmp_proto *p, struct oid *cap, char *descr,
uint descr_len, struct agentx_context *c)
{
ASSUME(descr != NULL && descr_len > 0);
sock *sk = p->sock;
//byte *buf = sk->tbuf;
//uint size = sk->tbsize;
// TODO rename to pkt and add pkt_start
byte *buf = sk->tpos;
uint size = sk->tbuf + sk->tbsize - sk->tpos;
if (size < AGENTX_HEADER_SIZE + snmp_context_size(c) + snmp_oid_size(cap) + snmp_str_size_from_len(descr_len))
{
/* TODO need more mem */
return;
}
struct agentx_header *h;
SNMP_CREATE(buf, struct agentx_header, h);
SNMP_BLANK_HEADER(h, AGENTX_ADD_AGENT_CAPS_PDU);
SNMP_SESSION(h, p);
ADVANCE(buf, size, AGENTX_HEADER_SIZE);
2023-07-26 12:30:34 +00:00
uint in_pkt;
if (c && c->length)
{
SNMP_HAS_CONTEXT(h);
in_pkt = snmp_put_nstr(buf, c->context, c->length) - buf;
ADVANCE(buf, size, in_pkt);
}
2023-09-04 07:25:51 +00:00
// memcpy(buf, cap, snmp_oid_size(cap));
2023-07-26 12:30:34 +00:00
ADVANCE(buf, size, snmp_oid_size(cap));
2023-07-26 12:30:34 +00:00
in_pkt = snmp_put_nstr(buf, descr, descr_len) - buf;
ADVANCE(buf, size, in_pkt);
// make a note in the snmp_proto structure
//int ret = sk_send(sk, buf - sk->tbuf);
int ret = sk_send(sk, buf - sk->tpos);
if (ret == 0)
snmp_log("sk_send sleep");
else if (ret < 0)
snmp_log("sk_send err");
else
log(L_INFO, "sk_send ok !!");
}
static void UNUSED
removeagentcaps_pdu(struct snmp_proto *p, struct oid *cap, struct agentx_context *c)
{
sock *sk = p->sock;
//byte *buf = sk->tbuf;
//uint size = sk->tbsize;
// TODO rename to pkt and add pkt_start
byte *buf = sk->tpos;
uint size = sk->tbuf + sk->tbsize - sk->tpos;
if (size < AGENTX_HEADER_SIZE + snmp_context_size(c) + snmp_oid_size(cap))
{
/* TODO need more mem */
return;
}
2023-07-26 12:30:34 +00:00
struct agentx_header *h;
SNMP_CREATE(buf, struct agentx_header, h);
SNMP_SESSION(h, p);
ADVANCE(buf, size, AGENTX_HEADER_SIZE);
2022-08-10 15:31:32 +00:00
2023-07-26 12:34:01 +00:00
uint in_pkt;
if (c && c->length)
{
SNMP_HAS_CONTEXT(h);
in_pkt = snmp_put_nstr(buf, c->context, c->length) - buf;
ADVANCE(buf, size, in_pkt);
2022-08-10 15:31:32 +00:00
}
2023-07-26 12:34:01 +00:00
memcpy(buf, cap, snmp_oid_size(cap));
ADVANCE(buf, size, snmp_oid_size(cap));
// update state in snmp_proto structure
//int ret = sk_send(sk, buf - sk->tbuf);
int ret = sk_send(sk, buf - sk->tpos);
if (ret == 0)
snmp_log("sk_send sleep");
else if (ret < 0)
snmp_log("sk_send err");
else
log(L_INFO, "sk_send ok !!");
2022-08-10 15:31:32 +00:00
}
2023-07-26 12:34:01 +00:00
#endif
2022-08-10 15:31:32 +00:00
2022-09-06 16:04:29 +00:00
static inline void
refresh_ids(struct snmp_proto *p, struct agentx_header *h)
{
int byte_ord = h->flags & AGENTX_NETWORK_BYTE_ORDER;
p->transaction_id = LOAD_U32(h->transaction_id, byte_ord);
p->packet_id = LOAD_U32(h->packet_id, byte_ord);
2022-09-06 16:04:29 +00:00
}
2023-03-14 13:10:08 +00:00
/**
* parse_pkt - parse recieved response packet
* @p:
* @pkt: packet buffer
* @size: number of packet bytes in buffer
* retval number of byte parsed
*
2023-03-24 14:02:23 +00:00
* function parse_ptk() parses response-pdu and calls do_response().
* returns number of bytes parsed by function excluding size of header.
2023-03-14 13:10:08 +00:00
*/
static uint
2023-07-26 12:30:34 +00:00
parse_pkt(struct snmp_proto *p, byte *pkt, uint size, uint *skip)
2022-08-10 15:31:32 +00:00
{
2023-03-14 13:10:08 +00:00
snmp_log("parse_ptk() pkt start: %p", pkt);
2023-07-26 12:30:34 +00:00
if (size < AGENTX_HEADER_SIZE)
2022-08-10 15:31:32 +00:00
return 0;
2023-03-14 13:10:08 +00:00
uint parsed_len = 0;
struct agentx_header *h = (void *) pkt;
2022-12-10 17:08:00 +00:00
2023-09-04 07:25:51 +00:00
snmp_log("parse_pkt got type %s (%d)", snmp_pkt_type[h->type], h->type);
snmp_log("parse_pkt rx size %u", size);
2023-07-26 12:30:34 +00:00
snmp_dump_packet((void *)h, MIN(h->payload, 256));
2022-08-10 15:31:32 +00:00
switch (h->type)
{
case AGENTX_RESPONSE_PDU:
2023-03-14 13:10:08 +00:00
snmp_log("parse_pkt returning parse_response");
parsed_len = parse_response(p, pkt, size);
break;
2022-09-06 16:04:29 +00:00
2023-09-04 07:25:51 +00:00
/*
case AGENTX_GET_PDU:
refresh_ids(p, h);
return parse_get_pdu(p, pkt, size);
*/
2022-09-20 12:28:57 +00:00
case AGENTX_GET_PDU:
case AGENTX_GET_NEXT_PDU:
2022-09-30 07:36:09 +00:00
case AGENTX_GET_BULK_PDU:
2022-09-20 12:28:57 +00:00
refresh_ids(p, h);
2023-09-04 07:25:51 +00:00
//parsed_len = parse_gets_pdu(p, &c);
//parsed_len = parse_gets_pdu(p, pkt, size, skip);
2023-07-26 12:30:34 +00:00
parsed_len = parse_gets2_pdu(p, pkt, size, skip);
2022-09-30 07:36:09 +00:00
break;
2022-08-10 15:31:32 +00:00
2023-03-14 13:10:08 +00:00
/* during testing the connection should stay opened (we die if we screw up
* and get CLOSE_PDU in response)
case AGENTX_CLOSE_PDU:
refresh_ids(p, h);
parsed_len = parse_close_pdu(p, pkt, size);
break;
*/
2022-08-10 15:31:32 +00:00
/* should not happen */
default:
2023-07-26 12:30:34 +00:00
snmp_log("unknown packet type %u", h->type);
return 0;
//die("unknown packet type %u", h->type);
2022-08-10 15:31:32 +00:00
}
2022-09-30 07:36:09 +00:00
2023-09-04 07:25:51 +00:00
/* We will process the same header again later * /
if (*skip || parsed_len < size)
{
/ * We split our answer to multiple packet, we should differentiate them * /
h->packet_id++;
}
*/
2023-03-14 13:10:08 +00:00
snmp_log("parse_pkt returning parsed length");
2023-09-04 07:25:51 +00:00
//snmp_dump_packet(p->sock->tbuf, 64);
2023-03-14 13:10:08 +00:00
return parsed_len;
2022-08-10 15:31:32 +00:00
}
2023-03-14 13:10:08 +00:00
static uint
parse_response(struct snmp_proto *p, byte *res, uint size)
2022-08-10 15:31:32 +00:00
{
2022-12-17 17:16:19 +00:00
snmp_log("parse_response() g%u h%u", size, sizeof(struct agentx_header));
2023-09-04 07:25:51 +00:00
//snmp_dump_packet(res, size);
2022-08-10 15:31:32 +00:00
if (size < sizeof(struct agentx_response))
return 0;
2023-03-14 13:10:08 +00:00
struct agentx_response *r = (void *) res;
2022-08-10 15:31:32 +00:00
struct agentx_header *h = &r->h;
2023-03-14 13:10:08 +00:00
int byte_ord = h->flags & AGENTX_NETWORK_BYTE_ORDER;
2023-07-26 12:30:34 +00:00
uint pkt_size = LOAD_U32(h->payload, byte_ord);
2023-03-14 13:10:08 +00:00
snmp_log("p_res pkt_size %u", pkt_size);
2023-07-26 12:30:34 +00:00
if (size < pkt_size + AGENTX_HEADER_SIZE) {
2023-03-14 13:10:08 +00:00
snmp_log("parse_response early return");
return 0;
}
2023-07-26 12:30:34 +00:00
snmp_log(" endianity: %s, session %u, transaction: %u",
(h->flags & AGENTX_NETWORK_BYTE_ORDER) ? "big end": "little end",
h->session_id, h->transaction_id);
2022-12-17 17:16:19 +00:00
snmp_log(" sid: %3u\ttid: %3u\tpid: %3u", p->session_id, p->transaction_id,
2023-03-24 14:02:23 +00:00
p->packet_id);
2022-09-06 16:04:29 +00:00
2022-12-10 17:08:00 +00:00
snmp_log(" pkt size %u", h->payload);
2022-09-06 16:04:29 +00:00
2023-07-26 12:30:34 +00:00
if (r->error == AGENTX_RES_NO_ERROR)
2023-03-14 13:10:08 +00:00
do_response(p, res, size);
2022-09-06 16:04:29 +00:00
else
2023-09-04 07:25:51 +00:00
/* erronous packet should be dropped quietly */
2023-07-26 12:30:34 +00:00
snmp_log("an error occured '%s'", snmp_errs[get_u16(&r->error) - SNMP_ERR_SHIFT]);
2022-09-06 16:04:29 +00:00
2023-07-26 12:30:34 +00:00
return pkt_size + AGENTX_HEADER_SIZE;
2022-09-06 16:04:29 +00:00
}
2022-12-10 17:08:00 +00:00
static inline int
snmp_registered_all(struct snmp_proto *p)
{
2022-12-17 17:16:19 +00:00
snmp_log("snmp_registered_all() %u", list_length(&p->register_queue));
2022-12-10 17:08:00 +00:00
return p->register_to_ack == 0;
}
static void
2023-07-26 12:34:01 +00:00
snmp_register_mibs(struct snmp_proto *p)
{
2022-12-10 17:08:00 +00:00
snmp_log("snmp_register_mibs()");
snmp_bgp_register(p);
snmp_log("registering all done");
}
2023-03-14 13:10:08 +00:00
static void
2022-09-20 12:28:57 +00:00
do_response(struct snmp_proto *p, byte *buf, uint size UNUSED)
2022-09-06 16:04:29 +00:00
{
2022-12-10 17:08:00 +00:00
snmp_log("do_response()");
2022-09-06 16:04:29 +00:00
struct agentx_response *r = (void *) buf;
struct agentx_header *h = &r->h;
2023-07-26 12:30:34 +00:00
int byte_ord = h->flags & AGENTX_NETWORK_BYTE_ORDER;
2022-12-17 17:16:19 +00:00
2023-08-08 19:51:38 +00:00
/* TODO make it asynchronous for better speed */
2022-09-06 16:04:29 +00:00
switch (p->state)
2022-08-10 15:31:32 +00:00
{
2022-09-06 16:04:29 +00:00
case SNMP_INIT:
2023-09-04 07:25:51 +00:00
case SNMP_LOCKED:
/* silent drop of recieved packet */
break;
case SNMP_OPEN:
/* copy session info from recieved packet */
p->session_id = LOAD_U32(h->session_id, byte_ord);
2023-07-26 12:30:34 +00:00
refresh_ids(p, h);
2022-09-06 16:04:29 +00:00
2023-09-04 07:25:51 +00:00
/* the state needs to be changed before sending registering PDUs to
2022-12-17 17:16:19 +00:00
* use correct do_response action on them
*/
2022-12-10 17:08:00 +00:00
snmp_log("changing state to REGISTER");
2022-12-17 17:16:19 +00:00
p->state = SNMP_REGISTER;
snmp_register_mibs(p);
snmp_log("do_response state SNMP_INIT register list %u", list_length(&p->register_queue));
2022-12-10 17:08:00 +00:00
2022-08-10 15:31:32 +00:00
break;
2022-09-06 16:04:29 +00:00
2022-12-17 17:16:19 +00:00
case SNMP_REGISTER:
snmp_log("do_response state SNMP_REGISTER register list %u", list_length(&p->register_queue));
snmp_register_ack(p ,h);
2022-12-10 17:08:00 +00:00
if (snmp_registered_all(p)) {
snmp_log("changing proto_snmp state to CONNECTED");
p->state = SNMP_CONN;
2023-09-04 07:25:51 +00:00
proto_notify_state(&p->p, PS_UP);
2022-12-10 17:08:00 +00:00
}
2022-08-10 15:31:32 +00:00
break;
2022-09-06 16:04:29 +00:00
case SNMP_CONN:
2022-12-10 17:08:00 +00:00
// proto_notify_state(&p->p, PS_UP);
2022-09-06 16:04:29 +00:00
break;
2022-12-17 17:16:19 +00:00
case SNMP_STOP:
2023-09-04 11:46:02 +00:00
snmp_down(p);
2022-12-17 17:16:19 +00:00
break;
2022-09-06 16:04:29 +00:00
default:
die("unkonwn SNMP state");
2022-08-10 15:31:32 +00:00
}
2022-09-06 16:04:29 +00:00
}
2023-07-26 12:30:34 +00:00
u8
snmp_get_mib_class(const struct oid *oid)
2022-11-05 15:29:00 +00:00
{
2023-09-04 07:25:51 +00:00
// TODO check code paths for oid->n_subid < 3
if (oid->prefix != 2 && oid->ids[0] != SNMP_MIB_2)
2022-11-05 15:29:00 +00:00
return SNMP_CLASS_INVALID;
switch (oid->ids[1])
{
case SNMP_BGP4_MIB:
return SNMP_CLASS_BGP;
default:
return SNMP_CLASS_END;
2022-12-10 17:08:00 +00:00
}
2022-11-05 15:29:00 +00:00
}
2022-09-20 12:28:57 +00:00
2023-07-26 12:30:34 +00:00
static void
snmp_get_next2(struct snmp_proto *p, struct oid *o_start, struct oid *o_end,
struct snmp_pdu_context *c)
{
snmp_log("get_next2()");
enum snmp_search_res r;
snmp_log("next2() o_end %p", o_end);
struct oid *o_copy = search_mib(p, o_start, o_end, NULL, c, &r);
snmp_log("next2()2 o_end %p", o_end);
2023-08-08 18:47:30 +00:00
struct agentx_varbind *vb = NULL;
switch (r)
{
case SNMP_SEARCH_NO_OBJECT:
case SNMP_SEARCH_NO_INSTANCE:
case SNMP_SEARCH_END_OF_VIEW:;
uint sz = snmp_varbind_hdr_size_from_oid(o_start);
if (c->size < sz)
{
2023-09-04 07:25:51 +00:00
/* TODO create NULL varbind */
c->error = AGENTX_RES_GEN_ERROR;
return;
2023-08-08 18:47:30 +00:00
}
vb = snmp_create_varbind(c->buffer, o_start);
vb->type = AGENTX_END_OF_MIB_VIEW;
ADVANCE(c->buffer, c->size, snmp_varbind_header_size(vb));
return;
case SNMP_SEARCH_OK:
2023-09-04 07:25:51 +00:00
default:
2023-08-08 18:47:30 +00:00
break;
}
2023-07-26 12:30:34 +00:00
if (o_copy)
2023-08-08 18:47:30 +00:00
{
2023-09-04 07:25:51 +00:00
/* basicaly snmp_create_varbind(c->buffer, o_copy), but without any copying */
2023-08-08 18:47:30 +00:00
vb = (void *) c->buffer;
2023-07-26 12:30:34 +00:00
snmp_mib_fill2(p, o_copy, c);
2023-08-08 18:47:30 +00:00
2023-09-04 07:25:51 +00:00
/* override the error for GetNext-PDU object not find */
2023-08-08 18:47:30 +00:00
switch (vb->type)
{
case AGENTX_NO_SUCH_OBJECT:
case AGENTX_NO_SUCH_INSTANCE:
case AGENTX_END_OF_MIB_VIEW:
vb->type = AGENTX_END_OF_MIB_VIEW;
break;
default: /* intentionally left blank */
break;
}
return;
}
if (c->size < snmp_varbind_hdr_size_from_oid(o_start))
2023-07-26 12:30:34 +00:00
{
2023-08-08 18:47:30 +00:00
// TODO FIXME this is a bit tricky as we need to renew all TX buffer pointers
snmp_manage_tbuf(p, c);
2023-07-26 12:30:34 +00:00
}
2023-08-08 18:47:30 +00:00
vb = snmp_create_varbind(c->buffer, o_start);
vb->type = AGENTX_END_OF_MIB_VIEW;
ADVANCE(c->buffer, c->size, snmp_varbind_header_size(vb));
2023-07-26 12:30:34 +00:00
}
static void
snmp_get_bulk2(struct snmp_proto *p, struct oid *o_start, struct oid *o_end,
struct agentx_bulk_state *state, struct snmp_pdu_context *c)
{
if (state->index <= state->getbulk.non_repeaters)
2023-09-04 07:25:51 +00:00
{
2023-07-26 12:30:34 +00:00
return snmp_get_next2(p, o_start, o_end, c);
2023-09-04 07:25:51 +00:00
/*
* Here we don't need to do any overriding, not even in case no object was
* found, as the GetNext-PDU override is same as GetBulk-PDU override
* (to AGENTX_RES_END_OF_MIB_VIEW)
*/
}
struct oid *o_curr = NULL;
struct oid *o_predecessor = NULL;
enum snmp_search_res r;
uint i = 0;
do
2023-07-26 12:30:34 +00:00
{
2023-09-04 07:25:51 +00:00
o_predecessor = o_curr;
o_curr = search_mib(p, o_start, o_end, o_curr, c, &r);
i++;
} while (o_curr && i <= state->repetition);
/* Object Identifier fall-backs */
if (!o_curr)
o_curr = o_predecessor;
2023-07-26 12:30:34 +00:00
2023-09-04 07:25:51 +00:00
if (!o_curr)
o_curr = o_start;
2023-07-26 12:30:34 +00:00
2023-09-04 07:25:51 +00:00
uint sz = snmp_varbind_hdr_size_from_oid(o_curr);
2022-11-19 22:00:02 +00:00
2023-09-04 07:25:51 +00:00
if (c->size < sz)
{
c->error = AGENTX_RES_GEN_ERROR;
return;
}
2023-07-26 12:30:34 +00:00
2023-09-04 07:25:51 +00:00
/* we need the varbind handle to be able to override it's type */
struct agentx_varbind *vb = (void *) c->buffer;
vb->type = AGENTX_END_OF_MIB_VIEW;
2023-07-26 12:30:34 +00:00
2023-09-04 07:25:51 +00:00
if (r == SNMP_SEARCH_OK)
/* the varbind will be recreated inside the snmp_mib_fill2() */
snmp_mib_fill2(p, o_curr, c);
else
ADVANCE(c->buffer, c->size, snmp_varbind_header_size(vb));
2023-07-26 12:30:34 +00:00
2023-09-04 07:25:51 +00:00
/* override the error for GetBulk-PDU object not found */
switch (vb->type)
{
case AGENTX_NO_SUCH_OBJECT:
case AGENTX_NO_SUCH_INSTANCE:
case AGENTX_END_OF_MIB_VIEW:
vb->type = AGENTX_END_OF_MIB_VIEW;
break;
2023-07-26 12:30:34 +00:00
2023-09-04 07:25:51 +00:00
default: /* intentionally left blank */
break;
2022-11-19 22:00:02 +00:00
}
}
2023-07-26 12:30:34 +00:00
static uint UNUSED
parse_close_pdu(struct snmp_proto UNUSED *p, byte UNUSED *req, uint UNUSED size)
2023-03-14 13:10:08 +00:00
{
/*
snmp_log("parse_close_pdu()");
// byte *pkt = req;
// sock *sk = p->sock;
if (size < sizeof(struct agentx_header))
{
snmp_log("p_close early return");
return 0;
}
// struct agentx_header *h = (void *) req;
ADVANCE(req, size, AGENTX_HEADER_SIZE);
//snmp_log("after header %p", req);
p->state = SNMP_ERR;
2023-09-04 07:25:51 +00:00
proto_notify(PS_DOWN, &p->p);
2023-03-14 13:10:08 +00:00
*/
return 0;
}
static inline uint
update_packet_size(struct snmp_proto *p, byte *start, byte *end)
2023-07-26 12:30:34 +00:00
{
2023-09-04 07:25:51 +00:00
/* work even for partial messages */
struct agentx_header *h = (void *) p->sock->tpos;
2023-07-26 12:30:34 +00:00
size_t s = snmp_pkt_len(start, end);
STORE_U32(h->payload, s);
return AGENTX_HEADER_SIZE + s;
2023-09-04 07:25:51 +00:00
#if 0
if (EMPTY_LIST(p->additional_buffers))
{
return AGENTX_HEADER_SIZE + STORE_U32(h->payload, snmp_pkt_len(start, end));
}
uint size = p->to_send; /* to_send contain also the AGENTX_HEADER_SIZE */
struct additional_buffer *b = TAIL(p->additional_buffers);
snmp_log("update_packet_size additional buf 0x%p pos 0x%p new pos 0x%p", b->buf, b->pos, end);
b->pos = end;
snmp_log("update_packet_size to_send %u", p->to_send);
/* TODO add packet size limiting
* we couldn't overflow the size because we limit the maximum packet size
*/
WALK_LIST(b, p->additional_buffers)
{
size += b->pos - b->buf;
snmp_log("update_packet_size add %u => %u", b->pos - b->buf, size);
}
STORE_U32(h->payload, size - AGENTX_HEADER_SIZE);
return size;
// if (p->additional_bufferson
// STORE_U32(h->payload, p->to_send + (end - start));
// else {}
//// STORE_U32(h->payload, snmp_pkt_len(start, end));
#endif
2023-07-26 12:30:34 +00:00
}
static inline void
response_err_ind(struct agentx_response *res, uint err, uint ind)
{
STORE_U32(res->error, err);
if (err != AGENTX_RES_NO_ERROR && err != AGENTX_RES_PARSE_ERROR)
STORE_U32(res->index, ind);
else
STORE_U32(res->index, 0);
}
static uint
parse_gets2_pdu(struct snmp_proto *p, byte * const pkt_start, uint size, uint *skip)
{
snmp_log("parse_gets2_pdu()");
struct oid *o_start = NULL, *o_end = NULL;
byte *pkt = pkt_start;
struct agentx_header *h = (void *) pkt;
ADVANCE(pkt, size, AGENTX_HEADER_SIZE);
uint pkt_size = LOAD_U32(h->payload, h->flags & AGENTX_NETWORK_BYTE_ORDER);
2023-07-26 12:30:34 +00:00
sock *sk = p->sock;
2023-09-04 11:46:02 +00:00
struct snmp_pdu_context c = SNMP_PDU_CONTEXT(sk);
// TODO better handling of endianness
c.byte_ord = 0; /* use little-endian */
2023-07-26 12:30:34 +00:00
2023-09-04 07:25:51 +00:00
uint clen; /* count of characters in context (without last '\0') */
char *context; /* newly allocated string of character */
2023-08-08 19:51:38 +00:00
2023-07-26 12:30:34 +00:00
/* alters pkt; assign context, clen */
SNMP_LOAD_CONTEXT(p, h, pkt, context, clen);
2023-08-08 19:51:38 +00:00
2023-07-26 12:30:34 +00:00
/*
* We need more data; for valid response we need to know full
* header picture, including the context octet string
*/
if (size < clen)
{
snmp_log("size %u < %u clen, returning 0", size, clen);
goto wait;
}
/*
* It is a malformed packet if the context octet string should be longer than
* whole packet.
*/
if (pkt_size < clen)
{
2023-09-04 07:25:51 +00:00
/* for malformed packets consume full pkt_size [or size] */
2023-07-26 12:30:34 +00:00
c.error = AGENTX_RES_PARSE_ERROR;
goto send;
}
/* The RFC does not consider the context octet string as a part of a header */
ADVANCE(pkt, pkt_size, clen);
size -= clen;
/* FIXME add support for c.context hashing
c.context = ...
*/
2023-09-04 07:25:51 +00:00
struct agentx_bulk_state bulk_state = { };
if (h->type == AGENTX_GET_BULK_PDU)
{
if (size < sizeof(struct agentx_getbulk))
goto wait;
2023-07-26 12:30:34 +00:00
2023-09-04 07:25:51 +00:00
if (pkt_size < sizeof(struct agentx_getbulk))
{
c.error = AGENTX_RES_PARSE_ERROR;
goto send;
}
struct agentx_getbulk *bulk_info = (void *) pkt;
ADVANCE(pkt, pkt_size, sizeof(struct agentx_getbulk));
bulk_state = (struct agentx_bulk_state) {
.getbulk = {
.non_repeaters = LOAD_U32(bulk_info->non_repeaters, c.byte_ord),
.max_repetitions = LOAD_U32(bulk_info->max_repetitions, c.byte_ord),
},
.index = 1,
.repetition = 1,
};
}
if (!p->partial_response && c.size < sizeof(struct agentx_response))
2023-07-26 12:30:34 +00:00
{
snmp_manage_tbuf(p, &c);
}
struct agentx_response *response_header = prepare_response(p, &c);
uint ind = 1;
while (c.error == AGENTX_RES_NO_ERROR && size > 0 && pkt_size > 0)
{
2023-09-04 07:25:51 +00:00
snmp_log("iter %u size %u remaining %u/%u", ind, c.buffer - sk->tpos, size, pkt_size);
#if 0
if (EMPTY_LIST(p->additional_buffers))
snmp_log("iter %u size %u remaining %u/%u", ind, c.buffer - sk->tpos, size, pkt_size);
else
snmp_log("iter+ %u size %u remaining %u/%u", ind, c.buffer - ((struct additional_buffer *) TAIL(p->additional_buffers))->buf, size, pkt_size);
#endif
2023-07-26 12:30:34 +00:00
if (size < snmp_oid_sizeof(0))
goto partial;
/* We load search range start OID */
const struct oid *o_start_b = (void *) pkt;
uint sz;
if ((sz = snmp_oid_size(o_start_b)) > pkt_size)
{
2023-09-04 07:25:51 +00:00
/* for malformed packets consume full pkt_size [or size] */
2023-07-26 12:30:34 +00:00
c.error = AGENTX_RES_PARSE_ERROR; /* Packet error, inconsistent values */
goto send;
}
/*
2023-09-04 07:25:51 +00:00
* If we already have written same relevant data to the TX buffer, then
2023-07-26 12:30:34 +00:00
* we send processed part, otherwise we don't have anything to send and
* need to wait for more data to be recieved.
*/
if (sz > size && ind > 1)
{
snmp_log("sz %u > %u size && ind %u > 1", sz, size, ind);
2023-09-04 07:25:51 +00:00
goto partial; /* send already processed part */
2023-07-26 12:30:34 +00:00
}
else if (sz > size)
{
snmp_log("sz %u > %u size; returning 0", sz, size);
goto wait;
}
/* update buffer pointer and remaining size counters */
ADVANCE(pkt, pkt_size, sz);
size -= sz;
2023-08-08 19:51:38 +00:00
/*
2023-09-04 07:25:51 +00:00
* We load search range end OID
2023-07-26 12:30:34 +00:00
* The exactly same process of sanity checking is preformed while loading
2023-09-04 07:25:51 +00:00
* the SearchRange's end OID
2023-07-26 12:30:34 +00:00
*/
const struct oid *o_end_b = (void *) pkt;
if ((sz = snmp_oid_size(o_end_b)) > pkt_size)
{
c.error = AGENTX_RES_PARSE_ERROR; /* Packet error, inconsistent values */
goto send;
}
if (sz > size && ind > 1)
{
snmp_log("sz2 %u > %u size && ind %u > 1", sz, size, ind);
size += snmp_oid_size(o_start_b);
goto partial;
}
else if (sz > size)
{
snmp_log("sz2 %u > %u size; returning 0", sz, size);
goto wait;
}
ADVANCE(pkt, pkt_size, sz);
size -= sz;
2023-08-08 18:47:30 +00:00
// TODO check for oversized oids before any allocation (in prefixize())
2023-07-26 12:30:34 +00:00
/* We create copy of OIDs outside of rx-buffer and also prefixize them */
o_start = snmp_prefixize(p, o_start_b, c.byte_ord);
o_end = snmp_prefixize(p, o_end_b, c.byte_ord);
if (!snmp_is_oid_empty(o_end) && snmp_oid_compare(o_start, o_end) > 0)
{
snmp_log("snmp_gets2() o_start does not preceed o_end, returning GEN_ERROR");
c.error = AGENTX_RES_GEN_ERROR;
goto send;
}
2023-09-04 07:25:51 +00:00
/* TODO find mib_class, check if type is GET of GET_NEXT, act acordingly */
2023-07-26 12:30:34 +00:00
switch (h->type)
{
case AGENTX_GET_PDU:
snmp_mib_fill2(p, o_start, &c);
break;
case AGENTX_GET_NEXT_PDU:
snmp_get_next2(p, o_start, o_end, &c);
break;
case AGENTX_GET_BULK_PDU:
snmp_get_bulk2(p, o_start, o_end, &bulk_state, &c);
break;
default:
die("incorrect usage");
}
mb_free(o_start);
o_start = NULL;
mb_free(o_end);
o_end = NULL;
ind++;
} /* while (c.error == AGENTX_RES_NO_ERROR && size > 0) */
send:
snmp_log("gets2: sending response ...");
2023-09-04 07:25:51 +00:00
struct agentx_response *res = (void *) sk->tbuf;
/* update the error, index pair on the beginning of the packet */
response_err_ind(res, c.error, ind);
2023-08-08 19:51:38 +00:00
uint s = update_packet_size(p, (byte *) response_header, c.buffer);
2023-09-04 07:25:51 +00:00
snmp_log("sending response to Get-PDU, GetNext-PDU or GetBulk-PDU request (size %u)...", s);
/* send the message in TX buffer */
int ret = sk_send(sk, s);
2023-09-04 07:25:51 +00:00
if (ret > 0)
snmp_log("sk_send OK!");
else if (ret == 0)
snmp_log("sk_send sleep");
else
snmp_log("sk_send error");
// TODO think through the error state
p->partial_response = NULL;
/*
int ret = sk_send(sk, c.buffer - sk->tpos);
2023-07-26 12:30:34 +00:00
if (ret == 0)
snmp_log("sk_send sleep (gets2");
else if (ret < 0)
snmp_log("sk_send err %d (gets2)", ret);
else
snmp_log("sk_send was successful (gets2) !");
2023-09-04 07:25:51 +00:00
*/
2023-07-26 12:30:34 +00:00
mb_free(context);
mb_free(o_start);
mb_free(o_end);
2023-09-04 07:25:51 +00:00
/* number of bytes parsed form rx-buffer */
2023-07-26 12:30:34 +00:00
return pkt - pkt_start;
partial:
snmp_log("partial packet");
/* The context octet is not added into response pdu */
2023-09-04 07:25:51 +00:00
/* need to tweak RX buffer packet size */
snmp_log("old rx-buffer size %u", h->payload);
(c.byte_ord) ? put_u32(&h->payload, pkt_size) : (h->payload = pkt_size);
snmp_log("new rx-buffer size %u", h->payload);
2023-07-26 12:30:34 +00:00
*skip = AGENTX_HEADER_SIZE;
p->partial_response = response_header;
return pkt - pkt_start;
2023-07-26 12:30:34 +00:00
wait:
mb_free(context);
mb_free(o_start);
mb_free(o_end);
return 0;
}
void
2022-08-10 15:31:32 +00:00
snmp_start_subagent(struct snmp_proto *p)
{
2022-12-10 12:22:37 +00:00
snmp_log("snmp_start_subagent() starting subagent");
2023-03-14 13:10:08 +00:00
snmp_log("DEBUG p->local_as %u", p->local_as);
2022-08-10 15:31:32 +00:00
2023-09-04 07:25:51 +00:00
/* blank oid means unsupported */
2022-12-10 17:08:00 +00:00
struct oid *blank = snmp_oid_blank(p);
open_pdu(p, blank);
2023-09-04 07:25:51 +00:00
p->state = SNMP_OPEN;
2022-12-10 17:08:00 +00:00
mb_free(blank);
}
void
2022-08-10 15:31:32 +00:00
snmp_stop_subagent(struct snmp_proto *p)
{
2023-09-04 07:25:51 +00:00
snmp_log("snmp_stop_subagent() state %d", p->state);
2022-08-10 15:31:32 +00:00
2022-12-17 17:16:19 +00:00
if (p->state == SNMP_STOP)
2022-09-20 12:28:57 +00:00
close_pdu(p, AGENTX_CLOSE_SHUTDOWN);
2022-08-10 15:31:32 +00:00
}
2022-11-05 15:29:00 +00:00
static inline int
oid_prefix(struct oid *o, u32 *prefix, uint len)
{
for (uint i = 0; i < len; i++)
if (o->ids[i] != prefix[i])
2023-08-08 19:51:38 +00:00
return 0;
2023-08-08 19:51:38 +00:00
return 1;
}
2023-03-14 13:10:08 +00:00
int
snmp_rx(sock *sk, uint size)
{
snmp_log("snmp_rx() size %u", size);
2023-09-04 07:25:51 +00:00
//snmp_dump_packet(sk->tbuf, 64);
2023-03-14 13:10:08 +00:00
struct snmp_proto *p = sk->data;
byte *pkt_start = sk->rbuf;
2023-07-26 12:30:34 +00:00
byte *end = pkt_start + size;
snmp_log("snmp_rx rbuf 0x%p rpos 0x%p", sk->rbuf, sk->rpos);
2023-07-26 12:30:34 +00:00
/*
* In some cases we want to save the header for future parsing, skip is number
* of bytes that should not be overriden by memmove()
*/
uint skip = 0;
2023-03-14 13:10:08 +00:00
snmp_log("snmp_rx before loop");
2023-09-04 07:25:51 +00:00
while (end >= pkt_start + AGENTX_HEADER_SIZE && skip == 0)
2023-03-14 13:10:08 +00:00
{
2023-07-26 12:30:34 +00:00
uint parsed_len = parse_pkt(p, pkt_start, size, &skip);
2023-03-14 13:10:08 +00:00
snmp_log("snmp_rx loop end %p parsed >>> %u <<< curr %p", end, parsed_len,
pkt_start + parsed_len);
2023-09-04 07:25:51 +00:00
snmp_log("snmp_rx loop2 rpos 0x%p", sk->rpos);
2023-03-14 13:10:08 +00:00
if (parsed_len == 0)
break;
pkt_start += parsed_len;
size -= parsed_len;
}
snmp_log("snmp_rx loop finished");
2023-08-08 18:47:30 +00:00
/* Incomplete packets */
if (skip != 0 || pkt_start != end)
2023-03-14 13:10:08 +00:00
{
2023-08-08 18:47:30 +00:00
snmp_log("snmp_rx memmove");
snmp_dump_packet(sk->rbuf, SNMP_RX_BUFFER_SIZE);
memmove(sk->rbuf + skip, pkt_start, size);
snmp_log("after change; sk->rbuf 0x%p sk->rpos 0x%p", sk->rbuf, sk->rpos);
snmp_dump_packet(sk->rbuf, size + skip);
snmp_log("tweaking rpos 0x%p (size %u skip %u)", sk->rpos, size, skip);
sk->rpos = sk->rbuf + size + skip;
snmp_log("snmp_rx returing 0");
2023-03-14 13:10:08 +00:00
return 0;
}
snmp_log("snmp_rx returning 1");
return 1;
}
2022-09-06 16:04:29 +00:00
/* ping pdu */
void
snmp_ping(struct snmp_proto *p)
{
2023-09-04 07:25:51 +00:00
{
#define SNMP_OID_SIZE_FROM_LEN(x) (sizeof(struct oid) + (x) * sizeof(u32))
/* trap OID bgpEstablishedNotification (.1.3.6.1.2.1.0.1) */
struct oid *head = mb_alloc(p->p.pool, SNMP_OID_SIZE_FROM_LEN(3));
head->n_subid = 3;
head->prefix = 2;
head->include = head->pad = 0;
u32 trap_ids[] = { 1, 0, 1 };
for (uint i = 0; i < head->n_subid; i++)
head->ids[i] = trap_ids[i];
/* paylaod oids */
uint sz = 3 * SNMP_OID_SIZE_FROM_LEN(9) + 3 * 4 + 2 * 8 + 4 - 20;
void *data = mb_alloc(p->p.pool, sz);
struct agentx_varbind *addr_vb = data;
/* +4 for varbind header, +8 for octet string */
struct agentx_varbind *error_vb = data + SNMP_OID_SIZE_FROM_LEN(9) + 4 + 8;
struct agentx_varbind *state_vb = (void *) error_vb + SNMP_OID_SIZE_FROM_LEN(9) + 4 + 8;
#undef SNMP_OID_SIZE_FROM_LEN
addr_vb->pad = error_vb->pad = state_vb->pad = 0;
struct oid *addr = &addr_vb->name;
struct oid *error = &error_vb->name;
struct oid *state = &state_vb->name;
addr->n_subid = error->n_subid = state->n_subid = 9;
addr->prefix = error->prefix = state->prefix = 2;
addr->include = error->include = state->include = 0;
addr->pad = error->pad = state->pad = 0;
u32 oid_ids[] = {
SNMP_MIB_2, SNMP_BGP4_MIB, SNMP_BGP_PEER_TABLE, SNMP_BGP_PEER_ENTRY, 0,
10, 1, 2, 1
};
for (uint i = 0; i < addr->n_subid; i++)
addr->ids[i] = error->ids[i] = state->ids[i] = oid_ids[i];
addr->ids[4] = SNMP_BGP_REMOTE_ADDR;
error->ids[4] = SNMP_BGP_LAST_ERROR;
state->ids[4] = SNMP_BGP_STATE;
ip4_addr ip4 = ip4_build(10,1,2,1);
snmp_varbind_ip4(addr_vb, 100, ip4);
char error_str[] = { 0, 0 };
snmp_varbind_nstr(error_vb, 100, error_str, 2);
snmp_varbind_int(state_vb, 100, 6);
u32 now = current_time() TO_S;
if (now % 4 == 0)
notify_pdu(p, head, data, sz, 1);
else if (now % 2 == 0)
notify_pdu(p, head, data, sz, 0);
}
2022-08-10 15:31:32 +00:00
sock *sk = p->sock;
2023-09-04 07:25:51 +00:00
snmp_dump_packet(sk->tpos, AGENTX_HEADER_SIZE + 4);
snmp_log("snmp_ping sk->tpos 0x%p", sk->tpos);
2023-09-04 11:46:02 +00:00
struct snmp_pdu_context c = SNMP_PDU_CONTEXT(sk);
2022-09-06 16:04:29 +00:00
2023-07-26 12:30:34 +00:00
if (c.size < AGENTX_HEADER_SIZE)
snmp_manage_tbuf(p, &c);
2022-09-30 07:36:09 +00:00
2023-07-26 12:30:34 +00:00
snmp_log("ping_pdu()");
struct agentx_header *h = (struct agentx_header *) c.buffer;
ADVANCE(c.buffer, c.size, AGENTX_HEADER_SIZE);
SNMP_BLANK_HEADER(h, AGENTX_PING_PDU);
p->packet_id++;
2023-07-26 12:30:34 +00:00
SNMP_SESSION(h, p);
2023-09-04 11:46:02 +00:00
c.byte_ord = AGENTX_NETWORK_BYTE_ORDER;
2023-09-04 07:25:51 +00:00
snmp_log("sending ping packet ... tpos 0x%p", sk->tpos);
snmp_dump_packet(sk->tpos, AGENTX_HEADER_SIZE + 4);
/* sending only header -> pkt - buf */
uint s = update_packet_size(p, sk->tpos, c.buffer);
int ret = sk_send(sk, s);
if (ret > 0)
snmp_log("sk_send OK!");
else if (ret == 0)
snmp_log("sk_send sleep");
else
snmp_log("sk_send error");
}
2022-08-10 15:31:32 +00:00
/*
2022-09-06 16:04:29 +00:00
void
snmp_agent_reconfigure(void)
2022-08-10 15:31:32 +00:00
{
2022-09-06 16:04:29 +00:00
}
2023-09-04 07:25:51 +00:00
2022-09-30 07:36:09 +00:00
*/
2022-09-20 12:28:57 +00:00
static inline int
is_bgp4_mib_prefix(struct oid *o)
{
if (o->prefix == 2 && o->ids[0] == 15)
2022-09-30 07:36:09 +00:00
return 1;
2022-09-20 12:28:57 +00:00
else
return 0;
}
2022-09-30 07:36:09 +00:00
static inline int
has_inet_prefix(struct oid *o)
{
return (o->n_subid > 4 && o->ids[0] == 1 &&
o->ids[1] == 3 && o->ids[2] == 6 &&
o->ids[3] == 1);
}
2022-12-17 17:16:19 +00:00
/**
2023-07-26 12:30:34 +00:00
* snmp_search_check_end_oid - check if oid is before SearchRange end
2022-12-17 17:16:19 +00:00
*
* @found: best oid found in MIB tree
* @bound: upper bound specified in SearchRange
*
* check if found oid meet the SearchRange upper bound condition in
* lexicographical order, returns boolean value
*/
2023-07-26 12:30:34 +00:00
int snmp_search_check_end_oid(const struct oid *found, const struct oid *bound)
2022-09-20 12:28:57 +00:00
{
2023-03-24 14:02:23 +00:00
snmp_log("upper_bound_check(*f, *b) %p %p is_empty() %d", found, bound,
snmp_is_oid_empty(bound));
2022-12-17 17:16:19 +00:00
if (snmp_is_oid_empty(bound))
return 1;
2023-07-26 12:30:34 +00:00
return (snmp_oid_compare(found, bound) < 0);
2022-12-17 17:16:19 +00:00
}
2023-07-26 12:30:34 +00:00
/* tree is tree with "internet" prefix .1.3.6.1
working only with o_start, o_end allocated in heap (not from buffer)*/
static struct oid *
search_mib(struct snmp_proto *p, const struct oid *o_start, const struct oid *o_end,
struct oid *o_curr, struct snmp_pdu_context *c,
enum snmp_search_res *result)
2022-12-17 17:16:19 +00:00
{
2023-07-26 12:30:34 +00:00
snmp_log("search_mib()");
ASSUME(o_start != NULL);
2022-09-30 07:36:09 +00:00
2023-07-26 12:30:34 +00:00
if (o_curr && (o_curr->n_subid < 2 || o_curr->ids[0] != 1))
return NULL;
if (!o_curr && (o_start->n_subid < 2 || o_start->ids[0] != 1))
2022-11-05 15:29:00 +00:00
return NULL;
2022-09-30 07:36:09 +00:00
2022-11-05 15:29:00 +00:00
if (!o_curr)
{
2023-07-26 12:30:34 +00:00
o_curr = snmp_oid_duplicate(p->p.pool, o_start);
2022-11-05 15:29:00 +00:00
// XXX is it right time to free o_start right now (here) ?
2023-07-26 12:30:34 +00:00
// not for use in snmp_get_next2() the o_start comes and ends in _gets2_()
2022-11-05 15:29:00 +00:00
}
2022-09-30 07:36:09 +00:00
2023-07-26 12:30:34 +00:00
const struct oid *blank = NULL;
if (!snmp_is_oid_empty(o_end) &&
snmp_get_mib_class(o_curr) < snmp_get_mib_class(o_end))
2022-09-30 07:36:09 +00:00
{
2023-07-26 12:30:34 +00:00
o_end = blank = snmp_oid_blank(p);
snmp_log("search_mib() o_end points to blank oid now %p", o_end);
}
2022-11-19 22:00:02 +00:00
2023-07-26 12:30:34 +00:00
enum snmp_search_res r;
switch (o_curr->ids[1])
{
case SNMP_BGP4_MIB:
r = snmp_bgp_search2(p, &o_curr, o_end, c->context);
2022-12-17 17:16:19 +00:00
2023-07-26 12:30:34 +00:00
if (r == SNMP_SEARCH_OK)
{
*result = r;
break;
return o_curr;
}
2022-11-19 22:00:02 +00:00
2023-07-26 12:30:34 +00:00
// TODO add early break for o_end less then thinkable maximum in each tree
2022-11-19 22:00:02 +00:00
2023-07-26 12:30:34 +00:00
/* fall through */
2022-12-17 17:16:19 +00:00
2023-07-26 12:30:34 +00:00
default:
if (o_curr) mb_free(o_curr);
2023-07-26 12:30:34 +00:00
o_curr = snmp_oid_duplicate(p->p.pool, o_start);
*result = SNMP_SEARCH_END_OF_VIEW;
break;
2022-09-30 07:36:09 +00:00
}
2022-11-05 15:29:00 +00:00
2023-08-08 19:51:38 +00:00
if (o_end == blank)
2023-09-04 07:25:51 +00:00
mb_free((void *) blank);
2022-12-17 17:16:19 +00:00
2023-09-04 07:25:51 +00:00
return o_curr;
2022-09-06 16:04:29 +00:00
}
2022-08-10 15:31:32 +00:00
2022-12-17 17:24:05 +00:00
/**
* snmp_prefixize - return prefixed oid copy if possible
2022-11-05 15:29:00 +00:00
* @proto: allocation pool holder
* @oid: from packet loaded object identifier
2022-12-17 17:24:05 +00:00
* @byte_ord: byte order of @oid
2022-11-05 15:29:00 +00:00
*
* Returns prefixed (meaning with nonzero prefix field) oid copy of @oid if
2022-12-10 17:08:00 +00:00
* possible, NULL otherwise. Returned pointer is always allocated from @proto's
2022-11-05 15:29:00 +00:00
* pool not a pointer to recieve buffer (from which is most likely @oid).
*/
struct oid *
2023-07-26 12:30:34 +00:00
snmp_prefixize(struct snmp_proto *proto, const struct oid *oid, int byte_ord)
2022-09-20 12:28:57 +00:00
{
2023-07-26 12:30:34 +00:00
ASSERT(oid != NULL);
2023-03-24 14:02:23 +00:00
snmp_log("snmp_prefixize()");
2022-09-20 12:28:57 +00:00
const u32 prefix[] = {1, 3, 6, 1};
if (snmp_is_oid_empty(oid))
{
2023-09-04 07:25:51 +00:00
/* allocate new zeroed oid */
2023-03-24 14:02:23 +00:00
snmp_log("blank");
2022-12-10 17:08:00 +00:00
return snmp_oid_blank(proto);
}
2023-03-24 14:02:23 +00:00
2023-09-04 07:25:51 +00:00
/* already in prefixed form */
2023-03-24 14:02:23 +00:00
else if (oid->prefix != 0) {
2023-07-26 12:30:34 +00:00
struct oid *new = snmp_oid_duplicate(proto->p.pool, oid);
2023-03-24 14:02:23 +00:00
snmp_log("already prefixed");
return new;
}
2022-11-05 15:29:00 +00:00
if (oid->n_subid < 5)
2023-09-04 07:25:51 +00:00
{ snmp_log("too small"); return NULL; }
2022-09-20 12:28:57 +00:00
for (int i = 0; i < 4; i++)
if (LOAD_U32(oid->ids[i], byte_ord) != prefix[i])
2023-03-24 14:02:23 +00:00
{ snmp_log("different prefix"); return NULL; }
2022-09-20 12:28:57 +00:00
2023-09-04 07:25:51 +00:00
/* validity check here */
2022-11-05 15:29:00 +00:00
if (oid->ids[4] >= 256)
2023-03-24 14:02:23 +00:00
{ snmp_log("outside byte first id"); return NULL; }
2022-11-05 15:29:00 +00:00
2022-12-17 17:24:05 +00:00
struct oid *new = mb_alloc(proto->p.pool,
2022-11-05 15:29:00 +00:00
sizeof(struct oid) + MAX((oid->n_subid - 5) * sizeof(u32), 0));
2022-09-20 12:28:57 +00:00
2022-11-05 15:29:00 +00:00
memcpy(new, oid, sizeof(struct oid));
new->n_subid = oid->n_subid - 5;
2022-09-20 12:28:57 +00:00
2023-09-04 07:25:51 +00:00
/* validity check before allocation => ids[4] < 256
and can be copied to one byte new->prefix */
2022-11-05 15:29:00 +00:00
new->prefix = oid->ids[4];
2022-09-20 12:28:57 +00:00
2022-11-05 15:29:00 +00:00
memcpy(&new->ids, &oid->ids[5], new->n_subid * sizeof(u32));
2022-09-20 12:28:57 +00:00
return new;
}
2023-07-26 12:30:34 +00:00
static void
snmp_mib_fill2(struct snmp_proto *p, struct oid *oid,
struct snmp_pdu_context *c)
2022-11-19 22:00:02 +00:00
{
2023-07-26 12:30:34 +00:00
ASSUME(oid != NULL);
2022-11-05 15:29:00 +00:00
2023-09-04 07:25:51 +00:00
snmp_log("critical part");
2023-07-26 12:30:34 +00:00
if (c->size < snmp_varbind_hdr_size_from_oid(oid))
{
2023-08-08 19:51:38 +00:00
snmp_manage_tbuf(p, c);
2023-07-26 12:30:34 +00:00
}
2023-09-04 07:25:51 +00:00
else
{
u32 *ptr = mb_alloc(p->p.pool, sizeof(u32) * 4);
*ptr = 0xbeef;
mb_free(ptr);
}
snmp_log("critical part done");
2022-11-19 22:00:02 +00:00
2023-07-26 12:30:34 +00:00
struct agentx_varbind *vb = snmp_create_varbind(c->buffer, oid);
2022-11-05 15:29:00 +00:00
if (oid->n_subid < 2 || (oid->prefix != 2 && oid->ids[0] != 1))
2022-11-19 22:00:02 +00:00
{
2022-12-17 17:24:05 +00:00
vb->type = AGENTX_NO_SUCH_OBJECT;
2023-07-26 12:30:34 +00:00
ADVANCE(c->buffer, c->size, snmp_varbind_header_size(vb));
return;
2022-11-19 22:00:02 +00:00
}
2022-11-05 15:29:00 +00:00
2023-07-26 12:30:34 +00:00
u8 mib_class = snmp_get_mib_class(oid);
2022-11-05 15:29:00 +00:00
switch (mib_class)
{
case SNMP_CLASS_BGP:
2023-07-26 12:30:34 +00:00
snmp_bgp_fill(p, vb, c);
break;
2023-07-26 12:30:34 +00:00
case SNMP_CLASS_INVALID:
case SNMP_CLASS_END:
default:
vb->type = AGENTX_NO_SUCH_OBJECT;
ADVANCE(c->buffer, c->size, snmp_varbind_header_size(vb));
2022-11-05 15:29:00 +00:00
}
2023-07-26 12:30:34 +00:00
}
/**
*
* Important note: After managing insufficient buffer size all in buffer pointers
* are invalidated!
*/
void
2023-09-04 07:25:51 +00:00
snmp_manage_tbuf(struct snmp_proto UNUSED *p, struct snmp_pdu_context *c)
2022-09-06 16:04:29 +00:00
{
2023-07-26 12:30:34 +00:00
snmp_log("snmp_manage_tbuf()");
sock *sk = p->sock;
2023-09-04 07:25:51 +00:00
sk_set_tbsize(sk, sk->tbsize + 2048);
c->size += 2048;
2023-09-04 07:25:51 +00:00
//sk_set_tbsize(sk, sk->tbsize + SNMP_TX_BUFFER_SIZE);
//c->size += SNMP_TX_BUFFER_SIZE;
return;
#if 0
if (!EMPTY_LIST(p->additional_buffers))
{
struct additional_buffer *t = TAIL(p->additional_buffers);
t->pos = c->buffer;
}
else
p->to_send = c->buffer - p->sock->tpos;
struct additional_buffer *b = mb_allocz(p->p.pool, sizeof(struct additional_buffer));
b->buf = b->pos = mb_alloc(p->p.pool, SNMP_TX_BUFFER_SIZE);
add_tail(&p->additional_buffers, &b->n);
c->buffer = b->buf;
c->size = SNMP_TX_BUFFER_SIZE;
#endif
2022-08-10 15:31:32 +00:00
}
2023-09-04 07:25:51 +00:00
#if 0
static int
send_remaining_buffers(sock *sk)
{
struct snmp_proto *p = sk->data;
while (!EMPTY_LIST(p->additional_buffers))
{
struct additional_buffer *b = HEAD(p->additional_buffers);
p->to_send = b->pos - b->buf;
snmp_log("send_remaining_buffers sending next %u bytes", p->to_send);
ASSUME(sk->tbuf == sk->tpos);
memcpy(sk->tbuf, b->buf, p->to_send);
sk->tpos = sk->tbuf + p->to_send;
rem_node(&b->n);
snmp_log("state of additional b at 0x%p .buf = 0x%p .pos = 0x%p", b, b->buf, b->pos);
mb_free(b->buf);
snmp_log("b->buf fried, cause is b");
mb_free(b);
snmp_log("packet byte stream part next");
snmp_dump_packet(sk->tpos, p->to_send);
int ret;
if ((ret = sk_send(sk, p->to_send)) <= 0)
{
snmp_log("sending_remaining - error or sleep;returning");
return ret;
}
}
snmp_log("sending_remaining all done returning 1");
return 1;
}
#endif
2023-09-04 11:46:02 +00:00
/*
2023-07-26 12:30:34 +00:00
void
2023-08-08 18:47:30 +00:00
snmp_tx(sock UNUSED *sk)
2022-09-20 12:28:57 +00:00
{
2023-09-04 07:25:51 +00:00
snmp_log("snmp_tx");
2022-09-20 12:28:57 +00:00
}
2023-09-04 11:46:02 +00:00
*/
2022-09-20 12:28:57 +00:00
2023-07-26 12:30:34 +00:00
static struct agentx_response *
prepare_response(struct snmp_proto *p, struct snmp_pdu_context *c)
{
snmp_log("prepare_response()");
2023-09-04 07:25:51 +00:00
if (!p->partial_response)
{
struct agentx_response *r = (void *) c->buffer;
struct agentx_header *h = &r->h;
2023-08-08 18:47:30 +00:00
2023-09-04 07:25:51 +00:00
SNMP_BLANK_HEADER(h, AGENTX_RESPONSE_PDU);
SNMP_SESSION(h, p);
2023-07-26 12:30:34 +00:00
2023-09-04 07:25:51 +00:00
/* protocol doesn't care about subagent upTime */
STORE_U32(r->uptime, 0);
STORE_U16(r->error, AGENTX_RES_NO_ERROR);
STORE_U16(r->index, 0);
2023-07-26 12:30:34 +00:00
2023-09-04 07:25:51 +00:00
ADVANCE(c->buffer, c->size, sizeof(struct agentx_response));
return r;
}
2023-07-26 12:30:34 +00:00
2023-09-04 07:25:51 +00:00
return p->partial_response;
2023-07-26 12:30:34 +00:00
}
2022-08-10 15:31:32 +00:00
#undef SNMP_ERR_SHIFT