/* * BIRD Resource Manager -- A SLAB-like Memory Allocator * * Heavily inspired by the original SLAB paper by Jeff Bonwick. * * (c) 1998--2000 Martin Mares * (c) 2020 Maria Matejka * * Can be freely distributed and used under the terms of the GNU GPL. */ /** * DOC: Slabs * * Slabs are collections of memory blocks of a fixed size. * They support very fast allocation and freeing of such blocks, prevent memory * fragmentation and optimize L2 cache usage. Slabs have been invented by Jeff Bonwick * and published in USENIX proceedings as `The Slab Allocator: An Object-Caching Kernel * Memory Allocator'. Our implementation follows this article except that we don't use * constructors and destructors. * * When the |DEBUGGING| switch is turned on, we automatically fill all * newly allocated and freed blocks with a special pattern to make detection * of use of uninitialized or already freed memory easier. * * Example: Nodes of a FIB are allocated from a per-FIB Slab. */ #include #include #include "nest/bird.h" #include "lib/resource.h" #include "lib/string.h" #include "lib/tlists.h" #undef FAKE_SLAB /* Turn on if you want to debug memory allocations */ #ifdef DEBUGGING #define POISON /* Poison all regions after they are freed */ #endif static void slab_free(resource *r); static void slab_dump(resource *r, unsigned indent); static resource *slab_lookup(resource *r, unsigned long addr); static struct resmem slab_memsize(resource *r); #ifdef FAKE_SLAB /* * Fake version used for debugging. */ struct slab { resource r; uint size; list objs; }; static struct resclass sl_class = { "FakeSlab", sizeof(struct slab), slab_free, slab_dump, NULL, slab_memsize }; struct sl_obj { node n; uintptr_t data_align[0]; byte data[0]; }; slab * sl_new(pool *p, uint size) { slab *s = ralloc(p, &sl_class); s->size = size; init_list(&s->objs); return s; } void * sl_alloc(slab *s) { struct sl_obj *o = xmalloc(sizeof(struct sl_obj) + s->size); add_tail(&s->objs, &o->n); return o->data; } void * sl_allocz(slab *s) { void *obj = sl_alloc(s); memset(obj, 0, s->size); return obj; } void sl_free(void *oo) { SKIP_BACK_DECLARE(struct sl_obj, o, data, oo); rem_node(&o->n); xfree(o); } static void slab_free(resource *r) { slab *s = (slab *) r; struct sl_obj *o, *p; for(o = HEAD(s->objs); p = (struct sl_obj *) o->n.next; o = p) xfree(o); } static void slab_dump(resource *r, unsigned indent UNUSED) { slab *s = (slab *) r; int cnt = 0; struct sl_obj *o; WALK_LIST(o, s->objs) cnt++; debug("(%d objects per %d bytes)\n", cnt, s->size); } static struct resmem slab_memsize(resource *r) { slab *s = (slab *) r; size_t cnt = 0; struct sl_obj *o; WALK_LIST(o, s->objs) cnt++; return (struct resmem) { .effective = cnt * s->size, .overhead = ALLOC_OVERHEAD + sizeof(struct slab) + cnt * ALLOC_OVERHEAD, }; } #else /* * Real efficient version. */ #define MAX_EMPTY_HEADS 1 enum sl_head_state { slh_empty = 2, slh_partial = 0, slh_full = 1, } PACKED; struct sl_head { struct slab *slab; TLIST_NODE(sl_head, struct sl_head) n; u16 num_full; enum sl_head_state state; u32 used_bits[0]; }; struct sl_alignment { /* Magic structure for testing of alignment */ byte data; int x[0]; _Atomic u64 y[0]; void *z[0]; }; #define TLIST_PREFIX sl_head #define TLIST_TYPE struct sl_head #define TLIST_ITEM n #define TLIST_WANT_WALK #define TLIST_WANT_ADD_HEAD #include "lib/tlists.h" struct slab { resource r; uint obj_size, head_size, head_bitfield_len; uint objs_per_slab, num_empty_heads, data_size; struct sl_head_list empty_heads, partial_heads, full_heads; }; static struct resclass sl_class = { "Slab", sizeof(struct slab), slab_free, slab_dump, slab_lookup, slab_memsize }; #define SL_GET_HEAD(x) PAGE_HEAD(x) #define SL_HEAD_CHANGE_STATE(_s, _h, _from, _to) ({ \ ASSERT_DIE(_h->state == slh_##_from); \ sl_head_rem_node(&_s->_from##_heads, _h); \ sl_head_add_head(&_s->_to##_heads, _h); \ _h->state = slh_##_to; \ }) /** * sl_new - create a new Slab * @p: resource pool * @size: block size * * This function creates a new Slab resource from which * objects of size @size can be allocated. */ slab * sl_new(pool *p, uint size) { slab *s = ralloc(p, &sl_class); uint align = sizeof(struct sl_alignment); if (align < sizeof(void *)) align = sizeof(void *); s->data_size = size; size = (size + align - 1) / align * align; s->obj_size = size; s->head_size = sizeof(struct sl_head); do { s->objs_per_slab = (page_size - s->head_size) / size; s->head_bitfield_len = (s->objs_per_slab + 31) / 32; s->head_size = ( sizeof(struct sl_head) + sizeof(u32) * s->head_bitfield_len + align - 1) / align * align; } while (s->objs_per_slab * size + s->head_size > (size_t) page_size); if (!s->objs_per_slab) bug("Slab: object too large"); s->num_empty_heads = 0; return s; } /** * sl_delete - destroy an existing Slab * @s: slab * * This function destroys the given Slab. */ void sl_delete(slab *s) { rfree(&s->r); } /** * sl_alloc - allocate an object from Slab * @s: slab * * sl_alloc() allocates space for a single object from the * Slab and returns a pointer to the object. */ void * sl_alloc(slab *s) { struct sl_head *h; ASSERT_DIE(DG_IS_LOCKED(resource_parent(&s->r)->domain)); redo: if (!(h = s->partial_heads.first)) goto no_partial; okay: for (uint i=0; ihead_bitfield_len; i++) if (~h->used_bits[i]) { uint pos = u32_ctz(~h->used_bits[i]); if (i * 32 + pos >= s->objs_per_slab) break; h->used_bits[i] |= 1 << pos; h->num_full++; void *out = ((void *) h) + s->head_size + (i * 32 + pos) * s->obj_size; #ifdef POISON memset(out, 0xcd, s->data_size); #endif return out; } SL_HEAD_CHANGE_STATE(s, h, partial, full); goto redo; no_partial: if (h = s->empty_heads.first) { SL_HEAD_CHANGE_STATE(s, h, empty, partial); s->num_empty_heads--; goto okay; } h = alloc_page(); ASSERT_DIE(SL_GET_HEAD(h) == h); #ifdef POISON memset(h, 0xba, page_size); #endif memset(h, 0, s->head_size); h->slab = s; sl_head_add_head(&s->partial_heads, h); goto okay; } /** * sl_allocz - allocate an object from Slab and zero it * @s: slab * * sl_allocz() allocates space for a single object from the * Slab and returns a pointer to the object after zeroing out * the object memory. */ void * sl_allocz(slab *s) { void *obj = sl_alloc(s); memset(obj, 0, s->data_size); return obj; } /** * sl_free - return a free object back to a Slab * @s: slab * @oo: object returned by sl_alloc() * * This function frees memory associated with the object @oo * and returns it back to the Slab @s. */ void sl_free(void *oo) { struct sl_head *h = SL_GET_HEAD(oo); struct slab *s = h->slab; ASSERT_DIE(DG_IS_LOCKED(resource_parent(&s->r)->domain)); #ifdef POISON memset(oo, 0xdb, s->data_size); #endif uint offset = oo - ((void *) h) - s->head_size; ASSERT_DIE(offset % s->obj_size == 0); uint pos = offset / s->obj_size; ASSERT_DIE(pos < s->objs_per_slab); h->used_bits[pos / 32] &= ~(1 << (pos % 32)); if ((h->num_full-- == s->objs_per_slab) && (h->state == slh_full)) SL_HEAD_CHANGE_STATE(s, h, full, partial); else if (!h->num_full) { sl_head_rem_node(&s->partial_heads, h); if (s->num_empty_heads >= MAX_EMPTY_HEADS) { #ifdef POISON memset(h, 0xde, page_size); #endif free_page(h); } else { sl_head_add_head(&s->empty_heads, h); h->state = slh_empty; s->num_empty_heads++; } } } static void slab_free(resource *r) { slab *s = (slab *) r; WALK_TLIST_DELSAFE(sl_head, h, &s->empty_heads) free_page(h); WALK_TLIST_DELSAFE(sl_head, h, &s->partial_heads) free_page(h); WALK_TLIST_DELSAFE(sl_head, h, &s->full_heads) free_page(h); } static void slab_dump(resource *r, unsigned indent UNUSED) { slab *s = (slab *) r; int ec=0, pc=0, fc=0; WALK_TLIST(sl_head, h, &s->empty_heads) ec++; WALK_TLIST(sl_head, h, &s->partial_heads) pc++; WALK_TLIST(sl_head, h, &s->full_heads) fc++; debug("(%de+%dp+%df blocks per %d objs per %d bytes)\n", ec, pc, fc, s->objs_per_slab, s->obj_size); char x[16]; bsprintf(x, "%%%ds%%s %%p\n", indent + 2); WALK_TLIST(sl_head, h, &s->full_heads) debug(x, "", "full", h); WALK_TLIST(sl_head, h, &s->partial_heads) debug(x, "", "partial", h); WALK_TLIST(sl_head, h, &s->empty_heads) debug(x, "", "empty", h); } static struct resmem slab_memsize(resource *r) { slab *s = (slab *) r; size_t heads = 0; WALK_TLIST(sl_head, h, &s->full_heads) heads++; size_t items = heads * s->objs_per_slab; WALK_TLIST(sl_head, h, &s->partial_heads) { heads++; items += h->num_full; } WALK_TLIST(sl_head, h, &s->empty_heads) heads++; size_t eff = items * s->data_size; return (struct resmem) { .effective = eff, .overhead = ALLOC_OVERHEAD + sizeof(struct slab) + heads * page_size - eff, }; } static resource * slab_lookup(resource *r, unsigned long a) { slab *s = (slab *) r; WALK_TLIST(sl_head, h, &s->partial_heads) if ((unsigned long) h < a && (unsigned long) h + page_size < a) return r; WALK_TLIST(sl_head, h, &s->full_heads) if ((unsigned long) h < a && (unsigned long) h + page_size < a) return r; return NULL; } #endif