mirror of
				https://git.proxmox.com/git/mirror_frr
				synced 2025-11-04 06:38:04 +00:00 
			
		
		
		
	it is handy to be able to see allocated bitfields in the show memory output. Signed-off-by: Pat Ruddy <pat@voltanet.io>
		
			
				
	
	
		
			193 lines
		
	
	
		
			4.9 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			193 lines
		
	
	
		
			4.9 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * Copyright (c) 2015-16  David Lamparter, for NetDEF, Inc.
 | 
						|
 *
 | 
						|
 * Permission to use, copy, modify, and distribute this software for any
 | 
						|
 * purpose with or without fee is hereby granted, provided that the above
 | 
						|
 * copyright notice and this permission notice appear in all copies.
 | 
						|
 *
 | 
						|
 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
 | 
						|
 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
 | 
						|
 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
 | 
						|
 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
 | 
						|
 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
 | 
						|
 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
 | 
						|
 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
 | 
						|
 */
 | 
						|
 | 
						|
#include <zebra.h>
 | 
						|
 | 
						|
#include <stdlib.h>
 | 
						|
#ifdef HAVE_MALLOC_H
 | 
						|
#include <malloc.h>
 | 
						|
#endif
 | 
						|
#ifdef HAVE_MALLOC_NP_H
 | 
						|
#include <malloc_np.h>
 | 
						|
#endif
 | 
						|
#ifdef HAVE_MALLOC_MALLOC_H
 | 
						|
#include <malloc/malloc.h>
 | 
						|
#endif
 | 
						|
 | 
						|
#include "memory.h"
 | 
						|
#include "log.h"
 | 
						|
#include "libfrr_trace.h"
 | 
						|
 | 
						|
static struct memgroup *mg_first = NULL;
 | 
						|
struct memgroup **mg_insert = &mg_first;
 | 
						|
 | 
						|
DEFINE_MGROUP(LIB, "libfrr");
 | 
						|
DEFINE_MTYPE(LIB, TMP, "Temporary memory");
 | 
						|
DEFINE_MTYPE(LIB, BITFIELD, "Bitfield memory");
 | 
						|
 | 
						|
static inline void mt_count_alloc(struct memtype *mt, size_t size, void *ptr)
 | 
						|
{
 | 
						|
	size_t current;
 | 
						|
	size_t oldsize;
 | 
						|
 | 
						|
	current = 1 + atomic_fetch_add_explicit(&mt->n_alloc, 1,
 | 
						|
						memory_order_relaxed);
 | 
						|
 | 
						|
	oldsize = atomic_load_explicit(&mt->n_max, memory_order_relaxed);
 | 
						|
	if (current > oldsize)
 | 
						|
		/* note that this may fail, but approximation is sufficient */
 | 
						|
		atomic_compare_exchange_weak_explicit(&mt->n_max, &oldsize,
 | 
						|
						      current,
 | 
						|
						      memory_order_relaxed,
 | 
						|
						      memory_order_relaxed);
 | 
						|
 | 
						|
	oldsize = atomic_load_explicit(&mt->size, memory_order_relaxed);
 | 
						|
	if (oldsize == 0)
 | 
						|
		oldsize = atomic_exchange_explicit(&mt->size, size,
 | 
						|
						   memory_order_relaxed);
 | 
						|
	if (oldsize != 0 && oldsize != size && oldsize != SIZE_VAR)
 | 
						|
		atomic_store_explicit(&mt->size, SIZE_VAR,
 | 
						|
				      memory_order_relaxed);
 | 
						|
 | 
						|
#ifdef HAVE_MALLOC_USABLE_SIZE
 | 
						|
	size_t mallocsz = malloc_usable_size(ptr);
 | 
						|
 | 
						|
	current = mallocsz + atomic_fetch_add_explicit(&mt->total, mallocsz,
 | 
						|
						       memory_order_relaxed);
 | 
						|
	oldsize = atomic_load_explicit(&mt->max_size, memory_order_relaxed);
 | 
						|
	if (current > oldsize)
 | 
						|
		/* note that this may fail, but approximation is sufficient */
 | 
						|
		atomic_compare_exchange_weak_explicit(&mt->max_size, &oldsize,
 | 
						|
						      current,
 | 
						|
						      memory_order_relaxed,
 | 
						|
						      memory_order_relaxed);
 | 
						|
#endif
 | 
						|
}
 | 
						|
 | 
						|
static inline void mt_count_free(struct memtype *mt, void *ptr)
 | 
						|
{
 | 
						|
	frrtrace(2, frr_libfrr, memfree, mt, ptr);
 | 
						|
 | 
						|
	assert(mt->n_alloc);
 | 
						|
	atomic_fetch_sub_explicit(&mt->n_alloc, 1, memory_order_relaxed);
 | 
						|
 | 
						|
#ifdef HAVE_MALLOC_USABLE_SIZE
 | 
						|
	size_t mallocsz = malloc_usable_size(ptr);
 | 
						|
 | 
						|
	atomic_fetch_sub_explicit(&mt->total, mallocsz, memory_order_relaxed);
 | 
						|
#endif
 | 
						|
}
 | 
						|
 | 
						|
static inline void *mt_checkalloc(struct memtype *mt, void *ptr, size_t size)
 | 
						|
{
 | 
						|
	frrtrace(3, frr_libfrr, memalloc, mt, ptr, size);
 | 
						|
 | 
						|
	if (__builtin_expect(ptr == NULL, 0)) {
 | 
						|
		if (size) {
 | 
						|
			/* malloc(0) is allowed to return NULL */
 | 
						|
			memory_oom(size, mt->name);
 | 
						|
		}
 | 
						|
		return NULL;
 | 
						|
	}
 | 
						|
	mt_count_alloc(mt, size, ptr);
 | 
						|
	return ptr;
 | 
						|
}
 | 
						|
 | 
						|
void *qmalloc(struct memtype *mt, size_t size)
 | 
						|
{
 | 
						|
	return mt_checkalloc(mt, malloc(size), size);
 | 
						|
}
 | 
						|
 | 
						|
void *qcalloc(struct memtype *mt, size_t size)
 | 
						|
{
 | 
						|
	return mt_checkalloc(mt, calloc(size, 1), size);
 | 
						|
}
 | 
						|
 | 
						|
void *qrealloc(struct memtype *mt, void *ptr, size_t size)
 | 
						|
{
 | 
						|
	if (ptr)
 | 
						|
		mt_count_free(mt, ptr);
 | 
						|
	return mt_checkalloc(mt, ptr ? realloc(ptr, size) : malloc(size), size);
 | 
						|
}
 | 
						|
 | 
						|
void *qstrdup(struct memtype *mt, const char *str)
 | 
						|
{
 | 
						|
	return str ? mt_checkalloc(mt, strdup(str), strlen(str) + 1) : NULL;
 | 
						|
}
 | 
						|
 | 
						|
void qcountfree(struct memtype *mt, void *ptr)
 | 
						|
{
 | 
						|
	if (ptr)
 | 
						|
		mt_count_free(mt, ptr);
 | 
						|
}
 | 
						|
 | 
						|
void qfree(struct memtype *mt, void *ptr)
 | 
						|
{
 | 
						|
	if (ptr)
 | 
						|
		mt_count_free(mt, ptr);
 | 
						|
	free(ptr);
 | 
						|
}
 | 
						|
 | 
						|
int qmem_walk(qmem_walk_fn *func, void *arg)
 | 
						|
{
 | 
						|
	struct memgroup *mg;
 | 
						|
	struct memtype *mt;
 | 
						|
	int rv;
 | 
						|
 | 
						|
	for (mg = mg_first; mg; mg = mg->next) {
 | 
						|
		if ((rv = func(arg, mg, NULL)))
 | 
						|
			return rv;
 | 
						|
		for (mt = mg->types; mt; mt = mt->next)
 | 
						|
			if ((rv = func(arg, mg, mt)))
 | 
						|
				return rv;
 | 
						|
	}
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
struct exit_dump_args {
 | 
						|
	FILE *fp;
 | 
						|
	const char *prefix;
 | 
						|
	int error;
 | 
						|
};
 | 
						|
 | 
						|
static int qmem_exit_walker(void *arg, struct memgroup *mg, struct memtype *mt)
 | 
						|
{
 | 
						|
	struct exit_dump_args *eda = arg;
 | 
						|
 | 
						|
	if (!mt) {
 | 
						|
		fprintf(eda->fp,
 | 
						|
			"%s: showing active allocations in memory group %s\n",
 | 
						|
			eda->prefix, mg->name);
 | 
						|
 | 
						|
	} else if (mt->n_alloc) {
 | 
						|
		char size[32];
 | 
						|
		if (!mg->active_at_exit)
 | 
						|
			eda->error++;
 | 
						|
		snprintf(size, sizeof(size), "%10zu", mt->size);
 | 
						|
		fprintf(eda->fp, "%s: memstats:  %-30s: %6zu * %s\n",
 | 
						|
			eda->prefix, mt->name, mt->n_alloc,
 | 
						|
			mt->size == SIZE_VAR ? "(variably sized)" : size);
 | 
						|
	}
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
int log_memstats(FILE *fp, const char *prefix)
 | 
						|
{
 | 
						|
	struct exit_dump_args eda = {.fp = fp, .prefix = prefix, .error = 0};
 | 
						|
	qmem_walk(qmem_exit_walker, &eda);
 | 
						|
	return eda.error;
 | 
						|
}
 |