mirror of
https://github.com/cryb-to/cryb-to.git
synced 2024-12-25 05:51:07 +00:00
8e0f4a293e
When cryb-test is used as a framework for another project, the compile-time test is useless since cryb-test itself will have been built with coverage disabled. Besides, it is not a reliable indicator of whether leak detection will work. Instead, check if the heap is already dirty when we first gain control.
575 lines
14 KiB
C
575 lines
14 KiB
C
/*
|
|
* Copyright (c) 2014-2017 Dag-Erling Smørgrav
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. The name of the author may not be used to endorse or promote
|
|
* products derived from this software without specific prior written
|
|
* permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*/
|
|
|
|
#include "cryb/impl.h"
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/mman.h>
|
|
|
|
#ifndef MAP_NOCORE
|
|
#define MAP_NOCORE 0
|
|
#endif
|
|
|
|
#ifndef MAP_NOSYNC
|
|
#define MAP_NOSYNC 0
|
|
#endif
|
|
|
|
#if HAVE_UTRACE
|
|
#if HAVE_SYS_KTRACE_H
|
|
#if HAVE_SYS_UIO_H
|
|
#include <sys/uio.h>
|
|
#endif
|
|
#include <sys/ktrace.h>
|
|
#endif
|
|
#endif
|
|
|
|
#include <errno.h>
|
|
#include <stdint.h>
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
|
|
#include <cryb/assert.h>
|
|
#include <cryb/test.h>
|
|
|
|
/*
|
|
* Very simple, non-thread-safe malloc() implementation tailored for unit
|
|
* tests. The most important feature of this implementation is the
|
|
* t_malloc_fail flag, which can be used to force malloc(), calloc() and
|
|
* realloc() calls to fail. It also emits jemalloc-compatible trace
|
|
* records on platforms that have utrace(2).
|
|
*
|
|
* Allocations are satisfied either from a bucket or by direct mapping.
|
|
* The allocation size is first rounded to the nearest power of two or 16,
|
|
* whichever is largest. If this number is larger than the maximum bucket
|
|
* block size, a direct mapping is used. Otherwise, the appropriate
|
|
* bucket is selected and the first free block from that bucket is
|
|
* returned. If there are no free blocks in the bucket, the allocation
|
|
* fails.
|
|
*
|
|
* Bucket metadata is stored in a static array; the buckets themselves are
|
|
* allocated using mmap(). The free list is maintained by keeping a
|
|
* pointer to the first free block in the bucket metadata, and storing a
|
|
* pointer to the next free block at the start of each free block. These
|
|
* pointers are not preinitialized, which avoid faulting in large amounts
|
|
* of memory that will never be used.
|
|
*
|
|
* Direct mappings are allocated using mmap(). Metadata for each mapping
|
|
* is stored in a malloc()ed struct in a linked list.
|
|
*
|
|
* Attempts to allocate 0 bytes return a pointer to address space which is
|
|
* mmap()ed with PROT_NONE, causing any attempt to use it to fail.
|
|
*
|
|
* The data structures used are arrays and linked lists, which would be
|
|
* unacceptably inefficient for production use but are good enough for
|
|
* testing.
|
|
*/
|
|
|
|
#define PADD(p, c) (void *)((intptr_t)(p) + (size_t)(c))
|
|
#define PSUB(p, c) (void *)((intptr_t)(p) + (size_t)(c))
|
|
#define PDIFF(p1, p2) (size_t)((char *)(p1) - (char *)(p2))
|
|
|
|
/* base 2 logarithm of the minimum and maximum block sizes */
|
|
#define BUCKET_MIN_SHIFT 4
|
|
#define BUCKET_MAX_SHIFT 16
|
|
|
|
/* bucket size */
|
|
#define BUCKET_SIZE (16*1024*1024)
|
|
|
|
/* byte values used to fill allocated and unallocated blocks */
|
|
#define BUCKET_FILL_ALLOC 0xaa
|
|
#define BUCKET_FILL_FREE 0x55
|
|
|
|
struct bucket {
|
|
void *base; /* bottom of bucket */
|
|
void *top; /* top of bucket */
|
|
void *free; /* first free block */
|
|
void *unused; /* first never-used block */
|
|
unsigned long nalloc;
|
|
unsigned long nfree;
|
|
};
|
|
|
|
struct mapping {
|
|
void *base; /* base address */
|
|
void *top; /* end address */
|
|
struct mapping *prev, *next; /* linked list */
|
|
};
|
|
|
|
/* bucket metadata */
|
|
static struct bucket buckets[BUCKET_MAX_SHIFT + 1];
|
|
|
|
/* mapping metadata */
|
|
static struct mapping *mappings;
|
|
static unsigned long nmapalloc, nmapfree;
|
|
|
|
/* if non-zero, all allocations fail */
|
|
int t_malloc_fail;
|
|
|
|
/* if non-zero, all allocations will fail after a countdown */
|
|
int t_malloc_fail_after;
|
|
|
|
/* if non-zero, unintentional allocation failures are fatal */
|
|
int t_malloc_fatal;
|
|
|
|
#if HAVE_UTRACE
|
|
|
|
/*
|
|
* Record malloc() / realloc() / free() events
|
|
*/
|
|
static void
|
|
trace_malloc_event(const void *o, size_t s, const void *p)
|
|
{
|
|
struct { const void *o; size_t s; const void *p; } mu = { o, s, p };
|
|
int serrno = errno;
|
|
|
|
(void)utrace(&mu, sizeof mu);
|
|
errno = serrno;
|
|
}
|
|
|
|
#define UTRACE_MALLOC(s, p) \
|
|
trace_malloc_event(NULL, (s), (p))
|
|
#define UTRACE_REALLOC(o, s, p) \
|
|
trace_malloc_event((o), (s), (p))
|
|
#define UTRACE_FREE(o) \
|
|
trace_malloc_event((o), 0, NULL)
|
|
|
|
#else
|
|
|
|
#define UTRACE_MALLOC(s, p) \
|
|
do { (void)(s); (void)(p); } while (0)
|
|
#define UTRACE_REALLOC(o, s, p) \
|
|
do { (void)(o); (void)(s); (void)(p); } while (0)
|
|
#define UTRACE_FREE(o) \
|
|
do { (void)(o); } while (0)
|
|
|
|
#endif
|
|
|
|
|
|
/*
|
|
* Return a pointer to inaccessible memory.
|
|
*/
|
|
static void *
|
|
t_malloc_null(void)
|
|
{
|
|
struct bucket *b;
|
|
|
|
b = &buckets[0];
|
|
if (b->base == NULL) {
|
|
b->base = mmap(NULL, BUCKET_SIZE, PROT_NONE,
|
|
MAP_ANON | MAP_NOCORE | MAP_NOSYNC | MAP_SHARED, -1, 0);
|
|
if (b->base == MAP_FAILED)
|
|
abort();
|
|
b->top = b->base + BUCKET_SIZE;
|
|
b->free = b->unused = b->base;
|
|
}
|
|
++b->nalloc;
|
|
return (b->base);
|
|
}
|
|
|
|
/*
|
|
* Allocate a direct mapping. Round up the size to the nearest multiple
|
|
* of 8192, call mmap() with the correct arguments, and verify the result.
|
|
*/
|
|
static void *
|
|
t_malloc_mapped(size_t size)
|
|
{
|
|
struct mapping *m;
|
|
size_t msize;
|
|
|
|
/* prepare metadata */
|
|
if ((m = malloc(sizeof *m)) == NULL)
|
|
return (NULL);
|
|
msize = ((size + 8191) >> 13) << 13;
|
|
|
|
/* map a sufficiently large region */
|
|
m->base = mmap(NULL, msize, PROT_READ | PROT_WRITE,
|
|
MAP_ANON | MAP_NOSYNC | MAP_SHARED, -1, 0);
|
|
if (m->base == MAP_FAILED) {
|
|
free(m);
|
|
errno = ENOMEM;
|
|
return (NULL);
|
|
}
|
|
m->top = PADD(m->base, msize);
|
|
|
|
/* insert into linked list */
|
|
m->next = mappings;
|
|
m->prev = NULL;
|
|
mappings = m;
|
|
|
|
/* fill the slop */
|
|
if (msize > size)
|
|
memset(PADD(m->base, size), BUCKET_FILL_FREE, msize - size);
|
|
|
|
/* done! */
|
|
++nmapalloc;
|
|
return (m->base);
|
|
}
|
|
|
|
/*
|
|
* Allocate from a bucket. Round up the size to the nearest power of two,
|
|
* select the appropriate bucket, and return the first free or unused
|
|
* block.
|
|
*/
|
|
static void *
|
|
t_malloc_bucket(size_t size)
|
|
{
|
|
unsigned int shift;
|
|
struct bucket *b;
|
|
size_t msize;
|
|
void *p;
|
|
|
|
/* select bucket */
|
|
for (shift = BUCKET_MIN_SHIFT; (1U << shift) < size; ++shift)
|
|
/* nothing */ ;
|
|
assert(shift >= BUCKET_MIN_SHIFT && shift <= BUCKET_MAX_SHIFT);
|
|
b = &buckets[shift];
|
|
msize = 1U << shift;
|
|
|
|
/* initialize bucket if necessary */
|
|
if (b->base == NULL) {
|
|
b->base = mmap(NULL, BUCKET_SIZE, PROT_READ | PROT_WRITE,
|
|
MAP_ANON | MAP_NOSYNC | MAP_SHARED, -1, 0);
|
|
if (b->base == MAP_FAILED)
|
|
abort();
|
|
b->top = b->base + BUCKET_SIZE;
|
|
b->free = b->unused = b->base;
|
|
}
|
|
|
|
/* the bucket is full */
|
|
if (b->free == b->top) {
|
|
errno = ENOMEM;
|
|
return (NULL);
|
|
}
|
|
|
|
/* we will return the first free block */
|
|
p = b->free;
|
|
|
|
/* update the free block pointer */
|
|
if (b->free == b->unused) {
|
|
/* never been used before, increment free pointer */
|
|
b->free = b->unused = b->unused + msize;
|
|
} else {
|
|
/* previously used, disconnect from free list */
|
|
b->free = *(char **)p;
|
|
assert(b->free >= b->base && b->free < b->top);
|
|
}
|
|
|
|
/* fill the slop */
|
|
if (msize > size)
|
|
memset(PADD(p, size), BUCKET_FILL_FREE, msize - size);
|
|
|
|
/* done! */
|
|
++b->nalloc;
|
|
return (p);
|
|
}
|
|
|
|
/*
|
|
* Core malloc() logic: select the correct backend based on the requested
|
|
* allocation size and call it.
|
|
*/
|
|
static void *
|
|
t_malloc(size_t size)
|
|
{
|
|
|
|
/* select and call the right backend */
|
|
if (size == 0)
|
|
return (t_malloc_null());
|
|
else if (size > (1U << BUCKET_MAX_SHIFT))
|
|
return (t_malloc_mapped(size));
|
|
else
|
|
return (t_malloc_bucket(size));
|
|
}
|
|
|
|
/*
|
|
* Allocate an object of the requested size. According to the standard,
|
|
* the content of the allocated memory is undefined; we fill it with
|
|
* easily recognizable garbage.
|
|
*/
|
|
void *
|
|
malloc(size_t size)
|
|
{
|
|
void *p;
|
|
|
|
if (t_malloc_fail) {
|
|
errno = ENOMEM;
|
|
return (NULL);
|
|
} else if (t_malloc_fail_after > 0 && --t_malloc_fail_after == 0) {
|
|
t_malloc_fail = 1;
|
|
}
|
|
p = t_malloc(size);
|
|
UTRACE_MALLOC(size, p);
|
|
if (p == NULL) {
|
|
if (t_malloc_fatal)
|
|
abort();
|
|
return (NULL);
|
|
}
|
|
memset(p, BUCKET_FILL_ALLOC, size);
|
|
return (p);
|
|
}
|
|
|
|
/*
|
|
* Allocate an array of n objects of the requested size and initialize it
|
|
* to zero.
|
|
*/
|
|
void *
|
|
calloc(size_t n, size_t size)
|
|
{
|
|
void *p;
|
|
|
|
if (t_malloc_fail) {
|
|
errno = ENOMEM;
|
|
return (NULL);
|
|
} else if (t_malloc_fail_after > 0 && --t_malloc_fail_after == 0) {
|
|
t_malloc_fail = 1;
|
|
}
|
|
p = t_malloc(n * size);
|
|
UTRACE_MALLOC(n * size, p);
|
|
if (p == NULL) {
|
|
if (t_malloc_fatal)
|
|
abort();
|
|
return (NULL);
|
|
}
|
|
memset(p, 0, n * size);
|
|
return (p);
|
|
}
|
|
|
|
/*
|
|
* Grow or shrink an allocated object, preserving its contents up to the
|
|
* smaller of the object's original and new size. According to the
|
|
* standard, the object may be either grown or shrunk in place or replaced
|
|
* with a new one. We always allocate a new object and free the old one.
|
|
*/
|
|
void *
|
|
realloc(void *o, size_t size)
|
|
{
|
|
struct mapping *m;
|
|
struct bucket *b;
|
|
void *p;
|
|
size_t osize;
|
|
unsigned int shift;
|
|
|
|
/* corner cases */
|
|
if (o == NULL || o == buckets[0].base)
|
|
return (malloc(size));
|
|
|
|
/* was this a direct mapping? */
|
|
for (m = mappings; m != NULL; m = m->next) {
|
|
if (o == m->base) {
|
|
/* found our mapping */
|
|
osize = PDIFF(m->top, m->base);
|
|
goto found;
|
|
}
|
|
assert(o < m->base || o >= m->top);
|
|
}
|
|
|
|
/* was this a bucket allocation? */
|
|
for (shift = BUCKET_MIN_SHIFT; shift <= BUCKET_MAX_SHIFT; ++shift) {
|
|
b = &buckets[shift];
|
|
if (o >= b->base && o < b->top) {
|
|
/* found our bucket */
|
|
assert(PDIFF(o, b->base) % (1U << shift) == 0);
|
|
osize = 1U << shift;
|
|
goto found;
|
|
}
|
|
}
|
|
|
|
/* oops */
|
|
abort();
|
|
|
|
found:
|
|
if (t_malloc_fail) {
|
|
errno = ENOMEM;
|
|
return (NULL);
|
|
} else if (t_malloc_fail_after > 0 && --t_malloc_fail_after == 0) {
|
|
t_malloc_fail = 1;
|
|
}
|
|
p = t_malloc(size);
|
|
UTRACE_REALLOC(o, size, p);
|
|
if (p == NULL) {
|
|
if (t_malloc_fatal)
|
|
abort();
|
|
return (NULL);
|
|
}
|
|
if (size > osize) {
|
|
memcpy(p, o, osize);
|
|
memset(p + osize, BUCKET_FILL_ALLOC, size - osize);
|
|
} else {
|
|
memcpy(p, o, size);
|
|
}
|
|
free(o);
|
|
return (p);
|
|
}
|
|
|
|
/*
|
|
* Free an allocated object. According to the standard, the content of
|
|
* the memory previously occupied by the object is undefined. We fill it
|
|
* with easily recognizable garbage to facilitate debugging use-after-free
|
|
* bugs.
|
|
*/
|
|
void
|
|
free(void *p)
|
|
{
|
|
struct mapping *m;
|
|
struct bucket *b;
|
|
unsigned int shift;
|
|
|
|
UTRACE_FREE(p);
|
|
|
|
/* free(NULL) */
|
|
if (p == NULL)
|
|
return;
|
|
|
|
/* was this a zero-size allocation? */
|
|
if (p == buckets[0].base) {
|
|
++buckets[0].nfree;
|
|
return;
|
|
}
|
|
|
|
/* was this a direct mapping? */
|
|
for (m = mappings; m != NULL; m = m->next) {
|
|
if (p == m->base) {
|
|
/* found our mapping */
|
|
if (munmap(m->base, PDIFF(m->top, m->base)) != 0)
|
|
abort();
|
|
if (m->prev != NULL)
|
|
m->prev->next = m->next;
|
|
if (m->next != NULL)
|
|
m->next->prev = m->prev;
|
|
if (m == mappings)
|
|
mappings = m->next;
|
|
/* fall through and free metadata */
|
|
UTRACE_FREE(p = m);
|
|
++nmapfree;
|
|
break;
|
|
}
|
|
assert(p < m->base || p >= m->top);
|
|
}
|
|
|
|
/* was this a bucket allocation? */
|
|
for (shift = BUCKET_MIN_SHIFT; shift <= BUCKET_MAX_SHIFT; ++shift) {
|
|
b = &buckets[shift];
|
|
if (p >= b->base && p < b->top) {
|
|
/* found our bucket */
|
|
assert(PDIFF(p, b->base) % (1U << shift) == 0);
|
|
memset(p, BUCKET_FILL_FREE, 1U << shift);
|
|
/* connect the block to the free list */
|
|
*(char **)p = b->free;
|
|
b->free = p;
|
|
++b->nfree;
|
|
return;
|
|
}
|
|
}
|
|
|
|
/* oops */
|
|
abort();
|
|
}
|
|
|
|
/*
|
|
* Return a snapshot of the allocator state
|
|
*/
|
|
size_t
|
|
t_malloc_snapshot(void *buf, size_t len)
|
|
{
|
|
unsigned long snapshot[BUCKET_MAX_SHIFT];
|
|
unsigned int i;
|
|
|
|
if (buf == NULL)
|
|
return (sizeof snapshot);
|
|
snapshot[0] = nmapalloc - nmapfree;
|
|
for (i = 2; i < BUCKET_MIN_SHIFT; ++i)
|
|
snapshot[i - 1] = 0;
|
|
for (i = BUCKET_MIN_SHIFT; i <= BUCKET_MAX_SHIFT; ++i)
|
|
snapshot[i - 1] = buckets[i].nalloc - buckets[i].nfree;
|
|
if (len > sizeof snapshot)
|
|
len = sizeof snapshot;
|
|
memcpy(buf, snapshot, len);
|
|
return (sizeof snapshot);
|
|
}
|
|
|
|
/*
|
|
* Print allocator statistics
|
|
*/
|
|
void
|
|
t_malloc_printstats(FILE *f)
|
|
{
|
|
struct bucket *b;
|
|
unsigned int shift;
|
|
|
|
fprintf(f, "%6s %9s %9s %9s\n", "bucket", "alloc", "free", "leaked");
|
|
for (shift = BUCKET_MIN_SHIFT; shift <= BUCKET_MAX_SHIFT; ++shift) {
|
|
b = &buckets[shift];
|
|
if (b->nalloc > 0)
|
|
fprintf(f, " 2^%-3u %9lu %9lu %9lu\n",
|
|
shift, b->nalloc, b->nfree,
|
|
b->nalloc - b->nfree);
|
|
}
|
|
if (nmapalloc > 0)
|
|
fprintf(f, "%6s %9lu %9lu %9lu\n", "mapped",
|
|
nmapalloc, nmapfree, nmapalloc - nmapfree);
|
|
}
|
|
|
|
/*
|
|
* Return number of outstanding allocations
|
|
*/
|
|
unsigned long
|
|
t_malloc_outstanding(void)
|
|
{
|
|
struct bucket *b;
|
|
unsigned int shift;
|
|
unsigned long n;
|
|
|
|
n = nmapalloc - nmapfree;
|
|
for (shift = BUCKET_MIN_SHIFT; shift <= BUCKET_MAX_SHIFT; ++shift) {
|
|
b = &buckets[shift];
|
|
n += b->nalloc - b->nfree;
|
|
}
|
|
return (n);
|
|
}
|
|
|
|
/*
|
|
* Test that fails if we leaked memory
|
|
*/
|
|
static int
|
|
t_malloc_leaked(char **desc, void *arg CRYB_UNUSED)
|
|
{
|
|
unsigned long nleaked;
|
|
|
|
nleaked = t_malloc_outstanding();
|
|
if (nleaked > 0)
|
|
(void)asprintf(desc, "%lu allocation(s) leaked", nleaked);
|
|
else
|
|
(void)asprintf(desc, "%s", "no memory leaked");
|
|
return (nleaked == 0);
|
|
}
|
|
|
|
struct t_test t_memory_leak = {
|
|
.func = &t_malloc_leaked,
|
|
.arg = NULL,
|
|
.desc = "memory leak check",
|
|
};
|