blob: e3d63d66d7e29127a72940c316f769d3a961b91a [file] [log] [blame]
/*
* Copyright (C) 2008 Advanced Micro Devices, Inc.
* Copyright (C) 2008-2010 coresystems GmbH
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* 3. The name of the author may not be used to endorse or promote products
* derived from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*/
/*
* This is a classically weak malloc() implementation. We have a relatively
* small and static heap, so we take the easy route with an O(N) loop
* through the tree for every malloc() and free(). Obviously, this doesn't
* scale past a few hundred KB (if that).
*
* We're also susceptible to the usual buffer overrun poisoning, though the
* risk is within acceptable ranges for this implementation (don't overrun
* your buffers, kids!).
*/
#define IN_MALLOC_C
#include <stdint.h>
#include <stdio.h>
#include <stdlib.h>
struct memory_type {
void *start;
void *end;
struct align_region_t* align_regions;
#if CONFIG_DEBUG_MALLOC
int magic_initialized;
size_t minimal_free;
const char *name;
#endif
};
static uint8_t heap_buffer[CONFIG_HEAP_SIZE] __attribute__((aligned(16)));
static struct memory_type default_type =
{ (void *)&heap_buffer[0], (void *)&heap_buffer[CONFIG_HEAP_SIZE], NULL
#if CONFIG_DEBUG_MALLOC
, 0, 0, "HEAP"
#endif
};
static struct memory_type *const heap = &default_type;
static struct memory_type *dma = &default_type;
typedef uint64_t hdrtype_t;
#define HDRSIZE (sizeof(hdrtype_t))
#define SIZE_BITS ((HDRSIZE << 3) - 7)
#define MAGIC (((hdrtype_t)0x2a) << (SIZE_BITS + 1))
#define FLAG_FREE (((hdrtype_t)0x01) << (SIZE_BITS + 0))
#define MAX_SIZE ((((hdrtype_t)0x01) << SIZE_BITS) - 1)
#define SIZE(_h) ((_h) & MAX_SIZE)
#define _HEADER(_s, _f) ((hdrtype_t) (MAGIC | (_f) | ((_s) & MAX_SIZE)))
#define FREE_BLOCK(_s) _HEADER(_s, FLAG_FREE)
#define USED_BLOCK(_s) _HEADER(_s, 0)
#define IS_FREE(_h) (((_h) & (MAGIC | FLAG_FREE)) == (MAGIC | FLAG_FREE))
#define HAS_MAGIC(_h) (((_h) & MAGIC) == MAGIC)
static int free_aligned(void* addr, struct memory_type *type);
void print_malloc_map(void);
void init_dma_memory(void *start, uint32_t size)
{
if (dma_initialized()) {
printf("ERROR: %s called twice!\n", __func__);
return;
}
/*
* DMA memory might not be zeroed by Coreboot on stage loading, so make
* sure we clear the magic cookie from last boot.
*/
*(hdrtype_t *)start = 0;
dma = malloc(sizeof(*dma));
dma->start = start;
dma->end = start + size;
dma->align_regions = NULL;
#if CONFIG_DEBUG_MALLOC
dma->minimal_free = 0;
dma->magic_initialized = 0;
dma->name = "DMA";
printf("Initialized cache-coherent DMA memory at [%p:%p]\n", start, start + size);
#endif
}
int dma_initialized()
{
return dma != heap;
}
/* For boards that don't initialize DMA we assume all locations are coherent */
int dma_coherent(void *ptr)
{
return !dma_initialized() || (dma->start <= ptr && dma->end > ptr);
}
static void *alloc(int len, struct memory_type *type)
{
hdrtype_t header;
hdrtype_t volatile *ptr = (hdrtype_t volatile *)type->start;
/* Align the size. */
len = ALIGN_UP(len, HDRSIZE);
if (!len || len > MAX_SIZE)
return (void *)NULL;
/* Make sure the region is setup correctly. */
if (!HAS_MAGIC(*ptr)) {
size_t size = (type->end - type->start) - HDRSIZE;
*ptr = FREE_BLOCK(size);
#if CONFIG_DEBUG_MALLOC
type->magic_initialized = 1;
type->minimal_free = size;
#endif
}
/* Find some free space. */
do {
header = *ptr;
int size = SIZE(header);
if (!HAS_MAGIC(header) || size == 0) {
printf("memory allocator panic. (%s%s)\n",
!HAS_MAGIC(header) ? " no magic " : "",
size == 0 ? " size=0 " : "");
halt();
}
if (header & FLAG_FREE) {
if (len <= size) {
hdrtype_t volatile *nptr = (hdrtype_t volatile *)((uintptr_t)ptr + HDRSIZE + len);
int nsize = size - (HDRSIZE + len);
/* If there is still room in this block,
* then mark it as such otherwise account
* the whole space for that block.
*/
if (nsize > 0) {
/* Mark the block as used. */
*ptr = USED_BLOCK(len);
/* Create a new free block. */
*nptr = FREE_BLOCK(nsize);
} else {
/* Mark the block as used. */
*ptr = USED_BLOCK(size);
}
return (void *)((uintptr_t)ptr + HDRSIZE);
}
}
ptr = (hdrtype_t volatile *)((uintptr_t)ptr + HDRSIZE + size);
} while (ptr < (hdrtype_t *) type->end);
/* Nothing available. */
return (void *)NULL;
}
static void _consolidate(struct memory_type *type)
{
void *ptr = type->start;
while (ptr < type->end) {
void *nptr;
hdrtype_t hdr = *((hdrtype_t *) ptr);
unsigned int size = 0;
if (!IS_FREE(hdr)) {
ptr += HDRSIZE + SIZE(hdr);
continue;
}
size = SIZE(hdr);
nptr = ptr + HDRSIZE + SIZE(hdr);
while (nptr < type->end) {
hdrtype_t nhdr = *((hdrtype_t *) nptr);
if (!(IS_FREE(nhdr)))
break;
size += SIZE(nhdr) + HDRSIZE;
*((hdrtype_t *) nptr) = 0;
nptr += (HDRSIZE + SIZE(nhdr));
}
*((hdrtype_t *) ptr) = FREE_BLOCK(size);
ptr = nptr;
}
}
void free(void *ptr)
{
hdrtype_t hdr;
struct memory_type *type = heap;
/* Sanity check. */
if (ptr < type->start || ptr >= type->end) {
type = dma;
if (ptr < type->start || ptr >= type->end)
return;
}
if (free_aligned(ptr, type)) return;
ptr -= HDRSIZE;
hdr = *((hdrtype_t *) ptr);
/* Not our header (we're probably poisoned). */
if (!HAS_MAGIC(hdr))
return;
/* Double free. */
if (hdr & FLAG_FREE)
return;
*((hdrtype_t *) ptr) = FREE_BLOCK(SIZE(hdr));
_consolidate(type);
}
void *malloc(size_t size)
{
return alloc(size, heap);
}
void *dma_malloc(size_t size)
{
return alloc(size, dma);
}
void *calloc(size_t nmemb, size_t size)
{
size_t total = nmemb * size;
void *ptr = alloc(total, heap);
if (ptr)
memset(ptr, 0, total);
return ptr;
}
void *realloc(void *ptr, size_t size)
{
void *ret, *pptr;
unsigned int osize;
struct memory_type *type = heap;
if (ptr == NULL)
return alloc(size, type);
pptr = ptr - HDRSIZE;
if (!HAS_MAGIC(*((hdrtype_t *) pptr)))
return NULL;
if (ptr < type->start || ptr >= type->end)
type = dma;
/* Get the original size of the block. */
osize = SIZE(*((hdrtype_t *) pptr));
/*
* Free the memory to update the tables - this won't touch the actual
* memory, so we can still use it for the copy after we have
* reallocated the new space.
*/
free(ptr);
ret = alloc(size, type);
/*
* if ret == NULL, then doh - failure.
* if ret == ptr then woo-hoo! no copy needed.
*/
if (ret == NULL || ret == ptr)
return ret;
/* Copy the memory to the new location. */
memcpy(ret, ptr, osize > size ? size : osize);
return ret;
}
struct align_region_t
{
/* If alignment is 0 then the region reqpresents a large region which
* has no metadata for tracking subelements. */
int alignment;
/* start in memory, and size in bytes */
void* start;
int size;
/* layout within a region:
- num_elements bytes, 0: free, 1: used, 2: used, combines with next
- padding to alignment
- data section
- waste space
start_data points to the start of the data section
*/
void* start_data;
/* number of free blocks sized "alignment" */
int free;
struct align_region_t *next;
};
static inline int region_is_large(const struct align_region_t *r)
{
return r->alignment == 0;
}
static inline int addr_in_region(const struct align_region_t *r, void *addr)
{
return ((addr >= r->start_data) && (addr < r->start_data + r->size));
}
/* num_elements == 0 indicates a large aligned region instead of a smaller
* region comprised of alignment-sized chunks. */
static struct align_region_t *allocate_region(int alignment, int num_elements,
size_t size, struct memory_type *type)
{
struct align_region_t *r;
size_t extra_space;
#if CONFIG_DEBUG_MALLOC
printf("%s(old align_regions=%p, alignment=%u, num_elements=%u, size=%zu)\n",
__func__, type->align_regions, alignment, num_elements, size);
#endif
r = malloc(sizeof(*r));
if (r == NULL)
return NULL;
memset(r, 0, sizeof(r));
if (num_elements != 0) {
r->alignment = alignment;
r->size = num_elements * alignment;
r->free = num_elements;
/* Allocate enough memory for alignment requirements and
* metadata for each chunk. */
extra_space = num_elements;
} else {
/* Large aligned allocation. Set alignment = 0. */
r->alignment = 0;
r->size = size;
extra_space = 0;
}
r->start = alloc(r->size + alignment + extra_space, type);
if (r->start == NULL) {
free(r);
return NULL;
}
r->start_data = (void *)ALIGN_UP((uintptr_t)r->start + extra_space,
alignment);
/* Clear any (if requested) metadata. */
memset(r->start, 0, extra_space);
/* Link the region with the rest. */
r->next = type->align_regions;
type->align_regions = r;
return r;
}
static void try_free_region(struct align_region_t **prev_link)
{
struct align_region_t *r = *prev_link;
/* All large regions are immediately free-able. Non-large regions
* need to be checked for the fully freed state. */
if (!region_is_large(r)) {
if (r->free != r->size / r->alignment)
return;
}
/* Unlink region from link list. */
*prev_link = r->next;
/* Free the data and metadata. */
free(r->start);
free(r);
}
static int free_aligned(void* addr, struct memory_type *type)
{
struct align_region_t **prev_link = &type->align_regions;
while (*prev_link != NULL)
{
if (!addr_in_region(*prev_link, addr)) {
prev_link = &((*prev_link)->next);
continue;
}
if (region_is_large(*prev_link)) {
try_free_region(prev_link);
return 1;
}
int i = (addr-(*prev_link)->start_data)/(*prev_link)->alignment;
uint8_t *meta = (*prev_link)->start;
while (meta[i] == 2)
{
meta[i++] = 0;
(*prev_link)->free++;
}
meta[i] = 0;
(*prev_link)->free++;
try_free_region(prev_link);
return 1;
}
return 0;
}
static void *alloc_aligned(size_t align, size_t size, struct memory_type *type)
{
/* Define a large request to be 1024 bytes for either alignment or
* size of allocation. */
const size_t large_request = 1024;
if (size == 0) return 0;
if (type->align_regions == 0) {
type->align_regions = malloc(sizeof(struct align_region_t));
if (type->align_regions == NULL)
return NULL;
memset(type->align_regions, 0, sizeof(struct align_region_t));
}
struct align_region_t *reg = type->align_regions;
if (size >= large_request || align >= large_request) {
reg = allocate_region(align, 0, size, type);
if (reg == NULL)
return NULL;
return reg->start_data;
}
look_further:
while (reg != 0)
{
if ((reg->alignment == align) && (reg->free >= (size + align - 1)/align))
{
#if CONFIG_DEBUG_MALLOC
printf(" found memalign region. %x free, %x required\n", reg->free, (size + align - 1)/align);
#endif
break;
}
reg = reg->next;
}
if (reg == 0)
{
#if CONFIG_DEBUG_MALLOC
printf(" need to allocate a new memalign region\n");
#endif
/* get align regions */
reg = allocate_region(align, large_request/align, size, type);
#if CONFIG_DEBUG_MALLOC
printf(" ... returned %p\n", reg);
#endif
}
if (reg == 0) {
/* Nothing available. */
return (void *)NULL;
}
int i, count = 0, target = (size+align-1)/align;
for (i = 0; i < (reg->size/align); i++)
{
if (((uint8_t *)reg->start)[i] == 0)
{
count++;
if (count == target) {
count = i+1-count;
for (i=0; i<target-1; i++)
{
((uint8_t *)reg->start)[count + i] = 2;
}
((uint8_t *)reg->start)[count + target - 1] = 1;
reg->free -= target;
return reg->start_data+(align*count);
}
} else {
count = 0;
}
}
/* The free space in this region is fragmented,
so we will move on and try the next one: */
reg = reg->next;
goto look_further; // end condition is once a new region is allocated - it always has enough space
}
void *memalign(size_t align, size_t size)
{
return alloc_aligned(align, size, heap);
}
void *dma_memalign(size_t align, size_t size)
{
return alloc_aligned(align, size, dma);
}
/* This is for debugging purposes. */
#if CONFIG_DEBUG_MALLOC
void print_malloc_map(void)
{
struct memory_type *type = heap;
void *ptr;
int free_memory;
again:
ptr = type->start;
free_memory = 0;
while (ptr < type->end) {
hdrtype_t hdr = *((hdrtype_t *) ptr);
if (!HAS_MAGIC(hdr)) {
if (type->magic_initialized)
printf("%s: Poisoned magic - we're toast\n", type->name);
else
printf("%s: No magic yet - going to initialize\n", type->name);
break;
}
/* FIXME: Verify the size of the block. */
printf("%s %x: %s (%x bytes)\n", type->name,
(unsigned int)(ptr - type->start),
hdr & FLAG_FREE ? "FREE" : "USED", SIZE(hdr));
if (hdr & FLAG_FREE)
free_memory += SIZE(hdr);
ptr += HDRSIZE + SIZE(hdr);
}
if (free_memory && (type->minimal_free > free_memory))
type->minimal_free = free_memory;
printf("%s: Maximum memory consumption: %u bytes\n", type->name,
(type->end - type->start) - HDRSIZE - type->minimal_free);
if (type != dma) {
type = dma;
goto again;
}
}
#endif