/* * malloc.c * * Very simple linked-list based malloc()/free(). * * Simplified even further by Alastair M. Robinson for TG68 project. * */ #ifdef __cplusplus extern "C" { #endif #include "stdio.h" #include "malloc.h" /* Both the arena list and the free memory list are double linked list with head node. This the head node. Note that the arena list is sorted in order of address. */ static struct free_arena_header __malloc_head = { { ARENA_TYPE_HEAD, 0, &__malloc_head, &__malloc_head, }, &__malloc_head, &__malloc_head }; static inline void mark_block_dead(struct free_arena_header *ah) { #ifdef DEBUG_MALLOC ah->a.type = ARENA_TYPE_DEAD; #endif } static inline void remove_from_main_chain(struct free_arena_header *ah) { struct free_arena_header *ap, *an; mark_block_dead(ah); ap = ah->a.prev; an = ah->a.next; ap->a.next = an; an->a.prev = ap; } static inline void remove_from_free_chain(struct free_arena_header *ah) { struct free_arena_header *ap, *an; ap = ah->prev_free; an = ah->next_free; ap->next_free = an; an->prev_free = ap; } static inline void remove_from_chains(struct free_arena_header *ah) { remove_from_free_chain(ah); remove_from_main_chain(ah); } void *__malloc_from_block(struct free_arena_header *fp, size_t size) { size_t fsize; struct free_arena_header *nfp, *na, *fpn, *fpp; fsize = fp->a.size; /* We need the 2* to account for the larger requirements of a free block */ if (fsize >= (size + 2 * sizeof(struct arena_header))) { /* Bigger block than required -- split block */ nfp = (struct free_arena_header *)((char *)fp + size); na = fp->a.next; nfp->a.type = ARENA_TYPE_FREE; nfp->a.size = fsize - size; fp->a.type = ARENA_TYPE_USED; fp->a.size = size; /* Insert into all-block chain */ nfp->a.prev = fp; nfp->a.next = na; na->a.prev = nfp; fp->a.next = nfp; /* Replace current block on free chain */ nfp->next_free = fpn = fp->next_free; nfp->prev_free = fpp = fp->prev_free; fpn->prev_free = nfp; fpp->next_free = nfp; } else { fp->a.type = ARENA_TYPE_USED; /* Allocate the whole block */ remove_from_free_chain(fp); } return (void *)(&fp->a + 1); } static struct free_arena_header *__free_block(struct free_arena_header *ah) { struct free_arena_header *pah, *nah; pah = ah->a.prev; nah = ah->a.next; if (pah->a.type == ARENA_TYPE_FREE && (char *)pah + pah->a.size == (char *)ah) { /* Coalesce into the previous block */ pah->a.size += ah->a.size; pah->a.next = nah; nah->a.prev = pah; mark_block_dead(ah); ah = pah; pah = ah->a.prev; } else { /* Need to add this block to the free chain */ ah->a.type = ARENA_TYPE_FREE; ah->next_free = __malloc_head.next_free; ah->prev_free = &__malloc_head; __malloc_head.next_free = ah; ah->next_free->prev_free = ah; } /* In either of the previous cases, we might be able to merge with the subsequent block... */ if (nah->a.type == ARENA_TYPE_FREE && (char *)ah + ah->a.size == (char *)nah) { ah->a.size += nah->a.size; /* Remove the old block from the chains */ remove_from_chains(nah); } /* Return the block that contains the called block */ return ah; } void malloc_add(void *p,size_t size) { struct free_arena_header *fp; struct free_arena_header *pah; fp=(struct free_arena_header *)p; fp->a.type = ARENA_TYPE_FREE; fp->a.size = size & ~MALLOC_CHUNK_MASK; // Round down size to fit chunk mask //printf("Adding %d bytes at %d to the memory pool\n",size,p); /* We need to insert this into the main block list in the proper place -- this list is required to be sorted. Since we most likely get memory assignments in ascending order, search backwards for the proper place. */ for (pah = __malloc_head.a.prev; pah->a.type != ARENA_TYPE_HEAD; pah = pah->a.prev) { if (pah < fp) break; } /* Now pah points to the node that should be the predecessor of the new node */ fp->a.next = pah->a.next; fp->a.prev = pah; pah->a.next = fp; fp->a.next->a.prev = fp; /* Insert into the free chain and coalesce with adjacent blocks */ fp = __free_block(fp); } void *malloc(size_t size) { struct free_arena_header *fp; struct free_arena_header *pah; size_t fsize; //printf("Custom malloc asking for 0x%x bytes\n",size); if (size == 0) return NULL; /* Add the obligatory arena header, and round up */ size = (size + 2 * sizeof(struct arena_header) - 1) & ARENA_SIZE_MASK; for (fp = __malloc_head.next_free; fp->a.type != ARENA_TYPE_HEAD; fp = fp->next_free) { if (fp->a.size >= size) { /* Found fit -- allocate out of this block */ return __malloc_from_block(fp, size); } } /* Nothing found... need to request a block from the kernel */ fsize = (size + MALLOC_CHUNK_MASK) & ~MALLOC_CHUNK_MASK; #if 0 // AMR - using a fixed arena. fp = (struct free_arena_header *) mmap(NULL, fsize, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANONYMOUS, 0, 0); if (fp == (struct free_arena_header *)MAP_FAILED) { return NULL; /* Failed to get a block */ } #endif fp = (struct free_arena_header *)_sbrk(fsize); if(fp==0 || fp == -1) return(NULL); /* Insert the block into the management chains. We need to set up the size and the main block list pointer, the rest of the work is logically identical to free(). */ fp->a.type = ARENA_TYPE_FREE; fp->a.size = fsize; /* We need to insert this into the main block list in the proper place -- this list is required to be sorted. Since we most likely get memory assignments in ascending order, search backwards for the proper place. */ for (pah = __malloc_head.a.prev; pah->a.type != ARENA_TYPE_HEAD; pah = pah->a.prev) { if (pah < fp) break; } /* Now pah points to the node that should be the predecessor of the new node */ fp->a.next = pah->a.next; fp->a.prev = pah; pah->a.next = fp; fp->a.next->a.prev = fp; /* Insert into the free chain and coalesce with adjacent blocks */ fp = __free_block(fp); /* Now we can allocate from this block */ return __malloc_from_block(fp, size); return NULL; } void free(void *ptr) { struct free_arena_header *ah; if (!ptr) return; ah = (struct free_arena_header *) ((struct arena_header *)ptr - 1); /* Merge into adjacent free blocks */ ah = __free_block(ah); #if 0 /* See if it makes sense to return memory to the system */ { size_t page_size = getpagesize(); size_t page_mask = page_size - 1; size_t head_portion = -(size_t)ah & page_mask; size_t tail_portion = ((size_t)ah + ah->a.size) & page_mask; size_t adj_size; /* Careful here... an individual chunk of memory must have a minimum size if it exists at all, so if either the head or the tail is below the minimum, then extend that chunk by a page. */ if (head_portion && head_portion < 2*sizeof(struct arena_header)) head_portion += page_size; if (tail_portion && tail_portion < 2*sizeof(struct arena_header)) tail_portion += page_size; adj_size = ah->a.size - head_portion - tail_portion; /* Worth it? This is written the way it is to guard against overflows... */ if (ah->a.size >= head_portion+tail_portion+ _KLIBC_MALLOC_CHUNK_SIZE) { struct free_arena_header *tah, *tan, *tap; if (tail_portion) { /* Make a new header, and insert into chains immediately after the current block */ tah = (struct free_arena_header *) ((char *)ah + head_portion + adj_size); tah->a.type = ARENA_TYPE_FREE; tah->a.size = tail_portion; tah->a.next = tan = ah->a.next; tan->a.prev = tah; tah->a.prev = ah; ah->a.next = tah; tah->prev_free = tap = ah->prev_free; tap->next_free = tah; tah->next_free = ah; ah->prev_free = tah; } if (head_portion) ah->a.size = head_portion; else remove_from_chains(ah); munmap((char *)ah + head_portion, adj_size); } } #endif } #ifdef __cplusplus } #endif