diff options
author | Rich Felker <dalias@aerifal.cx> | 2018-04-19 20:45:48 -0400 |
---|---|---|
committer | Rich Felker <dalias@aerifal.cx> | 2018-04-19 20:45:48 -0400 |
commit | 3c2cbbe7ba8b4486299ae0d5336ae01ab520d116 (patch) | |
tree | f269c212fa64c10f68e990550ef6c91c885592e3 /src/malloc | |
parent | 23389b1988b061e8487c316893a8a8eb77770a2f (diff) | |
download | musl-3c2cbbe7ba8b4486299ae0d5336ae01ab520d116.tar.gz musl-3c2cbbe7ba8b4486299ae0d5336ae01ab520d116.tar.bz2 musl-3c2cbbe7ba8b4486299ae0d5336ae01ab520d116.tar.xz musl-3c2cbbe7ba8b4486299ae0d5336ae01ab520d116.zip |
using malloc implementation types/macros/idioms for memalign
the generated code should be mostly unchanged, except for explicit use
of C_INUSE in place of copying the low bits from existing chunk
headers/footers.
these changes also remove mild UB due to dubious arithmetic on
pointers into imaginary size_t[] arrays.
Diffstat (limited to 'src/malloc')
-rw-r--r-- | src/malloc/memalign.c | 42 |
1 files changed, 22 insertions, 20 deletions
diff --git a/src/malloc/memalign.c b/src/malloc/memalign.c index 8fb2002c..9c420229 100644 --- a/src/malloc/memalign.c +++ b/src/malloc/memalign.c @@ -2,49 +2,51 @@ #include <stdint.h> #include <errno.h> #include "libc.h" +#include "malloc_impl.h" void *__memalign(size_t align, size_t len) { - unsigned char *mem, *new, *end; - size_t header, footer; + unsigned char *mem, *new; if ((align & -align) != align) { errno = EINVAL; - return NULL; + return 0; } if (len > SIZE_MAX - align) { errno = ENOMEM; - return NULL; + return 0; } - if (align <= 4*sizeof(size_t)) { - if (!(mem = malloc(len))) - return NULL; - return mem; - } + if (align <= SIZE_ALIGN) + return malloc(len); if (!(mem = malloc(len + align-1))) - return NULL; + return 0; new = (void *)((uintptr_t)mem + align-1 & -align); if (new == mem) return mem; - header = ((size_t *)mem)[-1]; + struct chunk *c = MEM_TO_CHUNK(mem); + struct chunk *n = MEM_TO_CHUNK(new); - if (!(header & 7)) { - ((size_t *)new)[-2] = ((size_t *)mem)[-2] + (new-mem); - ((size_t *)new)[-1] = ((size_t *)mem)[-1] - (new-mem); + if (IS_MMAPPED(c)) { + /* Apply difference between aligned and original + * address to the "extra" field of mmapped chunk. */ + n->psize = c->psize + (new-mem); + n->csize = c->csize - (new-mem); return new; } - end = mem + (header & -8); - footer = ((size_t *)end)[-2]; + struct chunk *t = NEXT_CHUNK(c); - ((size_t *)mem)[-1] = header&7 | new-mem; - ((size_t *)new)[-2] = footer&7 | new-mem; - ((size_t *)new)[-1] = header&7 | end-new; - ((size_t *)end)[-2] = footer&7 | end-new; + /* Split the allocated chunk into two chunks. The aligned part + * that will be used has the size in its footer reduced by the + * difference between the aligned and original addresses, and + * the resulting size copied to its header. A new header and + * footer are written for the split-off part to be freed. */ + n->psize = c->csize = C_INUSE | (new-mem); + n->csize = t->psize -= new-mem; free(mem); return new; |