heap.c 11 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427
  1. /*
  2. * Copyright (c) 2019 Intel Corporation
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. */
  6. #include <sys/sys_heap.h>
  7. #include <kernel.h>
  8. #include <string.h>
  9. #include "heap.h"
  10. static void *chunk_mem(struct z_heap *h, chunkid_t c)
  11. {
  12. chunk_unit_t *buf = chunk_buf(h);
  13. uint8_t *ret = ((uint8_t *)&buf[c]) + chunk_header_bytes(h);
  14. CHECK(!(((uintptr_t)ret) & (big_heap(h) ? 7 : 3)));
  15. return ret;
  16. }
  17. static void free_list_remove_bidx(struct z_heap *h, chunkid_t c, int bidx)
  18. {
  19. struct z_heap_bucket *b = &h->buckets[bidx];
  20. CHECK(!chunk_used(h, c));
  21. CHECK(b->next != 0);
  22. CHECK(h->avail_buckets & (1 << bidx));
  23. if (next_free_chunk(h, c) == c) {
  24. /* this is the last chunk */
  25. h->avail_buckets &= ~(1 << bidx);
  26. b->next = 0;
  27. } else {
  28. chunkid_t first = prev_free_chunk(h, c),
  29. second = next_free_chunk(h, c);
  30. b->next = second;
  31. set_next_free_chunk(h, first, second);
  32. set_prev_free_chunk(h, second, first);
  33. }
  34. }
  35. static void free_list_remove(struct z_heap *h, chunkid_t c)
  36. {
  37. if (!solo_free_header(h, c)) {
  38. int bidx = bucket_idx(h, chunk_size(h, c));
  39. free_list_remove_bidx(h, c, bidx);
  40. }
  41. }
  42. static void free_list_add_bidx(struct z_heap *h, chunkid_t c, int bidx)
  43. {
  44. struct z_heap_bucket *b = &h->buckets[bidx];
  45. if (b->next == 0U) {
  46. CHECK((h->avail_buckets & (1 << bidx)) == 0);
  47. /* Empty list, first item */
  48. h->avail_buckets |= (1 << bidx);
  49. b->next = c;
  50. set_prev_free_chunk(h, c, c);
  51. set_next_free_chunk(h, c, c);
  52. } else {
  53. CHECK(h->avail_buckets & (1 << bidx));
  54. /* Insert before (!) the "next" pointer */
  55. chunkid_t second = b->next;
  56. chunkid_t first = prev_free_chunk(h, second);
  57. set_prev_free_chunk(h, c, first);
  58. set_next_free_chunk(h, c, second);
  59. set_next_free_chunk(h, first, c);
  60. set_prev_free_chunk(h, second, c);
  61. }
  62. }
  63. static void free_list_add(struct z_heap *h, chunkid_t c)
  64. {
  65. if (!solo_free_header(h, c)) {
  66. int bidx = bucket_idx(h, chunk_size(h, c));
  67. free_list_add_bidx(h, c, bidx);
  68. }
  69. }
  70. /* Splits a chunk "lc" into a left chunk and a right chunk at "rc".
  71. * Leaves both chunks marked "free"
  72. */
  73. static void split_chunks(struct z_heap *h, chunkid_t lc, chunkid_t rc)
  74. {
  75. CHECK(rc > lc);
  76. CHECK(rc - lc < chunk_size(h, lc));
  77. chunksz_t sz0 = chunk_size(h, lc);
  78. chunksz_t lsz = rc - lc;
  79. chunksz_t rsz = sz0 - lsz;
  80. set_chunk_size(h, lc, lsz);
  81. set_chunk_size(h, rc, rsz);
  82. set_left_chunk_size(h, rc, lsz);
  83. set_left_chunk_size(h, right_chunk(h, rc), rsz);
  84. }
  85. /* Does not modify free list */
  86. static void merge_chunks(struct z_heap *h, chunkid_t lc, chunkid_t rc)
  87. {
  88. chunksz_t newsz = chunk_size(h, lc) + chunk_size(h, rc);
  89. set_chunk_size(h, lc, newsz);
  90. set_left_chunk_size(h, right_chunk(h, rc), newsz);
  91. }
  92. static void free_chunk(struct z_heap *h, chunkid_t c)
  93. {
  94. /* Merge with free right chunk? */
  95. if (!chunk_used(h, right_chunk(h, c))) {
  96. free_list_remove(h, right_chunk(h, c));
  97. merge_chunks(h, c, right_chunk(h, c));
  98. }
  99. /* Merge with free left chunk? */
  100. if (!chunk_used(h, left_chunk(h, c))) {
  101. free_list_remove(h, left_chunk(h, c));
  102. merge_chunks(h, left_chunk(h, c), c);
  103. c = left_chunk(h, c);
  104. }
  105. free_list_add(h, c);
  106. }
  107. /*
  108. * Return the closest chunk ID corresponding to given memory pointer.
  109. * Here "closest" is only meaningful in the context of sys_heap_aligned_alloc()
  110. * where wanted alignment might not always correspond to a chunk header
  111. * boundary.
  112. */
  113. static chunkid_t mem_to_chunkid(struct z_heap *h, void *p)
  114. {
  115. uint8_t *mem = p, *base = (uint8_t *)chunk_buf(h);
  116. return (mem - chunk_header_bytes(h) - base) / CHUNK_UNIT;
  117. }
  118. void sys_heap_free(struct sys_heap *heap, void *mem)
  119. {
  120. if (mem == NULL) {
  121. return; /* ISO C free() semantics */
  122. }
  123. struct z_heap *h = heap->heap;
  124. chunkid_t c = mem_to_chunkid(h, mem);
  125. /*
  126. * This should catch many double-free cases.
  127. * This is cheap enough so let's do it all the time.
  128. */
  129. __ASSERT(chunk_used(h, c),
  130. "unexpected heap state (double-free?) for memory at %p", mem);
  131. /*
  132. * It is easy to catch many common memory overflow cases with
  133. * a quick check on this and next chunk header fields that are
  134. * immediately before and after the freed memory.
  135. */
  136. __ASSERT(left_chunk(h, right_chunk(h, c)) == c,
  137. "corrupted heap bounds (buffer overflow?) for memory at %p",
  138. mem);
  139. set_chunk_used(h, c, false);
  140. free_chunk(h, c);
  141. }
  142. static chunkid_t alloc_chunk(struct z_heap *h, chunksz_t sz)
  143. {
  144. int bi = bucket_idx(h, sz);
  145. struct z_heap_bucket *b = &h->buckets[bi];
  146. CHECK(bi <= bucket_idx(h, h->end_chunk));
  147. /* First try a bounded count of items from the minimal bucket
  148. * size. These may not fit, trying (e.g.) three means that
  149. * (assuming that chunk sizes are evenly distributed[1]) we
  150. * have a 7/8 chance of finding a match, thus keeping the
  151. * number of such blocks consumed by allocation higher than
  152. * the number of smaller blocks created by fragmenting larger
  153. * ones.
  154. *
  155. * [1] In practice, they are never evenly distributed, of
  156. * course. But even in pathological situations we still
  157. * maintain our constant time performance and at worst see
  158. * fragmentation waste of the order of the block allocated
  159. * only.
  160. */
  161. if (b->next) {
  162. chunkid_t first = b->next;
  163. int i = CONFIG_SYS_HEAP_ALLOC_LOOPS;
  164. do {
  165. chunkid_t c = b->next;
  166. if (chunk_size(h, c) >= sz) {
  167. free_list_remove_bidx(h, c, bi);
  168. return c;
  169. }
  170. b->next = next_free_chunk(h, c);
  171. CHECK(b->next != 0);
  172. } while (--i && b->next != first);
  173. }
  174. /* Otherwise pick the smallest non-empty bucket guaranteed to
  175. * fit and use that unconditionally.
  176. */
  177. uint32_t bmask = h->avail_buckets & ~((1 << (bi + 1)) - 1);
  178. if (bmask != 0U) {
  179. int minbucket = __builtin_ctz(bmask);
  180. chunkid_t c = h->buckets[minbucket].next;
  181. free_list_remove_bidx(h, c, minbucket);
  182. CHECK(chunk_size(h, c) >= sz);
  183. return c;
  184. }
  185. return 0;
  186. }
  187. void *sys_heap_alloc(struct sys_heap *heap, size_t bytes)
  188. {
  189. struct z_heap *h = heap->heap;
  190. if (bytes == 0U || size_too_big(h, bytes)) {
  191. return NULL;
  192. }
  193. chunksz_t chunk_sz = bytes_to_chunksz(h, bytes);
  194. chunkid_t c = alloc_chunk(h, chunk_sz);
  195. if (c == 0U) {
  196. return NULL;
  197. }
  198. /* Split off remainder if any */
  199. if (chunk_size(h, c) > chunk_sz) {
  200. split_chunks(h, c, c + chunk_sz);
  201. free_list_add(h, c + chunk_sz);
  202. }
  203. set_chunk_used(h, c, true);
  204. return chunk_mem(h, c);
  205. }
  206. void *sys_heap_aligned_alloc(struct sys_heap *heap, size_t align, size_t bytes)
  207. {
  208. struct z_heap *h = heap->heap;
  209. size_t gap, rew;
  210. /*
  211. * Split align and rewind values (if any).
  212. * We allow for one bit of rewind in addition to the alignment
  213. * value to efficiently accommodate z_heap_aligned_alloc().
  214. * So if e.g. align = 0x28 (32 | 8) this means we align to a 32-byte
  215. * boundary and then rewind 8 bytes.
  216. */
  217. rew = align & -align;
  218. if (align != rew) {
  219. align -= rew;
  220. gap = MIN(rew, chunk_header_bytes(h));
  221. } else {
  222. if (align <= chunk_header_bytes(h)) {
  223. return sys_heap_alloc(heap, bytes);
  224. }
  225. rew = 0;
  226. gap = chunk_header_bytes(h);
  227. }
  228. __ASSERT((align & (align - 1)) == 0, "align must be a power of 2");
  229. if (bytes == 0 || size_too_big(h, bytes)) {
  230. return NULL;
  231. }
  232. /*
  233. * Find a free block that is guaranteed to fit.
  234. * We over-allocate to account for alignment and then free
  235. * the extra allocations afterwards.
  236. */
  237. chunksz_t padded_sz = bytes_to_chunksz(h, bytes + align - gap);
  238. chunkid_t c0 = alloc_chunk(h, padded_sz);
  239. if (c0 == 0) {
  240. return NULL;
  241. }
  242. uint8_t *mem = chunk_mem(h, c0);
  243. /* Align allocated memory */
  244. mem = (uint8_t *) ROUND_UP(mem + rew, align) - rew;
  245. chunk_unit_t *end = (chunk_unit_t *) ROUND_UP(mem + bytes, CHUNK_UNIT);
  246. /* Get corresponding chunks */
  247. chunkid_t c = mem_to_chunkid(h, mem);
  248. chunkid_t c_end = end - chunk_buf(h);
  249. CHECK(c >= c0 && c < c_end && c_end <= c0 + padded_sz);
  250. /* Split and free unused prefix */
  251. if (c > c0) {
  252. split_chunks(h, c0, c);
  253. free_list_add(h, c0);
  254. }
  255. /* Split and free unused suffix */
  256. if (right_chunk(h, c) > c_end) {
  257. split_chunks(h, c, c_end);
  258. free_list_add(h, c_end);
  259. }
  260. set_chunk_used(h, c, true);
  261. return mem;
  262. }
  263. void *sys_heap_aligned_realloc(struct sys_heap *heap, void *ptr,
  264. size_t align, size_t bytes)
  265. {
  266. struct z_heap *h = heap->heap;
  267. /* special realloc semantics */
  268. if (ptr == NULL) {
  269. return sys_heap_aligned_alloc(heap, align, bytes);
  270. }
  271. if (bytes == 0) {
  272. sys_heap_free(heap, ptr);
  273. return NULL;
  274. }
  275. __ASSERT((align & (align - 1)) == 0, "align must be a power of 2");
  276. if (size_too_big(h, bytes)) {
  277. return NULL;
  278. }
  279. chunkid_t c = mem_to_chunkid(h, ptr);
  280. chunkid_t rc = right_chunk(h, c);
  281. size_t align_gap = (uint8_t *)ptr - (uint8_t *)chunk_mem(h, c);
  282. chunksz_t chunks_need = bytes_to_chunksz(h, bytes + align_gap);
  283. if (align && ((uintptr_t)ptr & (align - 1))) {
  284. /* ptr is not sufficiently aligned */
  285. } else if (chunk_size(h, c) == chunks_need) {
  286. /* We're good already */
  287. return ptr;
  288. } else if (chunk_size(h, c) > chunks_need) {
  289. /* Shrink in place, split off and free unused suffix */
  290. split_chunks(h, c, c + chunks_need);
  291. set_chunk_used(h, c, true);
  292. free_chunk(h, c + chunks_need);
  293. return ptr;
  294. } else if (!chunk_used(h, rc) &&
  295. (chunk_size(h, c) + chunk_size(h, rc) >= chunks_need)) {
  296. /* Expand: split the right chunk and append */
  297. chunkid_t split_size = chunks_need - chunk_size(h, c);
  298. free_list_remove(h, rc);
  299. if (split_size < chunk_size(h, rc)) {
  300. split_chunks(h, rc, rc + split_size);
  301. free_list_add(h, rc + split_size);
  302. }
  303. merge_chunks(h, c, rc);
  304. set_chunk_used(h, c, true);
  305. return ptr;
  306. } else {
  307. ;
  308. }
  309. /* Fallback: allocate and copy */
  310. void *ptr2 = sys_heap_aligned_alloc(heap, align, bytes);
  311. if (ptr2 != NULL) {
  312. size_t prev_size = chunksz_to_bytes(h, chunk_size(h, c)) - align_gap;
  313. memcpy(ptr2, ptr, MIN(prev_size, bytes));
  314. sys_heap_free(heap, ptr);
  315. }
  316. return ptr2;
  317. }
  318. void sys_heap_init(struct sys_heap *heap, void *mem, size_t bytes)
  319. {
  320. /* Must fit in a 31 bit count of HUNK_UNIT */
  321. __ASSERT(bytes / CHUNK_UNIT <= 0x7fffffffU, "heap size is too big");
  322. /* Reserve the end marker chunk's header */
  323. __ASSERT(bytes > heap_footer_bytes(bytes), "heap size is too small");
  324. bytes -= heap_footer_bytes(bytes);
  325. /* Round the start up, the end down */
  326. uintptr_t addr = ROUND_UP(mem, CHUNK_UNIT);
  327. uintptr_t end = ROUND_DOWN((uint8_t *)mem + bytes, CHUNK_UNIT);
  328. chunksz_t heap_sz = (end - addr) / CHUNK_UNIT;
  329. CHECK(end > addr);
  330. __ASSERT(heap_sz > chunksz(sizeof(struct z_heap)), "heap size is too small");
  331. struct z_heap *h = (struct z_heap *)addr;
  332. heap->heap = h;
  333. h->end_chunk = heap_sz;
  334. h->avail_buckets = 0;
  335. int nb_buckets = bucket_idx(h, heap_sz) + 1;
  336. chunksz_t chunk0_size = chunksz(sizeof(struct z_heap) +
  337. nb_buckets * sizeof(struct z_heap_bucket));
  338. __ASSERT(chunk0_size + min_chunk_size(h) <= heap_sz, "heap size is too small");
  339. for (int i = 0; i < nb_buckets; i++) {
  340. h->buckets[i].next = 0;
  341. }
  342. /* chunk containing our struct z_heap */
  343. set_chunk_size(h, 0, chunk0_size);
  344. set_left_chunk_size(h, 0, 0);
  345. set_chunk_used(h, 0, true);
  346. /* chunk containing the free heap */
  347. set_chunk_size(h, chunk0_size, heap_sz - chunk0_size);
  348. set_left_chunk_size(h, chunk0_size, chunk0_size);
  349. /* the end marker chunk */
  350. set_chunk_size(h, heap_sz, 0);
  351. set_left_chunk_size(h, heap_sz, heap_sz - chunk0_size);
  352. set_chunk_used(h, heap_sz, true);
  353. free_list_add(h, chunk0_size);
  354. }