heap-validate.c 13 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404
  1. /*
  2. * Copyright (c) 2019 Intel Corporation
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. */
  6. #include <sys/sys_heap.h>
  7. #include <kernel.h>
  8. #include "heap.h"
  9. /* White-box sys_heap validation code. Uses internal data structures.
  10. * Not expected to be useful in production apps. This checks every
  11. * header field of every chunk and returns true if the totality of the
  12. * data structure is a valid heap. It doesn't necessarily tell you
  13. * that it is the CORRECT heap given the history of alloc/free calls
  14. * that it can't inspect. In a pathological case, you can imagine
  15. * something scribbling a copy of a previously-valid heap on top of a
  16. * running one and corrupting it. YMMV.
  17. */
  18. #define VALIDATE(cond) do { if (!(cond)) { return false; } } while (0)
  19. static bool in_bounds(struct z_heap *h, chunkid_t c)
  20. {
  21. VALIDATE(c >= right_chunk(h, 0));
  22. VALIDATE(c < h->end_chunk);
  23. VALIDATE(chunk_size(h, c) < h->end_chunk);
  24. return true;
  25. }
  26. static bool valid_chunk(struct z_heap *h, chunkid_t c)
  27. {
  28. VALIDATE(chunk_size(h, c) > 0);
  29. VALIDATE(c + chunk_size(h, c) <= h->end_chunk);
  30. VALIDATE(in_bounds(h, c));
  31. VALIDATE(right_chunk(h, left_chunk(h, c)) == c);
  32. VALIDATE(left_chunk(h, right_chunk(h, c)) == c);
  33. if (chunk_used(h, c)) {
  34. VALIDATE(!solo_free_header(h, c));
  35. } else {
  36. VALIDATE(chunk_used(h, left_chunk(h, c)));
  37. VALIDATE(chunk_used(h, right_chunk(h, c)));
  38. if (!solo_free_header(h, c)) {
  39. VALIDATE(in_bounds(h, prev_free_chunk(h, c)));
  40. VALIDATE(in_bounds(h, next_free_chunk(h, c)));
  41. }
  42. }
  43. return true;
  44. }
  45. /* Validate multiple state dimensions for the bucket "next" pointer
  46. * and see that they match. Probably should unify the design a
  47. * bit...
  48. */
  49. static inline void check_nexts(struct z_heap *h, int bidx)
  50. {
  51. struct z_heap_bucket *b = &h->buckets[bidx];
  52. bool emptybit = (h->avail_buckets & (1 << bidx)) == 0;
  53. bool emptylist = b->next == 0;
  54. bool empties_match = emptybit == emptylist;
  55. (void)empties_match;
  56. CHECK(empties_match);
  57. if (b->next != 0) {
  58. CHECK(valid_chunk(h, b->next));
  59. }
  60. }
  61. bool sys_heap_validate(struct sys_heap *heap)
  62. {
  63. struct z_heap *h = heap->heap;
  64. chunkid_t c;
  65. /*
  66. * Walk through the chunks linearly, verifying sizes and end pointer.
  67. */
  68. for (c = right_chunk(h, 0); c < h->end_chunk; c = right_chunk(h, c)) {
  69. if (!valid_chunk(h, c)) {
  70. return false;
  71. }
  72. }
  73. if (c != h->end_chunk) {
  74. return false; /* Should have exactly consumed the buffer */
  75. }
  76. /* Check the free lists: entry count should match, empty bit
  77. * should be correct, and all chunk entries should point into
  78. * valid unused chunks. Mark those chunks USED, temporarily.
  79. */
  80. for (int b = 0; b <= bucket_idx(h, h->end_chunk); b++) {
  81. chunkid_t c0 = h->buckets[b].next;
  82. uint32_t n = 0;
  83. check_nexts(h, b);
  84. for (c = c0; c != 0 && (n == 0 || c != c0);
  85. n++, c = next_free_chunk(h, c)) {
  86. if (!valid_chunk(h, c)) {
  87. return false;
  88. }
  89. set_chunk_used(h, c, true);
  90. }
  91. bool empty = (h->avail_buckets & (1 << b)) == 0;
  92. bool zero = n == 0;
  93. if (empty != zero) {
  94. return false;
  95. }
  96. if (empty && h->buckets[b].next != 0) {
  97. return false;
  98. }
  99. }
  100. /*
  101. * Walk through the chunks linearly again, verifying that all chunks
  102. * but solo headers are now USED (i.e. all free blocks were found
  103. * during enumeration). Mark all such blocks UNUSED and solo headers
  104. * USED.
  105. */
  106. chunkid_t prev_chunk = 0;
  107. for (c = right_chunk(h, 0); c < h->end_chunk; c = right_chunk(h, c)) {
  108. if (!chunk_used(h, c) && !solo_free_header(h, c)) {
  109. return false;
  110. }
  111. if (left_chunk(h, c) != prev_chunk) {
  112. return false;
  113. }
  114. prev_chunk = c;
  115. set_chunk_used(h, c, solo_free_header(h, c));
  116. }
  117. if (c != h->end_chunk) {
  118. return false; /* Should have exactly consumed the buffer */
  119. }
  120. /* Go through the free lists again checking that the linear
  121. * pass caught all the blocks and that they now show UNUSED.
  122. * Mark them USED.
  123. */
  124. for (int b = 0; b <= bucket_idx(h, h->end_chunk); b++) {
  125. chunkid_t c0 = h->buckets[b].next;
  126. int n = 0;
  127. if (c0 == 0) {
  128. continue;
  129. }
  130. for (c = c0; n == 0 || c != c0; n++, c = next_free_chunk(h, c)) {
  131. if (chunk_used(h, c)) {
  132. return false;
  133. }
  134. set_chunk_used(h, c, true);
  135. }
  136. }
  137. /* Now we are valid, but have managed to invert all the in-use
  138. * fields. One more linear pass to fix them up
  139. */
  140. for (c = right_chunk(h, 0); c < h->end_chunk; c = right_chunk(h, c)) {
  141. set_chunk_used(h, c, !chunk_used(h, c));
  142. }
  143. return true;
  144. }
  145. struct z_heap_stress_rec {
  146. void *(*alloc_fn)(void *arg, size_t bytes);
  147. void (*free_fn)(void *arg, void *p);
  148. void *arg;
  149. size_t total_bytes;
  150. struct z_heap_stress_block *blocks;
  151. size_t nblocks;
  152. size_t blocks_alloced;
  153. size_t bytes_alloced;
  154. uint32_t target_percent;
  155. };
  156. struct z_heap_stress_block {
  157. void *ptr;
  158. size_t sz;
  159. };
  160. /* Very simple LCRNG (from https://nuclear.llnl.gov/CNP/rng/rngman/node4.html)
  161. *
  162. * Here to guarantee cross-platform test repeatability.
  163. */
  164. static uint32_t rand32(void)
  165. {
  166. static uint64_t state = 123456789; /* seed */
  167. state = state * 2862933555777941757UL + 3037000493UL;
  168. return (uint32_t)(state >> 32);
  169. }
  170. static bool rand_alloc_choice(struct z_heap_stress_rec *sr)
  171. {
  172. /* Edge cases: no blocks allocated, and no space for a new one */
  173. if (sr->blocks_alloced == 0) {
  174. return true;
  175. } else if (sr->blocks_alloced >= sr->nblocks) {
  176. return false;
  177. } else {
  178. /* The way this works is to scale the chance of choosing to
  179. * allocate vs. free such that it's even odds when the heap is
  180. * at the target percent, with linear tapering on the low
  181. * slope (i.e. we choose to always allocate with an empty
  182. * heap, allocate 50% of the time when the heap is exactly at
  183. * the target, and always free when above the target). In
  184. * practice, the operations aren't quite symmetric (you can
  185. * always free, but your allocation might fail), and the units
  186. * aren't matched (we're doing math based on bytes allocated
  187. * and ignoring the overhead) but this is close enough. And
  188. * yes, the math here is coarse (in units of percent), but
  189. * that's good enough and fits well inside 32 bit quantities.
  190. * (Note precision issue when heap size is above 40MB
  191. * though!).
  192. */
  193. __ASSERT(sr->total_bytes < 0xffffffffU / 100, "too big for u32!");
  194. uint32_t full_pct = (100 * sr->bytes_alloced) / sr->total_bytes;
  195. uint32_t target = sr->target_percent ? sr->target_percent : 1;
  196. uint32_t free_chance = 0xffffffffU;
  197. if (full_pct < sr->target_percent) {
  198. free_chance = full_pct * (0x80000000U / target);
  199. }
  200. return rand32() > free_chance;
  201. }
  202. }
  203. /* Chooses a size of block to allocate, logarithmically favoring
  204. * smaller blocks (i.e. blocks twice as large are half as frequent
  205. */
  206. static size_t rand_alloc_size(struct z_heap_stress_rec *sr)
  207. {
  208. ARG_UNUSED(sr);
  209. /* Min scale of 4 means that the half of the requests in the
  210. * smallest size have an average size of 8
  211. */
  212. int scale = 4 + __builtin_clz(rand32());
  213. return rand32() & ((1 << scale) - 1);
  214. }
  215. /* Returns the index of a randomly chosen block to free */
  216. static size_t rand_free_choice(struct z_heap_stress_rec *sr)
  217. {
  218. return rand32() % sr->blocks_alloced;
  219. }
  220. /* General purpose heap stress test. Takes function pointers to allow
  221. * for testing multiple heap APIs with the same rig. The alloc and
  222. * free functions are passed back the argument as a context pointer.
  223. * The "log" function is for readable user output. The total_bytes
  224. * argument should reflect the size of the heap being tested. The
  225. * scratch array is used to store temporary state and should be sized
  226. * about half as large as the heap itself. Returns true on success.
  227. */
  228. void sys_heap_stress(void *(*alloc_fn)(void *arg, size_t bytes),
  229. void (*free_fn)(void *arg, void *p),
  230. void *arg, size_t total_bytes,
  231. uint32_t op_count,
  232. void *scratch_mem, size_t scratch_bytes,
  233. int target_percent,
  234. struct z_heap_stress_result *result)
  235. {
  236. struct z_heap_stress_rec sr = {
  237. .alloc_fn = alloc_fn,
  238. .free_fn = free_fn,
  239. .arg = arg,
  240. .total_bytes = total_bytes,
  241. .blocks = scratch_mem,
  242. .nblocks = scratch_bytes / sizeof(struct z_heap_stress_block),
  243. .target_percent = target_percent,
  244. };
  245. *result = (struct z_heap_stress_result) {0};
  246. for (uint32_t i = 0; i < op_count; i++) {
  247. if (rand_alloc_choice(&sr)) {
  248. size_t sz = rand_alloc_size(&sr);
  249. void *p = sr.alloc_fn(sr.arg, sz);
  250. result->total_allocs++;
  251. if (p != NULL) {
  252. result->successful_allocs++;
  253. sr.blocks[sr.blocks_alloced].ptr = p;
  254. sr.blocks[sr.blocks_alloced].sz = sz;
  255. sr.blocks_alloced++;
  256. sr.bytes_alloced += sz;
  257. }
  258. } else {
  259. int b = rand_free_choice(&sr);
  260. void *p = sr.blocks[b].ptr;
  261. size_t sz = sr.blocks[b].sz;
  262. result->total_frees++;
  263. sr.blocks[b] = sr.blocks[sr.blocks_alloced - 1];
  264. sr.blocks_alloced--;
  265. sr.bytes_alloced -= sz;
  266. sr.free_fn(sr.arg, p);
  267. }
  268. result->accumulated_in_use_bytes += sr.bytes_alloced;
  269. }
  270. }
  271. /*
  272. * Print heap info for debugging / analysis purpose
  273. */
  274. void heap_print_info(struct z_heap *h, bool dump_chunks)
  275. {
  276. int i, nb_buckets = bucket_idx(h, h->end_chunk) + 1;
  277. size_t free_bytes, allocated_bytes, total, overhead;
  278. #if !CONFIG_AEM_WATCH_SUPPORT
  279. printk("Heap at %p contains %d units in %d buckets\n\n", chunk_buf(h), h->end_chunk,
  280. nb_buckets);
  281. printk(" bucket# min units total largest largest\n"
  282. " threshold chunks (units) (bytes)\n"
  283. " -----------------------------------------------------------\n");
  284. #endif
  285. for (i = 0; i < nb_buckets; i++) {
  286. chunkid_t first = h->buckets[i].next;
  287. chunksz_t largest = 0;
  288. int count = 0;
  289. if (first) {
  290. chunkid_t curr = first;
  291. do {
  292. count++;
  293. largest = MAX(largest, chunk_size(h, curr));
  294. curr = next_free_chunk(h, curr);
  295. } while (curr != first);
  296. }
  297. #if !CONFIG_AEM_WATCH_SUPPORT
  298. if (count) {
  299. printk("%9d %12d %12d %12d %12zd\n",
  300. i, (1 << i) - 1 + min_chunk_size(h), count,
  301. largest, chunksz_to_bytes(h, largest));
  302. }
  303. #endif
  304. }
  305. #if !CONFIG_AEM_WATCH_SUPPORT
  306. if (dump_chunks) {
  307. printk("\nChunk dump:\n");
  308. }
  309. #endif
  310. free_bytes = allocated_bytes = 0;
  311. for (chunkid_t c = 0; ; c = right_chunk(h, c)) {
  312. if (chunk_used(h, c)) {
  313. if ((c != 0) && (c != h->end_chunk)) {
  314. /* 1st and last are always allocated for internal purposes */
  315. allocated_bytes += chunksz_to_bytes(h, chunk_size(h, c));
  316. }
  317. } else {
  318. if (!solo_free_header(h, c)) {
  319. free_bytes += chunksz_to_bytes(h, chunk_size(h, c));
  320. }
  321. }
  322. #if !CONFIG_AEM_WATCH_SUPPORT
  323. if (dump_chunks) {
  324. printk("chunk %4d: [%c] size=%-4d left=%-4d right=%d\n",
  325. c,
  326. chunk_used(h, c) ? '*'
  327. : solo_free_header(h, c) ? '.'
  328. : '-',
  329. chunk_size(h, c),
  330. left_chunk(h, c),
  331. right_chunk(h, c));
  332. }
  333. #endif
  334. if (c == h->end_chunk) {
  335. break;
  336. }
  337. }
  338. /* The end marker chunk has a header. It is part of the overhead. */
  339. total = h->end_chunk * CHUNK_UNIT + chunk_header_bytes(h);
  340. overhead = total - free_bytes - allocated_bytes;
  341. printk("\n%zd free bytes, %zd allocated bytes, overhead = %zd bytes (%zd.%zd%%)\n",
  342. free_bytes, allocated_bytes, overhead,
  343. (1000 * overhead + total/2) / total / 10,
  344. (1000 * overhead + total/2) / total % 10);
  345. #if CONFIG_AEM_WATCH_SUPPORT
  346. extern void aem_mem_refresh_ui(size_t free, size_t allocated);
  347. aem_mem_refresh_ui(free_bytes, allocated_bytes);
  348. #endif
  349. }
  350. void sys_heap_print_info(struct sys_heap *heap, bool dump_chunks)
  351. {
  352. heap_print_info(heap->heap, dump_chunks);
  353. }
  354. void sys_heap_dump(struct sys_heap *heap)
  355. {
  356. sys_heap_print_info(heap, true);
  357. }