kheap.c 3.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136
  1. /*
  2. * Copyright (c) 2020 Intel Corporation
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. */
  6. #include <kernel.h>
  7. #include <ksched.h>
  8. #include <wait_q.h>
  9. #include <init.h>
  10. #include <linker/linker-defs.h>
  11. void k_heap_init(struct k_heap *h, void *mem, size_t bytes)
  12. {
  13. z_waitq_init(&h->wait_q);
  14. sys_heap_init(&h->heap, mem, bytes);
  15. SYS_PORT_TRACING_OBJ_INIT(k_heap, h);
  16. }
  17. static int statics_init(const struct device *unused)
  18. {
  19. ARG_UNUSED(unused);
  20. STRUCT_SECTION_FOREACH(k_heap, h) {
  21. #if defined(CONFIG_DEMAND_PAGING) && !defined(CONFIG_LINKER_GENERIC_SECTIONS_PRESENT_AT_BOOT)
  22. /* Some heaps may not present at boot, so we need to wait for
  23. * paging mechanism to be initialized before we can initialize
  24. * each heap.
  25. */
  26. extern bool z_sys_post_kernel;
  27. bool do_clear = z_sys_post_kernel;
  28. /* During pre-kernel init, z_sys_post_kernel == false,
  29. * initialize if within pinned region. Otherwise skip.
  30. * In post-kernel init, z_sys_post_kernel == true, skip those in
  31. * pinned region as they have already been initialized and
  32. * possibly already in use. Otherwise initialize.
  33. */
  34. if (lnkr_is_pinned((uint8_t *)h) &&
  35. lnkr_is_pinned((uint8_t *)&h->wait_q) &&
  36. lnkr_is_region_pinned((uint8_t *)h->heap.init_mem,
  37. h->heap.init_bytes)) {
  38. do_clear = !do_clear;
  39. }
  40. if (do_clear)
  41. #endif /* CONFIG_DEMAND_PAGING && !CONFIG_LINKER_GENERIC_SECTIONS_PRESENT_AT_BOOT */
  42. {
  43. k_heap_init(h, h->heap.init_mem, h->heap.init_bytes);
  44. }
  45. }
  46. return 0;
  47. }
  48. SYS_INIT(statics_init, PRE_KERNEL_1, CONFIG_KERNEL_INIT_PRIORITY_OBJECTS);
  49. #if defined(CONFIG_DEMAND_PAGING) && !defined(CONFIG_LINKER_GENERIC_SECTIONS_PRESENT_AT_BOOT)
  50. /* Need to wait for paging mechanism to be initialized before
  51. * heaps that are not in pinned sections can be initialized.
  52. */
  53. SYS_INIT(statics_init, POST_KERNEL, 0);
  54. #endif /* CONFIG_DEMAND_PAGING && !CONFIG_LINKER_GENERIC_SECTIONS_PRESENT_AT_BOOT */
  55. void *k_heap_aligned_alloc(struct k_heap *h, size_t align, size_t bytes,
  56. k_timeout_t timeout)
  57. {
  58. int64_t now, end = sys_clock_timeout_end_calc(timeout);
  59. void *ret = NULL;
  60. k_spinlock_key_t key = k_spin_lock(&h->lock);
  61. SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_heap, aligned_alloc, h, timeout);
  62. __ASSERT(!arch_is_in_isr() || K_TIMEOUT_EQ(timeout, K_NO_WAIT), "");
  63. bool blocked_alloc = false;
  64. while (ret == NULL) {
  65. ret = sys_heap_aligned_alloc(&h->heap, align, bytes);
  66. now = sys_clock_tick_get();
  67. if (!IS_ENABLED(CONFIG_MULTITHREADING) ||
  68. (ret != NULL) || ((end - now) <= 0)) {
  69. break;
  70. }
  71. if (!blocked_alloc) {
  72. blocked_alloc = true;
  73. SYS_PORT_TRACING_OBJ_FUNC_BLOCKING(k_heap, aligned_alloc, h, timeout);
  74. } else {
  75. /**
  76. * @todo Trace attempt to avoid empty trace segments
  77. */
  78. }
  79. (void) z_pend_curr(&h->lock, key, &h->wait_q,
  80. K_TICKS(end - now));
  81. key = k_spin_lock(&h->lock);
  82. }
  83. SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_heap, aligned_alloc, h, timeout, ret);
  84. k_spin_unlock(&h->lock, key);
  85. if(ret == NULL){
  86. printk("kalloc %d fail\n", bytes);
  87. // sys_heap_dump(&h->heap);
  88. }
  89. return ret;
  90. }
  91. void *k_heap_alloc(struct k_heap *h, size_t bytes, k_timeout_t timeout)
  92. {
  93. SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_heap, alloc, h, timeout);
  94. void *ret = k_heap_aligned_alloc(h, sizeof(void *), bytes, timeout);
  95. SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_heap, alloc, h, timeout, ret);
  96. return ret;
  97. }
  98. void k_heap_free(struct k_heap *h, void *mem)
  99. {
  100. k_spinlock_key_t key = k_spin_lock(&h->lock);
  101. sys_heap_free(&h->heap, mem);
  102. SYS_PORT_TRACING_OBJ_FUNC(k_heap, free, h);
  103. if (IS_ENABLED(CONFIG_MULTITHREADING) && z_unpend_all(&h->wait_q) != 0) {
  104. z_reschedule(&h->lock, key);
  105. } else {
  106. k_spin_unlock(&h->lock, key);
  107. }
  108. }