display_composer.c 18 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645
  1. /*
  2. * Copyright (c) 2020 Actions Technology Co., Ltd
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. */
  6. #include <assert.h>
  7. #include <string.h>
  8. #include <sys/slist.h>
  9. #include <zephyr.h>
  10. #include <drivers/display/display_engine.h>
  11. #include <display/display_composer.h>
  12. #include <logging/log.h>
  13. #define SYS_LOG_DOMAIN "composer"
  14. #define CONFIG_COMPOSER_POST_NO_WAIT
  15. #define CONFIG_COMPOSER_NUM_POST_PARTS (7)
  16. /* must be power of 2; and must be 1 if both DMA and DE post available */
  17. #define NUM_POST_ENTRIES (CONFIG_COMPOSER_NUM_POST_PARTS * 2)
  18. #define NUM_POST_LAYERS (2)
  19. typedef struct post_entry {
  20. uint32_t flags;
  21. display_layer_t ovls[NUM_POST_LAYERS];
  22. display_buffer_t bufs[NUM_POST_LAYERS];
  23. graphic_buffer_t *graphic_bufs[NUM_POST_LAYERS];
  24. display_composer_post_cleanup_t cleanup_cb[NUM_POST_LAYERS];
  25. void *cleanup_data[NUM_POST_LAYERS];
  26. } post_entry_t;
  27. typedef struct display_composer {
  28. /* supported maximum layers */
  29. uint8_t max_layers;
  30. uint8_t post_pending : 1;
  31. uint8_t post_inprog : 1;
  32. uint8_t disp_active : 1;
  33. /* post entries */
  34. uint8_t free_idx;
  35. uint8_t post_idx;
  36. uint8_t complete_idx;
  37. uint8_t post_cnt;
  38. post_entry_t post_entries[NUM_POST_ENTRIES];
  39. struct k_spinlock post_lock;
  40. #ifndef CONFIG_COMPOSER_POST_NO_WAIT
  41. struct k_sem post_sem;
  42. #endif
  43. /* display engine device */
  44. const struct device *de_dev;
  45. int de_inst;
  46. /* display (panel) device */
  47. const struct device *disp_dev;
  48. /* display device callback */
  49. struct display_callback disp_cb;
  50. /* display write pixel format */
  51. uint32_t disp_pixel_formats;
  52. /* user display callback */
  53. const struct display_callback *user_cb;
  54. #ifdef CONFIG_DISPLAY_COMPOSER_DEBUG_FPS
  55. uint32_t frame_timestamp;
  56. uint16_t frame_cnt;
  57. #endif
  58. #ifdef CONFIG_DISPLAY_COMPOSER_DEBUG_VSYNC
  59. uint16_t vsync_cnt;
  60. uint32_t vsync_timestamp; /* measure in cycles */
  61. uint32_t vsync_period; /* measure in cycles */
  62. #endif
  63. } display_composer_t;
  64. /* global prototypes */
  65. extern const ui_region_t * display_composer_opt_full_screen_areas(int *num_regions);
  66. /* static prototypes */
  67. static void _composer_display_vsync_handler(const struct display_callback *callback, uint32_t timestamp);
  68. static void _composer_display_complete_handler(const struct display_callback *callback);
  69. static void _composer_display_pm_notify_handler(const struct display_callback *callback, uint32_t pm_action);
  70. static uint8_t _composer_num_free_entries_get(display_composer_t *composer);
  71. static post_entry_t *_composer_find_free_entry(display_composer_t *composer);
  72. static int _composer_post_top_entry(display_composer_t *composer, bool require_not_first);
  73. static void _composer_cleanup_entry(display_composer_t *composer, post_entry_t *entry);
  74. /* static variables */
  75. static display_composer_t display_composer __in_section_unique(ram.noinit.display_composer);
  76. static inline display_composer_t *_composer_get(void)
  77. {
  78. return &display_composer;
  79. }
  80. int display_composer_init(void)
  81. {
  82. display_composer_t *composer = _composer_get();
  83. union {
  84. struct display_capabilities panel;
  85. struct display_engine_capabilities engine;
  86. } capabilities;
  87. memset(composer, 0, sizeof(*composer));
  88. composer->disp_active = 1;
  89. composer->disp_dev = device_get_binding(CONFIG_LCD_DISPLAY_DEV_NAME);
  90. if (composer->disp_dev == NULL) {
  91. SYS_LOG_ERR("cannot find display " CONFIG_LCD_DISPLAY_DEV_NAME);
  92. return -ENODEV;
  93. }
  94. composer->de_dev = device_get_binding(CONFIG_DISPLAY_ENGINE_DEV_NAME);
  95. if (composer->de_dev) {
  96. composer->de_inst = display_engine_open(composer->de_dev,
  97. DISPLAY_ENGINE_FLAG_HIGH_PRIO | DISPLAY_ENGINE_FLAG_POST);
  98. } else {
  99. composer->de_inst = -1;
  100. }
  101. if (composer->de_inst >= 0) {
  102. display_engine_get_capabilities(composer->de_dev, &capabilities.engine);
  103. composer->max_layers = MIN(capabilities.engine.num_overlays, NUM_POST_LAYERS);
  104. } else {
  105. composer->max_layers = 1;
  106. }
  107. SYS_LOG_INF("supported layer num %d\n", composer->max_layers);
  108. #ifndef CONFIG_COMPOSER_POST_NO_WAIT
  109. k_sem_init(&composer->post_sem, NUM_POST_ENTRIES, NUM_POST_ENTRIES);
  110. #endif
  111. display_get_capabilities(composer->disp_dev, &capabilities.panel);
  112. composer->disp_pixel_formats = capabilities.panel.supported_pixel_formats;
  113. if (!(capabilities.panel.screen_info & SCREEN_INFO_VSYNC)) {
  114. SYS_LOG_WRN("vsync unsupported\n");
  115. }
  116. composer->disp_cb.vsync = _composer_display_vsync_handler;
  117. composer->disp_cb.complete = _composer_display_complete_handler;
  118. composer->disp_cb.pm_notify = _composer_display_pm_notify_handler;
  119. display_register_callback(composer->disp_dev, &composer->disp_cb);
  120. SYS_LOG_INF("composer initialized\n");
  121. return 0;
  122. }
  123. void display_composer_destroy(void)
  124. {
  125. display_composer_t *composer = _composer_get();
  126. if (composer->disp_dev == NULL) {
  127. return;
  128. }
  129. display_unregister_callback(composer->disp_dev, &composer->disp_cb);
  130. if (composer->de_inst >= 0) {
  131. display_engine_close(composer->de_dev, composer->de_inst);
  132. }
  133. SYS_LOG_INF("composer finalized\n");
  134. }
  135. void display_composer_register_callback(const struct display_callback *callback)
  136. {
  137. display_composer_t *composer = _composer_get();
  138. composer->user_cb = callback;
  139. }
  140. uint32_t display_composer_get_vsync_period(void)
  141. {
  142. display_composer_t *composer = _composer_get();
  143. struct display_capabilities capabilities;
  144. if (composer->disp_dev == NULL) {
  145. return -ENODEV;
  146. }
  147. capabilities.vsync_period = 0;
  148. display_get_capabilities(composer->disp_dev, &capabilities);
  149. /* fallback to refresh rate 60 Hz */
  150. return capabilities.vsync_period ? capabilities.vsync_period : 16667;
  151. }
  152. int display_composer_get_geometry(
  153. uint16_t *width, uint16_t *height, uint32_t *pixel_format)
  154. {
  155. display_composer_t *composer = _composer_get();
  156. struct display_capabilities capabilities;
  157. if (composer->disp_dev == NULL) {
  158. return -ENODEV;
  159. }
  160. display_get_capabilities(composer->disp_dev, &capabilities);
  161. if (width) {
  162. *width = capabilities.x_resolution;
  163. }
  164. if (height) {
  165. *height = capabilities.y_resolution;
  166. }
  167. if (pixel_format) {
  168. *pixel_format = capabilities.current_pixel_format;
  169. }
  170. return 0;
  171. }
  172. int display_composer_set_blanking(bool blanking_on)
  173. {
  174. display_composer_t *composer = _composer_get();
  175. int res = 0;
  176. if (composer->disp_dev == NULL) {
  177. return -ENODEV;
  178. }
  179. if (blanking_on) {
  180. res = display_blanking_on(composer->disp_dev);
  181. } else {
  182. res = display_blanking_off(composer->disp_dev);
  183. }
  184. return res;
  185. }
  186. int display_composer_set_brightness(uint8_t brightness)
  187. {
  188. display_composer_t *composer = _composer_get();
  189. if (composer->disp_dev == NULL) {
  190. return -ENODEV;
  191. }
  192. return display_set_brightness(composer->disp_dev, brightness);
  193. }
  194. int display_composer_set_contrast(uint8_t contrast)
  195. {
  196. display_composer_t *composer = _composer_get();
  197. if (composer->disp_dev == NULL) {
  198. return -ENODEV;
  199. }
  200. return display_set_contrast(composer->disp_dev, contrast);
  201. }
  202. void display_composer_round(ui_region_t *region)
  203. {
  204. /*
  205. * display_write() (DMA path) requires the buffer address, and the byte length
  206. * corresponding to with and stride respectively are 4-byte aligned.
  207. *
  208. * Also some LCD driver IC, like GC9C01, requires the position and dims are both even.
  209. */
  210. region->x1 &= ~0x1;
  211. region->y1 &= ~0x1;
  212. region->x2 |= 0x1;
  213. region->y2 |= 0x1;
  214. }
  215. static void _composer_display_vsync_handler(const struct display_callback *callback, uint32_t timestamp)
  216. {
  217. display_composer_t *composer = CONTAINER_OF(callback, display_composer_t, disp_cb);
  218. if (composer->post_cnt > 0 && !composer->post_inprog) {
  219. _composer_post_top_entry(composer, false);
  220. }
  221. if (composer->user_cb && composer->user_cb->vsync) {
  222. composer->user_cb->vsync(composer->user_cb, timestamp);
  223. }
  224. #ifdef CONFIG_DISPLAY_COMPOSER_DEBUG_VSYNC
  225. composer->vsync_period = timestamp - composer->vsync_timestamp;
  226. composer->vsync_timestamp = timestamp;
  227. if (++composer->vsync_cnt == 1024) {
  228. LOG_INF("vsync period %u us\n", k_cyc_to_us_near32(composer->vsync_period));
  229. composer->vsync_cnt = 0;
  230. }
  231. #endif
  232. }
  233. static void _composer_display_complete_handler(const struct display_callback *callback)
  234. {
  235. display_composer_t *composer = CONTAINER_OF(callback, display_composer_t, disp_cb);
  236. composer->post_inprog = 0;
  237. if (--composer->post_cnt > 0) {
  238. _composer_post_top_entry(composer, true);
  239. }
  240. post_entry_t *entry = &composer->post_entries[composer->complete_idx];
  241. _composer_cleanup_entry(composer, entry);
  242. if (++composer->complete_idx >= NUM_POST_ENTRIES)
  243. composer->complete_idx = 0;
  244. if (composer->post_cnt == 0 && composer->user_cb && composer->user_cb->complete) {
  245. composer->user_cb->complete(composer->user_cb);
  246. }
  247. }
  248. static void _composer_display_pm_notify_handler(const struct display_callback *callback, uint32_t pm_action)
  249. {
  250. display_composer_t *composer = CONTAINER_OF(callback, display_composer_t, disp_cb);
  251. composer->disp_active = (pm_action == PM_DEVICE_ACTION_LATE_RESUME);
  252. if (pm_action == PM_DEVICE_ACTION_EARLY_SUSPEND) {
  253. while (composer->post_cnt > 0) {
  254. SYS_LOG_INF("post cnt %d", composer->post_cnt);
  255. os_sleep(2);
  256. }
  257. }
  258. if (composer->user_cb && composer->user_cb->pm_notify) {
  259. composer->user_cb->pm_notify(composer->user_cb, pm_action);
  260. }
  261. }
  262. static uint8_t _composer_num_free_entries_get(display_composer_t *composer)
  263. {
  264. return NUM_POST_ENTRIES - composer->post_cnt;
  265. }
  266. static post_entry_t *_composer_find_free_entry(display_composer_t *composer)
  267. {
  268. post_entry_t *entry;
  269. #ifdef CONFIG_COMPOSER_POST_NO_WAIT
  270. if (composer->post_cnt >= NUM_POST_ENTRIES)
  271. return NULL;
  272. #else
  273. if (k_sem_take(&composer->post_sem, k_is_in_isr() ? K_NO_WAIT : K_FOREVER))
  274. return NULL;
  275. #endif
  276. entry = &composer->post_entries[composer->free_idx];
  277. if (++composer->free_idx >= NUM_POST_ENTRIES)
  278. composer->free_idx = 0;
  279. memset(entry, 0, sizeof(*entry));
  280. return entry;
  281. }
  282. static int _composer_post_top_entry(display_composer_t *composer, bool require_not_first)
  283. {
  284. post_entry_t *entry = &composer->post_entries[composer->post_idx];
  285. display_layer_t *ovls = entry->ovls;
  286. uint8_t num_layers = ovls[1].buffer ? 2 : 1;
  287. int res = -EINVAL;
  288. if (require_not_first && (entry->flags & FIRST_POST_IN_FRAME)) {
  289. return -EINVAL;
  290. }
  291. #if 0
  292. /* Try display_write() first, then display_engine_compose */
  293. if (num_layers == 1 && ovls[0].buffer != NULL &&
  294. (ovls[0].buffer->desc.pixel_format & composer->disp_pixel_formats)) {
  295. res = display_write(composer->disp_dev, ovls[0].frame.x, ovls[0].frame.y,
  296. &ovls[0].buffer->desc, (void *)ovls[0].buffer->addr);
  297. }
  298. if (res < 0 && composer->de_inst >= 0) {
  299. res = display_engine_compose(composer->de_dev, composer->de_inst,
  300. NULL, ovls, num_layers, false);
  301. }
  302. #else
  303. if (entry->flags & POST_PATH_BY_DE) {
  304. res = display_engine_compose(composer->de_dev, composer->de_inst,
  305. NULL, ovls, num_layers, false);
  306. } else {
  307. res = display_write(composer->disp_dev, ovls[0].frame.x, ovls[0].frame.y,
  308. &ovls[0].buffer->desc, (void *)ovls[0].buffer->addr);
  309. }
  310. #endif
  311. assert(res >= 0);
  312. composer->post_inprog = (res >= 0);
  313. if (++composer->post_idx >= NUM_POST_ENTRIES)
  314. composer->post_idx = 0;
  315. return res;
  316. }
  317. static void _composer_cleanup_entry(display_composer_t *composer, post_entry_t *entry)
  318. {
  319. for (int i = 0; i < ARRAY_SIZE(entry->graphic_bufs); i++) {
  320. if (entry->graphic_bufs[i]) {
  321. graphic_buffer_unref(entry->graphic_bufs[i]);
  322. }
  323. if (entry->cleanup_cb[i]) {
  324. entry->cleanup_cb[i](entry->cleanup_data[i]);
  325. }
  326. }
  327. #ifndef CONFIG_COMPOSER_POST_NO_WAIT
  328. k_sem_give(&composer->post_sem);
  329. #endif
  330. }
  331. int display_composer_simple_post(graphic_buffer_t *buffer,
  332. ui_region_t *crop, uint16_t x, uint16_t y)
  333. {
  334. ui_layer_t layer;
  335. memset(&layer, 0, sizeof(layer));
  336. layer.buffer = buffer;
  337. if (crop) {
  338. memcpy(&layer.crop, crop, sizeof(*crop));
  339. } else {
  340. layer.crop.x2 = graphic_buffer_get_width(buffer) - 1;
  341. layer.crop.y2 = graphic_buffer_get_height(buffer) - 1;
  342. }
  343. layer.frame.x1 = x;
  344. layer.frame.y1 = y;
  345. layer.frame.x2 = x + ui_region_get_width(&layer.crop) - 1;
  346. layer.frame.y2 = y + ui_region_get_height(&layer.crop) - 1;
  347. return display_composer_post(&layer, 1, FIRST_POST_IN_FRAME | LAST_POST_IN_FRAME);
  348. }
  349. static int _composer_post_inner(const ui_layer_t *layers, int num_layers, uint32_t post_flags)
  350. {
  351. display_composer_t *composer = _composer_get();
  352. post_entry_t *entry = NULL;
  353. int res = -EINVAL;
  354. int i;
  355. /* Get the free entry */
  356. entry = _composer_find_free_entry(composer);
  357. if (entry == NULL) {
  358. goto fail_cleanup_cb;
  359. }
  360. entry->flags = post_flags;
  361. /* Validate the buffer */
  362. for (i = 0; i < num_layers; i++) {
  363. if (layers[i].buffer) {
  364. entry->bufs[i].addr = (uint32_t)graphic_buffer_get_bufptr(
  365. layers[i].buffer, layers[i].crop.x1, layers[i].crop.y1);
  366. //if (entry->bufs[i].addr & 0x3)
  367. // entry->flags |= POST_PATH_BY_DE;
  368. entry->bufs[i].desc.pitch = graphic_buffer_get_stride(layers[i].buffer);
  369. entry->bufs[i].desc.pixel_format =
  370. graphic_buffer_get_pixel_format(layers[i].buffer);
  371. entry->bufs[i].desc.buf_size = entry->bufs[i].desc.pitch * entry->bufs[i].desc.height *
  372. display_format_get_bits_per_pixel(entry->bufs[i].desc.pixel_format) / 8;
  373. } else {
  374. entry->bufs[i].addr = 0;
  375. entry->bufs[i].desc.pitch = 0;
  376. entry->bufs[i].desc.pixel_format = 0;
  377. }
  378. entry->bufs[i].desc.width = ui_region_get_width(&layers[i].frame);
  379. entry->bufs[i].desc.height = ui_region_get_height(&layers[i].frame);
  380. entry->ovls[i].buffer = (entry->bufs[i].addr > 0) ? &entry->bufs[i] : NULL;
  381. entry->ovls[i].frame.x = layers[i].frame.x1;
  382. entry->ovls[i].frame.y = layers[i].frame.y1;
  383. entry->ovls[i].frame.w = entry->bufs[i].desc.width;
  384. entry->ovls[i].frame.h = entry->bufs[i].desc.height;
  385. entry->ovls[i].color.full = layers[i].color.full;
  386. entry->ovls[i].blending = layers[i].blending;
  387. if (!layers[i].buf_resident) {
  388. entry->graphic_bufs[i] = layers[i].buffer;
  389. if (layers[i].buffer)
  390. graphic_buffer_ref(layers[i].buffer);
  391. }
  392. entry->cleanup_cb[i] = layers[i].cleanup_cb;
  393. entry->cleanup_data[i] = layers[i].cleanup_data;
  394. SYS_LOG_DBG("L-%d: ptr=0x%08x fmt=0x%02x stride=%u color=0x%08x blend=0x%x frame=(%d,%d,%d,%d)%s",
  395. i, entry->bufs[i].addr, entry->bufs[i].desc.pixel_format, entry->bufs[i].desc.pitch,
  396. layers[i].color.full, layers[i].blending, layers[i].frame.x1,
  397. layers[i].frame.y1, layers[i].frame.x2, layers[i].frame.y2,
  398. i == num_layers - 1 ? "\n" : "");
  399. }
  400. k_spinlock_key_t key = k_spin_lock(&composer->post_lock);
  401. composer->post_cnt++;
  402. if (!composer->post_inprog) {
  403. _composer_post_top_entry(composer, true);
  404. }
  405. k_spin_unlock(&composer->post_lock, key);
  406. return 0;
  407. fail_cleanup_cb:
  408. for (i = 0; i < num_layers; i++) {
  409. if (layers[i].cleanup_cb)
  410. layers[i].cleanup_cb(layers[i].cleanup_data);
  411. }
  412. return res;
  413. }
  414. static int _composer_post_part(const ui_layer_t *layers, int num_layers,
  415. uint32_t *post_flags, const ui_region_t *window)
  416. {
  417. ui_layer_t tmp_layers[NUM_POST_LAYERS];
  418. int i, j = 0;
  419. for (i = 0; i < num_layers; i++) {
  420. if (window->y1 <= layers[i].frame.y2 && window->y2 >= layers[i].frame.y2) {
  421. tmp_layers[j].cleanup_cb = layers[i].cleanup_cb;
  422. tmp_layers[j].cleanup_data = layers[i].cleanup_data;
  423. } else {
  424. tmp_layers[j].cleanup_cb = NULL;
  425. }
  426. if (ui_region_intersect(&tmp_layers[j].frame, &layers[i].frame, window) == false) {
  427. if (tmp_layers[j].cleanup_cb) {
  428. tmp_layers[j].cleanup_cb(tmp_layers[j].cleanup_data);
  429. tmp_layers[j].cleanup_cb = NULL;
  430. }
  431. continue;
  432. }
  433. tmp_layers[j].buffer = layers[i].buffer;
  434. tmp_layers[j].buf_resident = layers[i].buf_resident;
  435. tmp_layers[j].color = layers[i].color;
  436. tmp_layers[j].blending = layers[i].blending;
  437. if (layers[i].buffer) {
  438. tmp_layers[j].crop.x1 = layers[i].crop.x1 + tmp_layers[j].frame.x1 - layers[i].frame.x1;
  439. tmp_layers[j].crop.y1 = layers[i].crop.y1 + tmp_layers[j].frame.y1 - layers[i].frame.y1;
  440. }
  441. j++;
  442. }
  443. if (j > 0 && _composer_post_inner(tmp_layers, j, *post_flags) == 0) {
  444. *post_flags &= ~FIRST_POST_IN_FRAME;
  445. return 0;
  446. }
  447. return -EINVAL;
  448. }
  449. int display_composer_post(const ui_layer_t *layers, int num_layers, uint32_t post_flags)
  450. {
  451. display_composer_t *composer = _composer_get();
  452. const ui_region_t * clip_areas;
  453. int res, num, i;
  454. #ifdef CONFIG_DISPLAY_COMPOSER_DEBUG_FPS
  455. if (post_flags & LAST_POST_IN_FRAME) {
  456. uint32_t timestamp = k_cycle_get_32();
  457. ++composer->frame_cnt;
  458. if ((timestamp - composer->frame_timestamp) >= sys_clock_hw_cycles_per_sec()) {
  459. LOG_INF("post fps %u\n", composer->frame_cnt);
  460. composer->frame_cnt = 0;
  461. composer->frame_timestamp = timestamp;
  462. }
  463. }
  464. #endif
  465. if (composer->disp_dev == NULL) {
  466. SYS_LOG_ERR("composer not initialized");
  467. goto fail_cleanup_cb;
  468. }
  469. if (composer->disp_active == 0) {
  470. SYS_LOG_DBG("display inactive");
  471. goto fail_cleanup_cb;
  472. }
  473. if (num_layers <= 0 || num_layers > composer->max_layers) {
  474. SYS_LOG_ERR("invalid layer num %d", num_layers);
  475. goto fail_cleanup_cb;
  476. }
  477. if (num_layers > 1 || layers[0].buffer == NULL ||
  478. !(graphic_buffer_get_pixel_format(layers[0].buffer) & composer->disp_pixel_formats)) {
  479. post_flags |= POST_PATH_BY_DE;
  480. if (composer->de_inst < 0) {
  481. SYS_LOG_ERR("DE path not unavailable");
  482. goto fail_cleanup_cb;
  483. }
  484. }
  485. clip_areas = display_composer_opt_full_screen_areas(&num);
  486. if (clip_areas == NULL) num = 1;
  487. #ifdef CONFIG_COMPOSER_POST_NO_WAIT
  488. if (_composer_num_free_entries_get(composer) < num)
  489. #else
  490. if (k_is_in_isr() && _composer_num_free_entries_get(composer) < num)
  491. #endif
  492. {
  493. #ifdef CONFIG_DISPLAY_COMPOSER_DEBUG_VSYNC
  494. SYS_LOG_WRN("drop 1 frame (last vsync %u)", k_cyc_to_us_near32(composer->vsync_period));
  495. #else
  496. SYS_LOG_WRN("drop 1 frame");
  497. #endif
  498. goto fail_cleanup_cb;
  499. }
  500. if (clip_areas && (post_flags & POST_FULL_SCREEN_OPT)) {
  501. uint32_t flags = post_flags & ~LAST_POST_IN_FRAME;
  502. _composer_post_part(layers, num_layers, &flags, &clip_areas[0]);
  503. for (i = 1; i < num - 1; i++)
  504. _composer_post_part(layers, num_layers, &flags, &clip_areas[i]);
  505. flags |= (post_flags & LAST_POST_IN_FRAME);
  506. res = _composer_post_part(layers, num_layers, &flags, &clip_areas[num - 1]);
  507. } else {
  508. return _composer_post_inner(layers, num_layers, post_flags);
  509. }
  510. return 0;
  511. fail_cleanup_cb:
  512. for (i = 0; i < num_layers; i++) {
  513. if (layers[i].cleanup_cb)
  514. layers[i].cleanup_cb(layers[i].cleanup_data);
  515. }
  516. return -EINVAL;
  517. }