mm_page.c 22 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884
  1. /*
  2. * Copyright (c) 2006-2022, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Change Logs:
  7. * Date Author Notes
  8. * 2019-11-01 Jesven The first version
  9. * 2022-12-13 WangXiaoyao Hot-pluggable, extensible
  10. * page management algorithm
  11. * 2023-02-20 WangXiaoyao Multi-list page-management
  12. */
  13. #include <rtthread.h>
  14. #include <stddef.h>
  15. #include <stdint.h>
  16. #include <string.h>
  17. #include "mm_fault.h"
  18. #include "mm_private.h"
  19. #include "mm_aspace.h"
  20. #include "mm_flag.h"
  21. #include "mm_page.h"
  22. #include <mmu.h>
  23. #define DBG_TAG "mm.page"
  24. #define DBG_LVL DBG_WARNING
  25. #include <rtdbg.h>
  26. RT_CTASSERT(order_huge_pg, RT_PAGE_MAX_ORDER > ARCH_PAGE_SHIFT - 2);
  27. RT_CTASSERT(size_width, sizeof(rt_size_t) == sizeof(void *));
  28. #ifdef RT_USING_SMART
  29. #include "lwp_arch_comm.h"
  30. #endif /* RT_USING_SMART */
  31. static rt_size_t init_mpr_align_start;
  32. static rt_size_t init_mpr_align_end;
  33. static void *init_mpr_cont_start;
  34. static struct rt_varea mpr_varea;
  35. static struct rt_page *page_list_low[RT_PAGE_MAX_ORDER];
  36. static struct rt_page *page_list_high[RT_PAGE_MAX_ORDER];
  37. #define page_start ((rt_page_t)rt_mpr_start)
  38. static rt_size_t page_nr;
  39. static rt_size_t _high_pages_nr;
  40. static rt_size_t early_offset;
  41. static const char *get_name(rt_varea_t varea)
  42. {
  43. return "master-page-record";
  44. }
  45. static void hint_free(rt_mm_va_hint_t hint)
  46. {
  47. hint->flags = MMF_MAP_FIXED;
  48. hint->limit_start = rt_kernel_space.start;
  49. hint->limit_range_size = rt_kernel_space.size;
  50. hint->prefer = rt_mpr_start;
  51. }
  52. static void on_page_fault(struct rt_varea *varea, struct rt_aspace_fault_msg *msg)
  53. {
  54. char *init_start = (void *)init_mpr_align_start;
  55. char *init_end = (void *)init_mpr_align_end;
  56. if ((char *)msg->fault_vaddr < init_end && (char *)msg->fault_vaddr >= init_start)
  57. {
  58. rt_size_t offset = (char *)msg->fault_vaddr - init_start;
  59. msg->response.status = MM_FAULT_STATUS_OK;
  60. msg->response.vaddr = (char *)init_mpr_cont_start + offset;
  61. msg->response.size = ARCH_PAGE_SIZE;
  62. }
  63. else
  64. {
  65. rt_mm_dummy_mapper.on_page_fault(varea, msg);
  66. }
  67. }
  68. static struct rt_mem_obj mm_page_mapper = {
  69. .get_name = get_name,
  70. .on_page_fault = on_page_fault,
  71. .hint_free = hint_free,
  72. };
  73. #ifdef RT_DEBUG_PAGE_LEAK
  74. static volatile int enable;
  75. static rt_page_t _trace_head;
  76. #define TRACE_ALLOC(pg, size) _trace_alloc(pg, __builtin_return_address(0), size)
  77. #define TRACE_FREE(pgaddr, size) _trace_free(pgaddr, __builtin_return_address(0), size)
  78. void rt_page_leak_trace_start()
  79. {
  80. // TODO multicore safety
  81. _trace_head = NULL;
  82. enable = 1;
  83. }
  84. MSH_CMD_EXPORT(rt_page_leak_trace_start, start page leak tracer);
  85. static void _collect()
  86. {
  87. rt_page_t page = _trace_head;
  88. if (!page)
  89. {
  90. rt_kputs("ok!\n");
  91. }
  92. while (page)
  93. {
  94. rt_page_t next = page->next;
  95. void *pg_va = rt_page_page2addr(page);
  96. LOG_W("LEAK: %p, allocator: %p, size bits: %lx", pg_va, page->caller, page->trace_size);
  97. rt_pages_free(pg_va, page->trace_size);
  98. page = next;
  99. }
  100. }
  101. void rt_page_leak_trace_stop()
  102. {
  103. // TODO multicore safety
  104. enable = 0;
  105. _collect();
  106. }
  107. MSH_CMD_EXPORT(rt_page_leak_trace_stop, stop page leak tracer);
  108. static void _trace_alloc(rt_page_t page, void *caller, size_t size_bits)
  109. {
  110. if (enable)
  111. {
  112. page->caller = caller;
  113. page->trace_size = size_bits;
  114. page->tl_prev = NULL;
  115. page->tl_next = NULL;
  116. if (_trace_head == NULL)
  117. {
  118. _trace_head = page;
  119. }
  120. else
  121. {
  122. _trace_head->tl_prev = page;
  123. page->tl_next = _trace_head;
  124. _trace_head = page;
  125. }
  126. }
  127. }
  128. void _report(rt_page_t page, size_bits, char *msg)
  129. {
  130. void *pg_va = rt_page_page2addr(page);
  131. LOG_W("%s: %p, allocator: %p, size bits: %lx", msg, pg_va, page->caller, page->trace_size);
  132. rt_kputs("backtrace\n");
  133. rt_hw_backtrace(0, 0);
  134. }
  135. static void _trace_free(rt_page_t page, void *caller, size_t size_bits)
  136. {
  137. if (enable)
  138. {
  139. /* free after free */
  140. if (page->trace_size == 0xabadcafe)
  141. {
  142. _report("free after free")
  143. return ;
  144. }
  145. else if (page->trace_size != size_bits)
  146. {
  147. rt_kprintf("free with size bits %lx\n", size_bits);
  148. _report("incompatible size bits parameter");
  149. return ;
  150. }
  151. if (page->ref_cnt == 1)
  152. {
  153. if (page->tl_prev)
  154. page->tl_prev->tl_next = page->tl_next;
  155. if (page->tl_next)
  156. page->tl_next->tl_prev = page->tl_prev;
  157. if (page == _trace_head)
  158. _trace_head = page->next;
  159. page->tl_prev = NULL;
  160. page->tl_next = NULL;
  161. page->trace_size = 0xabadcafe;
  162. }
  163. }
  164. }
  165. #else
  166. #define TRACE_ALLOC(x, y)
  167. #define TRACE_FREE(x, y)
  168. #endif
  169. static inline void *page_to_addr(rt_page_t page)
  170. {
  171. return (void *)(((page - page_start) << ARCH_PAGE_SHIFT) - PV_OFFSET);
  172. }
  173. static inline rt_page_t addr_to_page(rt_page_t pg_start, void *addr)
  174. {
  175. addr = (char *)addr + PV_OFFSET;
  176. return &pg_start[((rt_ubase_t)addr >> ARCH_PAGE_SHIFT)];
  177. }
  178. #define FLOOR(val, align) (((rt_size_t)(val) + (align)-1) & ~((align)-1))
  179. const rt_size_t shadow_mask =
  180. ((1ul << (RT_PAGE_MAX_ORDER + ARCH_PAGE_SHIFT - 1)) - 1);
  181. const rt_size_t rt_mpr_size = FLOOR(
  182. ((1ul << (ARCH_VADDR_WIDTH - ARCH_PAGE_SHIFT))) * sizeof(struct rt_page),
  183. ARCH_PAGE_SIZE);
  184. void *rt_mpr_start;
  185. rt_weak int rt_hw_clz(unsigned long n)
  186. {
  187. return __builtin_clzl(n);
  188. }
  189. rt_weak int rt_hw_ctz(unsigned long n)
  190. {
  191. return __builtin_ctzl(n);
  192. }
  193. rt_size_t rt_page_bits(rt_size_t size)
  194. {
  195. int bit = sizeof(rt_size_t) * 8 - rt_hw_clz(size) - 1;
  196. if ((size ^ (1UL << bit)) != 0)
  197. {
  198. bit++;
  199. }
  200. bit -= ARCH_PAGE_SHIFT;
  201. if (bit < 0)
  202. {
  203. bit = 0;
  204. }
  205. return bit;
  206. }
  207. struct rt_page *rt_page_addr2page(void *addr)
  208. {
  209. return addr_to_page(page_start, addr);
  210. }
  211. void *rt_page_page2addr(struct rt_page *p)
  212. {
  213. return page_to_addr(p);
  214. }
  215. static inline struct rt_page *_buddy_get(struct rt_page *p,
  216. rt_uint32_t size_bits)
  217. {
  218. rt_size_t addr;
  219. addr = (rt_size_t)rt_page_page2addr(p);
  220. addr ^= (1UL << (size_bits + ARCH_PAGE_SHIFT));
  221. return rt_page_addr2page((void *)addr);
  222. }
  223. static void _page_remove(rt_page_t page_list[], struct rt_page *p, rt_uint32_t size_bits)
  224. {
  225. if (p->pre)
  226. {
  227. p->pre->next = p->next;
  228. }
  229. else
  230. {
  231. page_list[size_bits] = p->next;
  232. }
  233. if (p->next)
  234. {
  235. p->next->pre = p->pre;
  236. }
  237. p->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  238. }
  239. static void _page_insert(rt_page_t page_list[], struct rt_page *p, rt_uint32_t size_bits)
  240. {
  241. p->next = page_list[size_bits];
  242. if (p->next)
  243. {
  244. p->next->pre = p;
  245. }
  246. p->pre = 0;
  247. page_list[size_bits] = p;
  248. p->size_bits = size_bits;
  249. }
  250. static void _pages_ref_inc(struct rt_page *p, rt_uint32_t size_bits)
  251. {
  252. struct rt_page *page_head;
  253. int idx;
  254. /* find page group head */
  255. idx = p - page_start;
  256. idx = idx & ~((1UL << size_bits) - 1);
  257. page_head = page_start + idx;
  258. page_head = (void *)((char *)page_head + early_offset);
  259. page_head->ref_cnt++;
  260. }
  261. static int _pages_ref_get(struct rt_page *p, rt_uint32_t size_bits)
  262. {
  263. struct rt_page *page_head;
  264. int idx;
  265. /* find page group head */
  266. idx = p - page_start;
  267. idx = idx & ~((1UL << size_bits) - 1);
  268. page_head = page_start + idx;
  269. return page_head->ref_cnt;
  270. }
  271. static int _pages_free(rt_page_t page_list[], struct rt_page *p, rt_uint32_t size_bits)
  272. {
  273. rt_uint32_t level = size_bits;
  274. struct rt_page *buddy;
  275. RT_ASSERT(p >= page_start);
  276. RT_ASSERT((char *)p < (char *)rt_mpr_start + rt_mpr_size);
  277. RT_ASSERT(rt_kmem_v2p(p));
  278. RT_ASSERT(p->ref_cnt > 0);
  279. RT_ASSERT(p->size_bits == ARCH_ADDRESS_WIDTH_BITS);
  280. RT_ASSERT(size_bits < RT_PAGE_MAX_ORDER);
  281. p->ref_cnt--;
  282. if (p->ref_cnt != 0)
  283. {
  284. return 0;
  285. }
  286. while (level < RT_PAGE_MAX_ORDER - 1)
  287. {
  288. buddy = _buddy_get(p, level);
  289. if (buddy && buddy->size_bits == level)
  290. {
  291. _page_remove(page_list, buddy, level);
  292. p = (p < buddy) ? p : buddy;
  293. level++;
  294. }
  295. else
  296. {
  297. break;
  298. }
  299. }
  300. _page_insert(page_list, p, level);
  301. return 1;
  302. }
  303. static struct rt_page *_pages_alloc(rt_page_t page_list[], rt_uint32_t size_bits)
  304. {
  305. struct rt_page *p;
  306. if (page_list[size_bits])
  307. {
  308. p = page_list[size_bits];
  309. _page_remove(page_list, p, size_bits);
  310. }
  311. else
  312. {
  313. rt_uint32_t level;
  314. for (level = size_bits + 1; level < RT_PAGE_MAX_ORDER; level++)
  315. {
  316. if (page_list[level])
  317. {
  318. break;
  319. }
  320. }
  321. if (level == RT_PAGE_MAX_ORDER)
  322. {
  323. return 0;
  324. }
  325. p = page_list[level];
  326. _page_remove(page_list, p, level);
  327. while (level > size_bits)
  328. {
  329. _page_insert(page_list, p, level - 1);
  330. p = _buddy_get(p, level - 1);
  331. level--;
  332. }
  333. }
  334. p->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  335. p->ref_cnt = 1;
  336. return p;
  337. }
  338. static void _early_page_remove(rt_page_t page_list[], rt_page_t page, rt_uint32_t size_bits)
  339. {
  340. rt_page_t page_cont = (rt_page_t)((char *)page + early_offset);
  341. if (page_cont->pre)
  342. {
  343. rt_page_t pre_cont = (rt_page_t)((char *)page_cont->pre + early_offset);
  344. pre_cont->next = page_cont->next;
  345. }
  346. else
  347. {
  348. page_list[size_bits] = page_cont->next;
  349. }
  350. if (page_cont->next)
  351. {
  352. rt_page_t next_cont = (rt_page_t)((char *)page_cont->next + early_offset);
  353. next_cont->pre = page_cont->pre;
  354. }
  355. page_cont->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  356. }
  357. static void _early_page_insert(rt_page_t page_list[], rt_page_t page, int size_bits)
  358. {
  359. RT_ASSERT((void *)page >= rt_mpr_start &&
  360. ((char *)page - (char *)rt_mpr_start) < rt_mpr_size);
  361. rt_page_t page_cont = (rt_page_t)((char *)page + early_offset);
  362. page_cont->next = page_list[size_bits];
  363. if (page_cont->next)
  364. {
  365. rt_page_t next_cont = (rt_page_t)((char *)page_cont->next + early_offset);
  366. next_cont->pre = page;
  367. }
  368. page_cont->pre = 0;
  369. page_list[size_bits] = page;
  370. page_cont->size_bits = size_bits;
  371. }
  372. static struct rt_page *_early_pages_alloc(rt_page_t page_list[], rt_uint32_t size_bits)
  373. {
  374. struct rt_page *p;
  375. if (page_list[size_bits])
  376. {
  377. p = page_list[size_bits];
  378. _early_page_remove(page_list, p, size_bits);
  379. }
  380. else
  381. {
  382. rt_uint32_t level;
  383. for (level = size_bits + 1; level < RT_PAGE_MAX_ORDER; level++)
  384. {
  385. if (page_list[level])
  386. {
  387. break;
  388. }
  389. }
  390. if (level == RT_PAGE_MAX_ORDER)
  391. {
  392. return 0;
  393. }
  394. p = page_list[level];
  395. _early_page_remove(page_list, p, level);
  396. while (level > size_bits)
  397. {
  398. _early_page_insert(page_list, p, level - 1);
  399. p = _buddy_get(p, level - 1);
  400. level--;
  401. }
  402. }
  403. rt_page_t page_cont = (rt_page_t)((char *)p + early_offset);
  404. page_cont->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  405. page_cont->ref_cnt = 1;
  406. return p;
  407. }
  408. static rt_page_t *_get_page_list(void *vaddr)
  409. {
  410. rt_ubase_t pa_int = (rt_ubase_t)vaddr + PV_OFFSET;
  411. rt_page_t *list;
  412. if (pa_int > UINT32_MAX)
  413. {
  414. list = page_list_high;
  415. }
  416. else
  417. {
  418. list = page_list_low;
  419. }
  420. return list;
  421. }
  422. int rt_page_ref_get(void *addr, rt_uint32_t size_bits)
  423. {
  424. struct rt_page *p;
  425. rt_base_t level;
  426. int ref;
  427. p = rt_page_addr2page(addr);
  428. level = rt_hw_interrupt_disable();
  429. ref = _pages_ref_get(p, size_bits);
  430. rt_hw_interrupt_enable(level);
  431. return ref;
  432. }
  433. void rt_page_ref_inc(void *addr, rt_uint32_t size_bits)
  434. {
  435. struct rt_page *p;
  436. rt_base_t level;
  437. p = rt_page_addr2page(addr);
  438. level = rt_hw_interrupt_disable();
  439. _pages_ref_inc(p, size_bits);
  440. rt_hw_interrupt_enable(level);
  441. }
  442. static rt_page_t (*pages_alloc_handler)(rt_page_t page_list[], rt_uint32_t size_bits);
  443. /* if not, we skip the finding on page_list_high */
  444. static size_t _high_page_configured = 0;
  445. static rt_page_t *_flag_to_page_list(size_t flags)
  446. {
  447. rt_page_t *page_list;
  448. if (_high_page_configured && (flags & PAGE_ANY_AVAILABLE))
  449. {
  450. page_list = page_list_high;
  451. }
  452. else
  453. {
  454. page_list = page_list_low;
  455. }
  456. return page_list;
  457. }
  458. static void *_do_pages_alloc(rt_uint32_t size_bits, size_t flags)
  459. {
  460. void *alloc_buf = RT_NULL;
  461. struct rt_page *p;
  462. rt_base_t level;
  463. rt_page_t *page_list = _flag_to_page_list(flags);
  464. level = rt_hw_interrupt_disable();
  465. p = pages_alloc_handler(page_list, size_bits);
  466. rt_hw_interrupt_enable(level);
  467. if (!p && page_list != page_list_low)
  468. {
  469. /* fall back */
  470. page_list = page_list_low;
  471. level = rt_hw_interrupt_disable();
  472. p = pages_alloc_handler(page_list, size_bits);
  473. rt_hw_interrupt_enable(level);
  474. }
  475. if (p)
  476. {
  477. alloc_buf = page_to_addr(p);
  478. #ifdef RT_DEBUG_PAGE_LEAK
  479. level = rt_hw_interrupt_disable();
  480. TRACE_ALLOC(p, size_bits);
  481. rt_hw_interrupt_enable(level);
  482. #endif
  483. }
  484. return alloc_buf;
  485. }
  486. void *rt_pages_alloc(rt_uint32_t size_bits)
  487. {
  488. return _do_pages_alloc(size_bits, 0);
  489. }
  490. void *rt_pages_alloc_ext(rt_uint32_t size_bits, size_t flags)
  491. {
  492. return _do_pages_alloc(size_bits, flags);
  493. }
  494. int rt_pages_free(void *addr, rt_uint32_t size_bits)
  495. {
  496. struct rt_page *p;
  497. rt_page_t *page_list = _get_page_list(addr);
  498. int real_free = 0;
  499. p = rt_page_addr2page(addr);
  500. if (p)
  501. {
  502. rt_base_t level;
  503. level = rt_hw_interrupt_disable();
  504. real_free = _pages_free(page_list, p, size_bits);
  505. if (real_free)
  506. TRACE_FREE(p, size_bits);
  507. rt_hw_interrupt_enable(level);
  508. }
  509. return real_free;
  510. }
  511. void rt_page_list(void) __attribute__((alias("list_page")));
  512. #warning TODO: improve list page
  513. void list_page(void)
  514. {
  515. int i;
  516. rt_size_t total = 0;
  517. rt_base_t level;
  518. level = rt_hw_interrupt_disable();
  519. for (i = 0; i < RT_PAGE_MAX_ORDER; i++)
  520. {
  521. struct rt_page *p = page_list_low[i];
  522. rt_kprintf("level %d ", i);
  523. while (p)
  524. {
  525. total += (1UL << i);
  526. rt_kprintf("[0x%08p]", rt_page_page2addr(p));
  527. p = p->next;
  528. }
  529. rt_kprintf("\n");
  530. }
  531. for (i = 0; i < RT_PAGE_MAX_ORDER; i++)
  532. {
  533. struct rt_page *p = page_list_high[i];
  534. rt_kprintf("level %d ", i);
  535. while (p)
  536. {
  537. total += (1UL << i);
  538. rt_kprintf("[0x%08p]", rt_page_page2addr(p));
  539. p = p->next;
  540. }
  541. rt_kprintf("\n");
  542. }
  543. rt_hw_interrupt_enable(level);
  544. rt_kprintf("free pages is 0x%08lx (%ld KB)\n", total, total * ARCH_PAGE_SIZE / 1024);
  545. rt_kprintf("-------------------------------\n");
  546. }
  547. MSH_CMD_EXPORT(list_page, show page info);
  548. void rt_page_get_info(rt_size_t *total_nr, rt_size_t *free_nr)
  549. {
  550. int i;
  551. rt_size_t total_free = 0;
  552. rt_base_t level;
  553. level = rt_hw_interrupt_disable();
  554. for (i = 0; i < RT_PAGE_MAX_ORDER; i++)
  555. {
  556. struct rt_page *p = page_list_low[i];
  557. while (p)
  558. {
  559. total_free += (1UL << i);
  560. p = p->next;
  561. }
  562. }
  563. for (i = 0; i < RT_PAGE_MAX_ORDER; i++)
  564. {
  565. struct rt_page *p = page_list_high[i];
  566. while (p)
  567. {
  568. total_free += (1UL << i);
  569. p = p->next;
  570. }
  571. }
  572. rt_hw_interrupt_enable(level);
  573. *total_nr = page_nr;
  574. *free_nr = total_free;
  575. }
  576. void rt_page_high_get_info(rt_size_t *total_nr, rt_size_t *free_nr)
  577. {
  578. int i;
  579. rt_size_t total_free = 0;
  580. rt_base_t level;
  581. level = rt_hw_interrupt_disable();
  582. for (i = 0; i < RT_PAGE_MAX_ORDER; i++)
  583. {
  584. struct rt_page *p = page_list_high[i];
  585. while (p)
  586. {
  587. total_free += (1UL << i);
  588. p = p->next;
  589. }
  590. }
  591. rt_hw_interrupt_enable(level);
  592. *total_nr = _high_pages_nr;
  593. *free_nr = total_free;
  594. }
  595. static void _install_page(rt_page_t mpr_head, rt_region_t region, void *insert_handler)
  596. {
  597. void (*insert)(rt_page_t *page_list, rt_page_t page, int size_bits) = insert_handler;
  598. rt_region_t shadow;
  599. shadow.start = region.start & ~shadow_mask;
  600. shadow.end = FLOOR(region.end, shadow_mask + 1);
  601. if (shadow.end + PV_OFFSET > UINT32_MAX)
  602. _high_page_configured = 1;
  603. rt_page_t shad_head = addr_to_page(mpr_head, (void *)shadow.start);
  604. rt_page_t shad_tail = addr_to_page(mpr_head, (void *)shadow.end);
  605. rt_page_t head = addr_to_page(mpr_head, (void *)region.start);
  606. rt_page_t tail = addr_to_page(mpr_head, (void *)region.end);
  607. /* mark shadow pages as illegal */
  608. for (rt_page_t iter = shad_head; iter < head; iter++)
  609. {
  610. iter->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  611. }
  612. for (rt_page_t iter = tail; iter < shad_tail; iter++)
  613. {
  614. iter->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  615. }
  616. /* insert reserved pages to list */
  617. const int max_order = RT_PAGE_MAX_ORDER + ARCH_PAGE_SHIFT - 1;
  618. while (region.start != region.end)
  619. {
  620. struct rt_page *p;
  621. int align_bits;
  622. int size_bits;
  623. size_bits =
  624. ARCH_ADDRESS_WIDTH_BITS - 1 - rt_hw_clz(region.end - region.start);
  625. align_bits = rt_hw_ctz(region.start);
  626. if (align_bits < size_bits)
  627. {
  628. size_bits = align_bits;
  629. }
  630. if (size_bits > max_order)
  631. {
  632. size_bits = max_order;
  633. }
  634. p = addr_to_page(mpr_head, (void *)region.start);
  635. p->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  636. p->ref_cnt = 0;
  637. /* insert to list */
  638. rt_page_t *page_list = _get_page_list((void *)region.start);
  639. if (page_list == page_list_high)
  640. {
  641. _high_pages_nr += 1 << (size_bits - ARCH_PAGE_SHIFT);
  642. }
  643. insert(page_list, (rt_page_t)((char *)p - early_offset), size_bits - ARCH_PAGE_SHIFT);
  644. region.start += (1UL << size_bits);
  645. }
  646. }
  647. void rt_page_init(rt_region_t reg)
  648. {
  649. int i;
  650. rt_region_t shadow;
  651. /* inclusive start, exclusive end */
  652. reg.start += ARCH_PAGE_MASK;
  653. reg.start &= ~ARCH_PAGE_MASK;
  654. reg.end &= ~ARCH_PAGE_MASK;
  655. if (reg.end <= reg.start)
  656. {
  657. LOG_E("region end(%p) must greater than start(%p)", reg.start, reg.end);
  658. RT_ASSERT(0);
  659. }
  660. page_nr = ((reg.end - reg.start) >> ARCH_PAGE_SHIFT);
  661. shadow.start = reg.start & ~shadow_mask;
  662. shadow.end = FLOOR(reg.end, shadow_mask + 1);
  663. LOG_D("[Init page] start: 0x%lx, end: 0x%lx, total: 0x%lx", reg.start,
  664. reg.end, page_nr);
  665. int err;
  666. /* init free list */
  667. for (i = 0; i < RT_PAGE_MAX_ORDER; i++)
  668. {
  669. page_list_low[i] = 0;
  670. page_list_high[i] = 0;
  671. }
  672. /* map MPR area */
  673. err = rt_aspace_map_static(&rt_kernel_space, &mpr_varea, &rt_mpr_start,
  674. rt_mpr_size, MMU_MAP_K_RWCB, MMF_MAP_FIXED,
  675. &mm_page_mapper, 0);
  676. if (err != RT_EOK)
  677. {
  678. LOG_E("MPR map failed with size %lx at %p", rt_mpr_size, rt_mpr_start);
  679. RT_ASSERT(0);
  680. }
  681. /* calculate footprint */
  682. init_mpr_align_start =
  683. (rt_size_t)addr_to_page(page_start, (void *)shadow.start) &
  684. ~ARCH_PAGE_MASK;
  685. init_mpr_align_end =
  686. FLOOR(addr_to_page(page_start, (void *)shadow.end), ARCH_PAGE_SIZE);
  687. rt_size_t init_mpr_size = init_mpr_align_end - init_mpr_align_start;
  688. rt_size_t init_mpr_npage = init_mpr_size >> ARCH_PAGE_SHIFT;
  689. init_mpr_cont_start = (void *)reg.start;
  690. rt_size_t init_mpr_cont_end = (rt_size_t)init_mpr_cont_start + init_mpr_size;
  691. early_offset = (rt_size_t)init_mpr_cont_start - init_mpr_align_start;
  692. rt_page_t mpr_cont = (void *)((char *)rt_mpr_start + early_offset);
  693. /* mark init mpr pages as illegal */
  694. rt_page_t head_cont = addr_to_page(mpr_cont, (void *)reg.start);
  695. rt_page_t tail_cont = addr_to_page(mpr_cont, (void *)reg.end);
  696. for (rt_page_t iter = head_cont; iter < tail_cont; iter++)
  697. {
  698. iter->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  699. }
  700. reg.start = init_mpr_cont_end;
  701. _install_page(mpr_cont, reg, _early_page_insert);
  702. pages_alloc_handler = _early_pages_alloc;
  703. /* doing the page table bushiness */
  704. if (rt_aspace_load_page(&rt_kernel_space, (void *)init_mpr_align_start, init_mpr_npage))
  705. {
  706. LOG_E("%s: failed to load pages", __func__);
  707. RT_ASSERT(0);
  708. }
  709. if (rt_hw_mmu_tbl_get() == rt_kernel_space.page_table)
  710. rt_page_cleanup();
  711. }
  712. static int _load_mpr_area(void *head, void *tail)
  713. {
  714. int err = 0;
  715. char *iter = (char *)((rt_ubase_t)head & ~ARCH_PAGE_MASK);
  716. tail = (void *)FLOOR(tail, ARCH_PAGE_SIZE);
  717. while (iter != tail)
  718. {
  719. void *paddr = rt_kmem_v2p(iter);
  720. if (paddr == ARCH_MAP_FAILED)
  721. {
  722. err = rt_aspace_load_page(&rt_kernel_space, iter, 1);
  723. if (err != RT_EOK)
  724. {
  725. LOG_E("%s: failed to load page", __func__);
  726. break;
  727. }
  728. }
  729. iter += ARCH_PAGE_SIZE;
  730. }
  731. return err;
  732. }
  733. int rt_page_install(rt_region_t region)
  734. {
  735. int err = -RT_EINVAL;
  736. if (region.end != region.start && !(region.start & ARCH_PAGE_MASK) &&
  737. !(region.end & ARCH_PAGE_MASK) &&
  738. !((region.end - region.start) & shadow_mask))
  739. {
  740. void *head = addr_to_page(page_start, (void *)region.start);
  741. void *tail = addr_to_page(page_start, (void *)region.end);
  742. page_nr += ((region.end - region.start) >> ARCH_PAGE_SHIFT);
  743. err = _load_mpr_area(head, tail);
  744. if (err == RT_EOK)
  745. {
  746. _install_page(rt_mpr_start, region, _page_insert);
  747. }
  748. }
  749. return err;
  750. }
  751. void rt_page_cleanup(void)
  752. {
  753. early_offset = 0;
  754. pages_alloc_handler = _pages_alloc;
  755. }