mm_page.c 23 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893
  1. /*
  2. * Copyright (c) 2006-2022, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Change Logs:
  7. * Date Author Notes
  8. * 2019-11-01 Jesven The first version
  9. * 2022-12-13 WangXiaoyao Hot-pluggable, extensible
  10. * page management algorithm
  11. * 2023-02-20 WangXiaoyao Multi-list page-management
  12. * 2023-11-28 Shell Bugs fix for page_install on shadow region
  13. */
  14. #include <rtthread.h>
  15. #include <stddef.h>
  16. #include <stdint.h>
  17. #include <string.h>
  18. #include "mm_fault.h"
  19. #include "mm_private.h"
  20. #include "mm_aspace.h"
  21. #include "mm_flag.h"
  22. #include "mm_page.h"
  23. #include <mmu.h>
  24. #define DBG_TAG "mm.page"
  25. #define DBG_LVL DBG_WARNING
  26. #include <rtdbg.h>
  27. RT_STATIC_ASSERT(order_huge_pg, RT_PAGE_MAX_ORDER > ARCH_PAGE_SHIFT - 2);
  28. RT_STATIC_ASSERT(size_width, sizeof(rt_size_t) == sizeof(void *));
  29. #ifdef RT_USING_SMART
  30. #include "lwp_arch_comm.h"
  31. #endif /* RT_USING_SMART */
  32. static rt_size_t init_mpr_align_start;
  33. static rt_size_t init_mpr_align_end;
  34. static void *init_mpr_cont_start;
  35. static struct rt_varea mpr_varea;
  36. static struct rt_page *page_list_low[RT_PAGE_MAX_ORDER];
  37. static struct rt_page *page_list_high[RT_PAGE_MAX_ORDER];
  38. static RT_DEFINE_SPINLOCK(_spinlock);
  39. #define page_start ((rt_page_t)rt_mpr_start)
  40. static rt_size_t page_nr;
  41. static rt_size_t _high_pages_nr;
  42. static rt_size_t early_offset;
  43. static const char *get_name(rt_varea_t varea)
  44. {
  45. return "master-page-record";
  46. }
  47. static void hint_free(rt_mm_va_hint_t hint)
  48. {
  49. hint->flags = MMF_MAP_FIXED;
  50. hint->limit_start = rt_kernel_space.start;
  51. hint->limit_range_size = rt_kernel_space.size;
  52. hint->prefer = rt_mpr_start;
  53. }
  54. static void on_page_fault(struct rt_varea *varea, struct rt_aspace_fault_msg *msg)
  55. {
  56. char *init_start = (void *)init_mpr_align_start;
  57. char *init_end = (void *)init_mpr_align_end;
  58. if ((char *)msg->fault_vaddr < init_end && (char *)msg->fault_vaddr >= init_start)
  59. {
  60. rt_size_t offset = (char *)msg->fault_vaddr - init_start;
  61. msg->response.status = MM_FAULT_STATUS_OK;
  62. msg->response.vaddr = (char *)init_mpr_cont_start + offset;
  63. msg->response.size = ARCH_PAGE_SIZE;
  64. }
  65. else
  66. {
  67. rt_mm_dummy_mapper.on_page_fault(varea, msg);
  68. }
  69. }
  70. static struct rt_mem_obj mm_page_mapper = {
  71. .get_name = get_name,
  72. .on_page_fault = on_page_fault,
  73. .hint_free = hint_free,
  74. };
  75. #ifdef RT_DEBUGING_PAGE_LEAK
  76. static volatile int enable;
  77. static rt_page_t _trace_head;
  78. #define TRACE_ALLOC(pg, size) _trace_alloc(pg, __builtin_return_address(0), size)
  79. #define TRACE_FREE(pgaddr, size) _trace_free(pgaddr, __builtin_return_address(0), size)
  80. static long _alloc_cnt;
  81. void rt_page_leak_trace_start()
  82. {
  83. // TODO multicore safety
  84. _trace_head = NULL;
  85. _alloc_cnt = 0;
  86. enable = 1;
  87. }
  88. MSH_CMD_EXPORT(rt_page_leak_trace_start, start page leak tracer);
  89. static void _collect()
  90. {
  91. rt_page_t page = _trace_head;
  92. if (!page)
  93. {
  94. rt_kprintf("ok! ALLOC CNT %ld\n", _alloc_cnt);
  95. }
  96. else
  97. {
  98. while (page)
  99. {
  100. rt_page_t next = page->tl_next;
  101. void *pg_va = rt_page_page2addr(page);
  102. LOG_W("LEAK: %p, allocator: %p, size bits: %lx", pg_va, page->caller, page->trace_size);
  103. rt_pages_free(pg_va, page->trace_size);
  104. page = next;
  105. }
  106. }
  107. }
  108. void rt_page_leak_trace_stop()
  109. {
  110. // TODO multicore safety
  111. enable = 0;
  112. _collect();
  113. }
  114. MSH_CMD_EXPORT(rt_page_leak_trace_stop, stop page leak tracer);
  115. static void _trace_alloc(rt_page_t page, void *caller, size_t size_bits)
  116. {
  117. if (enable)
  118. {
  119. page->caller = caller;
  120. page->trace_size = size_bits;
  121. page->tl_prev = NULL;
  122. page->tl_next = NULL;
  123. _alloc_cnt++;
  124. if (_trace_head == NULL)
  125. {
  126. _trace_head = page;
  127. }
  128. else
  129. {
  130. _trace_head->tl_prev = page;
  131. page->tl_next = _trace_head;
  132. _trace_head = page;
  133. }
  134. }
  135. }
  136. void _report(rt_page_t page, size_t size_bits, char *msg)
  137. {
  138. void *pg_va = rt_page_page2addr(page);
  139. LOG_W("%s: %p, allocator: %p, size bits: %lx", msg, pg_va, page->caller, page->trace_size);
  140. rt_kprintf("backtrace\n");
  141. rt_backtrace();
  142. }
  143. static void _trace_free(rt_page_t page, void *caller, size_t size_bits)
  144. {
  145. if (enable)
  146. {
  147. /* free after free */
  148. if (page->trace_size == 0xabadcafe)
  149. {
  150. _report(page, size_bits, "free after free");
  151. return ;
  152. }
  153. else if (page->trace_size != size_bits)
  154. {
  155. rt_kprintf("free with size bits %lx\n", size_bits);
  156. _report(page, size_bits, "incompatible size bits parameter");
  157. return ;
  158. }
  159. if (page->ref_cnt == 0)
  160. {
  161. _alloc_cnt--;
  162. if (page->tl_prev)
  163. page->tl_prev->tl_next = page->tl_next;
  164. if (page->tl_next)
  165. page->tl_next->tl_prev = page->tl_prev;
  166. if (page == _trace_head)
  167. _trace_head = page->tl_next;
  168. page->tl_prev = NULL;
  169. page->tl_next = NULL;
  170. page->trace_size = 0xabadcafe;
  171. }
  172. }
  173. }
  174. #else
  175. #define TRACE_ALLOC(x, y)
  176. #define TRACE_FREE(x, y)
  177. #endif
  178. static inline void *page_to_addr(rt_page_t page)
  179. {
  180. return (void *)(((page - page_start) << ARCH_PAGE_SHIFT) - PV_OFFSET);
  181. }
  182. static inline rt_page_t addr_to_page(rt_page_t pg_start, void *addr)
  183. {
  184. addr = (char *)addr + PV_OFFSET;
  185. return &pg_start[((rt_ubase_t)addr >> ARCH_PAGE_SHIFT)];
  186. }
  187. #define FLOOR(val, align) (((rt_size_t)(val) + (align)-1) & ~((align)-1))
  188. const rt_size_t shadow_mask =
  189. ((1ul << (RT_PAGE_MAX_ORDER + ARCH_PAGE_SHIFT - 1)) - 1);
  190. const rt_size_t rt_mpr_size = FLOOR(
  191. ((1ul << (ARCH_VADDR_WIDTH - ARCH_PAGE_SHIFT))) * sizeof(struct rt_page),
  192. ARCH_PAGE_SIZE);
  193. void *rt_mpr_start;
  194. rt_weak int rt_hw_clz(unsigned long n)
  195. {
  196. return __builtin_clzl(n);
  197. }
  198. rt_weak int rt_hw_ctz(unsigned long n)
  199. {
  200. return __builtin_ctzl(n);
  201. }
  202. rt_size_t rt_page_bits(rt_size_t size)
  203. {
  204. int bit = sizeof(rt_size_t) * 8 - rt_hw_clz(size) - 1;
  205. if ((size ^ (1UL << bit)) != 0)
  206. {
  207. bit++;
  208. }
  209. bit -= ARCH_PAGE_SHIFT;
  210. if (bit < 0)
  211. {
  212. bit = 0;
  213. }
  214. return bit;
  215. }
  216. struct rt_page *rt_page_addr2page(void *addr)
  217. {
  218. return addr_to_page(page_start, addr);
  219. }
  220. void *rt_page_page2addr(struct rt_page *p)
  221. {
  222. return page_to_addr(p);
  223. }
  224. static inline struct rt_page *_buddy_get(struct rt_page *p,
  225. rt_uint32_t size_bits)
  226. {
  227. rt_size_t addr;
  228. addr = (rt_size_t)rt_page_page2addr(p);
  229. addr ^= (1UL << (size_bits + ARCH_PAGE_SHIFT));
  230. return rt_page_addr2page((void *)addr);
  231. }
  232. static void _page_remove(rt_page_t page_list[], struct rt_page *p, rt_uint32_t size_bits)
  233. {
  234. if (p->pre)
  235. {
  236. p->pre->next = p->next;
  237. }
  238. else
  239. {
  240. page_list[size_bits] = p->next;
  241. }
  242. if (p->next)
  243. {
  244. p->next->pre = p->pre;
  245. }
  246. p->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  247. }
  248. static void _page_insert(rt_page_t page_list[], struct rt_page *p, rt_uint32_t size_bits)
  249. {
  250. p->next = page_list[size_bits];
  251. if (p->next)
  252. {
  253. p->next->pre = p;
  254. }
  255. p->pre = 0;
  256. page_list[size_bits] = p;
  257. p->size_bits = size_bits;
  258. }
  259. static void _pages_ref_inc(struct rt_page *p, rt_uint32_t size_bits)
  260. {
  261. struct rt_page *page_head;
  262. int idx;
  263. /* find page group head */
  264. idx = p - page_start;
  265. idx = idx & ~((1UL << size_bits) - 1);
  266. page_head = page_start + idx;
  267. page_head = (void *)((char *)page_head + early_offset);
  268. page_head->ref_cnt++;
  269. }
  270. static int _pages_ref_get(struct rt_page *p, rt_uint32_t size_bits)
  271. {
  272. struct rt_page *page_head;
  273. int idx;
  274. /* find page group head */
  275. idx = p - page_start;
  276. idx = idx & ~((1UL << size_bits) - 1);
  277. page_head = page_start + idx;
  278. return page_head->ref_cnt;
  279. }
  280. static int _pages_free(rt_page_t page_list[], struct rt_page *p, rt_uint32_t size_bits)
  281. {
  282. rt_uint32_t level = size_bits;
  283. struct rt_page *buddy;
  284. RT_ASSERT(p >= page_start);
  285. RT_ASSERT((char *)p < (char *)rt_mpr_start + rt_mpr_size);
  286. RT_ASSERT(rt_kmem_v2p(p));
  287. RT_ASSERT(p->ref_cnt > 0);
  288. RT_ASSERT(p->size_bits == ARCH_ADDRESS_WIDTH_BITS);
  289. RT_ASSERT(size_bits < RT_PAGE_MAX_ORDER);
  290. p->ref_cnt--;
  291. if (p->ref_cnt != 0)
  292. {
  293. return 0;
  294. }
  295. while (level < RT_PAGE_MAX_ORDER - 1)
  296. {
  297. buddy = _buddy_get(p, level);
  298. if (buddy && buddy->size_bits == level)
  299. {
  300. _page_remove(page_list, buddy, level);
  301. p = (p < buddy) ? p : buddy;
  302. level++;
  303. }
  304. else
  305. {
  306. break;
  307. }
  308. }
  309. _page_insert(page_list, p, level);
  310. return 1;
  311. }
  312. static struct rt_page *_pages_alloc(rt_page_t page_list[], rt_uint32_t size_bits)
  313. {
  314. struct rt_page *p;
  315. if (page_list[size_bits])
  316. {
  317. p = page_list[size_bits];
  318. _page_remove(page_list, p, size_bits);
  319. }
  320. else
  321. {
  322. rt_uint32_t level;
  323. for (level = size_bits + 1; level < RT_PAGE_MAX_ORDER; level++)
  324. {
  325. if (page_list[level])
  326. {
  327. break;
  328. }
  329. }
  330. if (level == RT_PAGE_MAX_ORDER)
  331. {
  332. return 0;
  333. }
  334. p = page_list[level];
  335. _page_remove(page_list, p, level);
  336. while (level > size_bits)
  337. {
  338. _page_insert(page_list, p, level - 1);
  339. p = _buddy_get(p, level - 1);
  340. level--;
  341. }
  342. }
  343. p->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  344. p->ref_cnt = 1;
  345. return p;
  346. }
  347. static void _early_page_remove(rt_page_t page_list[], rt_page_t page, rt_uint32_t size_bits)
  348. {
  349. rt_page_t page_cont = (rt_page_t)((char *)page + early_offset);
  350. if (page_cont->pre)
  351. {
  352. rt_page_t pre_cont = (rt_page_t)((char *)page_cont->pre + early_offset);
  353. pre_cont->next = page_cont->next;
  354. }
  355. else
  356. {
  357. page_list[size_bits] = page_cont->next;
  358. }
  359. if (page_cont->next)
  360. {
  361. rt_page_t next_cont = (rt_page_t)((char *)page_cont->next + early_offset);
  362. next_cont->pre = page_cont->pre;
  363. }
  364. page_cont->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  365. }
  366. static void _early_page_insert(rt_page_t page_list[], rt_page_t page, int size_bits)
  367. {
  368. RT_ASSERT((void *)page >= rt_mpr_start &&
  369. ((char *)page - (char *)rt_mpr_start) < rt_mpr_size);
  370. rt_page_t page_cont = (rt_page_t)((char *)page + early_offset);
  371. page_cont->next = page_list[size_bits];
  372. if (page_cont->next)
  373. {
  374. rt_page_t next_cont = (rt_page_t)((char *)page_cont->next + early_offset);
  375. next_cont->pre = page;
  376. }
  377. page_cont->pre = 0;
  378. page_list[size_bits] = page;
  379. page_cont->size_bits = size_bits;
  380. }
  381. static struct rt_page *_early_pages_alloc(rt_page_t page_list[], rt_uint32_t size_bits)
  382. {
  383. struct rt_page *p;
  384. if (page_list[size_bits])
  385. {
  386. p = page_list[size_bits];
  387. _early_page_remove(page_list, p, size_bits);
  388. }
  389. else
  390. {
  391. rt_uint32_t level;
  392. for (level = size_bits + 1; level < RT_PAGE_MAX_ORDER; level++)
  393. {
  394. if (page_list[level])
  395. {
  396. break;
  397. }
  398. }
  399. if (level == RT_PAGE_MAX_ORDER)
  400. {
  401. return 0;
  402. }
  403. p = page_list[level];
  404. _early_page_remove(page_list, p, level);
  405. while (level > size_bits)
  406. {
  407. _early_page_insert(page_list, p, level - 1);
  408. p = _buddy_get(p, level - 1);
  409. level--;
  410. }
  411. }
  412. rt_page_t page_cont = (rt_page_t)((char *)p + early_offset);
  413. page_cont->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  414. page_cont->ref_cnt = 1;
  415. return p;
  416. }
  417. static rt_page_t *_get_page_list(void *vaddr)
  418. {
  419. rt_ubase_t pa_int = (rt_ubase_t)vaddr + PV_OFFSET;
  420. rt_page_t *list;
  421. if (pa_int > UINT32_MAX)
  422. {
  423. list = page_list_high;
  424. }
  425. else
  426. {
  427. list = page_list_low;
  428. }
  429. return list;
  430. }
  431. int rt_page_ref_get(void *addr, rt_uint32_t size_bits)
  432. {
  433. struct rt_page *p;
  434. rt_base_t level;
  435. int ref;
  436. p = rt_page_addr2page(addr);
  437. level = rt_spin_lock_irqsave(&_spinlock);
  438. ref = _pages_ref_get(p, size_bits);
  439. rt_spin_unlock_irqrestore(&_spinlock, level);
  440. return ref;
  441. }
  442. void rt_page_ref_inc(void *addr, rt_uint32_t size_bits)
  443. {
  444. struct rt_page *p;
  445. rt_base_t level;
  446. p = rt_page_addr2page(addr);
  447. level = rt_spin_lock_irqsave(&_spinlock);
  448. _pages_ref_inc(p, size_bits);
  449. rt_spin_unlock_irqrestore(&_spinlock, level);
  450. }
  451. static rt_page_t (*pages_alloc_handler)(rt_page_t page_list[], rt_uint32_t size_bits);
  452. /* if not, we skip the finding on page_list_high */
  453. static size_t _high_page_configured = 0;
  454. static rt_page_t *_flag_to_page_list(size_t flags)
  455. {
  456. rt_page_t *page_list;
  457. if (_high_page_configured && (flags & PAGE_ANY_AVAILABLE))
  458. {
  459. page_list = page_list_high;
  460. }
  461. else
  462. {
  463. page_list = page_list_low;
  464. }
  465. return page_list;
  466. }
  467. rt_inline void *_do_pages_alloc(rt_uint32_t size_bits, size_t flags)
  468. {
  469. void *alloc_buf = RT_NULL;
  470. struct rt_page *p;
  471. rt_base_t level;
  472. rt_page_t *page_list = _flag_to_page_list(flags);
  473. level = rt_spin_lock_irqsave(&_spinlock);
  474. p = pages_alloc_handler(page_list, size_bits);
  475. rt_spin_unlock_irqrestore(&_spinlock, level);
  476. if (!p && page_list != page_list_low)
  477. {
  478. /* fall back */
  479. page_list = page_list_low;
  480. level = rt_spin_lock_irqsave(&_spinlock);
  481. p = pages_alloc_handler(page_list, size_bits);
  482. rt_spin_unlock_irqrestore(&_spinlock, level);
  483. }
  484. if (p)
  485. {
  486. alloc_buf = page_to_addr(p);
  487. #ifdef RT_DEBUGING_PAGE_LEAK
  488. level = rt_spin_lock_irqsave(&_spinlock);
  489. TRACE_ALLOC(p, size_bits);
  490. rt_spin_unlock_irqrestore(&_spinlock, level);
  491. #endif
  492. }
  493. return alloc_buf;
  494. }
  495. void *rt_pages_alloc(rt_uint32_t size_bits)
  496. {
  497. return _do_pages_alloc(size_bits, 0);
  498. }
  499. void *rt_pages_alloc_ext(rt_uint32_t size_bits, size_t flags)
  500. {
  501. return _do_pages_alloc(size_bits, flags);
  502. }
  503. int rt_pages_free(void *addr, rt_uint32_t size_bits)
  504. {
  505. struct rt_page *p;
  506. rt_page_t *page_list = _get_page_list(addr);
  507. int real_free = 0;
  508. p = rt_page_addr2page(addr);
  509. if (p)
  510. {
  511. rt_base_t level;
  512. level = rt_spin_lock_irqsave(&_spinlock);
  513. real_free = _pages_free(page_list, p, size_bits);
  514. if (real_free)
  515. TRACE_FREE(p, size_bits);
  516. rt_spin_unlock_irqrestore(&_spinlock, level);
  517. }
  518. return real_free;
  519. }
  520. void rt_page_list(void) __attribute__((alias("list_page")));
  521. #define PGNR2SIZE(nr) ((nr) * ARCH_PAGE_SIZE / 1024)
  522. void list_page(void)
  523. {
  524. int i;
  525. rt_size_t free = 0;
  526. rt_size_t installed = page_nr;
  527. rt_base_t level;
  528. level = rt_spin_lock_irqsave(&_spinlock);
  529. for (i = 0; i < RT_PAGE_MAX_ORDER; i++)
  530. {
  531. struct rt_page *lp = page_list_low[i];
  532. struct rt_page *hp = page_list_high[i];
  533. rt_kprintf("level %d ", i);
  534. while (lp)
  535. {
  536. free += (1UL << i);
  537. rt_kprintf("[0x%08p]", rt_page_page2addr(lp));
  538. lp = lp->next;
  539. }
  540. while (hp)
  541. {
  542. free += (1UL << i);
  543. rt_kprintf("[0x%08p]", rt_page_page2addr(hp));
  544. hp = hp->next;
  545. }
  546. rt_kprintf("\n");
  547. }
  548. rt_spin_unlock_irqrestore(&_spinlock, level);
  549. rt_kprintf("-------------------------------\n");
  550. rt_kprintf("Page Summary:\n => free/installed: 0x%lx/0x%lx (%ld/%ld KB)\n", free, installed, PGNR2SIZE(free), PGNR2SIZE(installed));
  551. rt_kprintf("-------------------------------\n");
  552. }
  553. MSH_CMD_EXPORT(list_page, show page info);
  554. void rt_page_get_info(rt_size_t *total_nr, rt_size_t *free_nr)
  555. {
  556. int i;
  557. rt_size_t total_free = 0;
  558. rt_base_t level;
  559. level = rt_spin_lock_irqsave(&_spinlock);
  560. for (i = 0; i < RT_PAGE_MAX_ORDER; i++)
  561. {
  562. struct rt_page *p = page_list_low[i];
  563. while (p)
  564. {
  565. total_free += (1UL << i);
  566. p = p->next;
  567. }
  568. }
  569. for (i = 0; i < RT_PAGE_MAX_ORDER; i++)
  570. {
  571. struct rt_page *p = page_list_high[i];
  572. while (p)
  573. {
  574. total_free += (1UL << i);
  575. p = p->next;
  576. }
  577. }
  578. rt_spin_unlock_irqrestore(&_spinlock, level);
  579. *total_nr = page_nr;
  580. *free_nr = total_free;
  581. }
  582. void rt_page_high_get_info(rt_size_t *total_nr, rt_size_t *free_nr)
  583. {
  584. int i;
  585. rt_size_t total_free = 0;
  586. rt_base_t level;
  587. level = rt_spin_lock_irqsave(&_spinlock);
  588. for (i = 0; i < RT_PAGE_MAX_ORDER; i++)
  589. {
  590. struct rt_page *p = page_list_high[i];
  591. while (p)
  592. {
  593. total_free += (1UL << i);
  594. p = p->next;
  595. }
  596. }
  597. rt_spin_unlock_irqrestore(&_spinlock, level);
  598. *total_nr = _high_pages_nr;
  599. *free_nr = total_free;
  600. }
  601. static void _install_page(rt_page_t mpr_head, rt_region_t region, void *insert_handler)
  602. {
  603. void (*insert)(rt_page_t *page_list, rt_page_t page, int size_bits) = insert_handler;
  604. rt_region_t shadow;
  605. shadow.start = region.start & ~shadow_mask;
  606. shadow.end = FLOOR(region.end, shadow_mask + 1);
  607. if (shadow.end + PV_OFFSET > UINT32_MAX)
  608. _high_page_configured = 1;
  609. rt_page_t shad_head = addr_to_page(mpr_head, (void *)shadow.start);
  610. rt_page_t shad_tail = addr_to_page(mpr_head, (void *)shadow.end);
  611. rt_page_t head = addr_to_page(mpr_head, (void *)region.start);
  612. rt_page_t tail = addr_to_page(mpr_head, (void *)region.end);
  613. /* mark shadow pages as illegal */
  614. for (rt_page_t iter = shad_head; iter < head; iter++)
  615. {
  616. iter->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  617. }
  618. for (rt_page_t iter = tail; iter < shad_tail; iter++)
  619. {
  620. iter->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  621. }
  622. /* insert reserved pages to list */
  623. const int max_order = RT_PAGE_MAX_ORDER + ARCH_PAGE_SHIFT - 1;
  624. while (region.start != region.end)
  625. {
  626. struct rt_page *p;
  627. int align_bits;
  628. int size_bits;
  629. size_bits =
  630. ARCH_ADDRESS_WIDTH_BITS - 1 - rt_hw_clz(region.end - region.start);
  631. align_bits = rt_hw_ctz(region.start);
  632. if (align_bits < size_bits)
  633. {
  634. size_bits = align_bits;
  635. }
  636. if (size_bits > max_order)
  637. {
  638. size_bits = max_order;
  639. }
  640. p = addr_to_page(mpr_head, (void *)region.start);
  641. p->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  642. p->ref_cnt = 0;
  643. /* insert to list */
  644. rt_page_t *page_list = _get_page_list((void *)region.start);
  645. if (page_list == page_list_high)
  646. {
  647. _high_pages_nr += 1 << (size_bits - ARCH_PAGE_SHIFT);
  648. }
  649. insert(page_list, (rt_page_t)((char *)p - early_offset), size_bits - ARCH_PAGE_SHIFT);
  650. region.start += (1UL << size_bits);
  651. }
  652. }
  653. void rt_page_init(rt_region_t reg)
  654. {
  655. int i;
  656. rt_region_t shadow;
  657. /* inclusive start, exclusive end */
  658. reg.start += ARCH_PAGE_MASK;
  659. reg.start &= ~ARCH_PAGE_MASK;
  660. reg.end &= ~ARCH_PAGE_MASK;
  661. if (reg.end <= reg.start)
  662. {
  663. LOG_E("region end(%p) must greater than start(%p)", reg.start, reg.end);
  664. RT_ASSERT(0);
  665. }
  666. page_nr = ((reg.end - reg.start) >> ARCH_PAGE_SHIFT);
  667. shadow.start = reg.start & ~shadow_mask;
  668. shadow.end = FLOOR(reg.end, shadow_mask + 1);
  669. LOG_D("[Init page] start: 0x%lx, end: 0x%lx, total: 0x%lx", reg.start,
  670. reg.end, page_nr);
  671. int err;
  672. /* init free list */
  673. for (i = 0; i < RT_PAGE_MAX_ORDER; i++)
  674. {
  675. page_list_low[i] = 0;
  676. page_list_high[i] = 0;
  677. }
  678. /* map MPR area */
  679. err = rt_aspace_map_static(&rt_kernel_space, &mpr_varea, &rt_mpr_start,
  680. rt_mpr_size, MMU_MAP_K_RWCB, MMF_MAP_FIXED,
  681. &mm_page_mapper, 0);
  682. if (err != RT_EOK)
  683. {
  684. LOG_E("MPR map failed with size %lx at %p", rt_mpr_size, rt_mpr_start);
  685. RT_ASSERT(0);
  686. }
  687. /* calculate footprint */
  688. init_mpr_align_start =
  689. (rt_size_t)addr_to_page(page_start, (void *)shadow.start) &
  690. ~ARCH_PAGE_MASK;
  691. init_mpr_align_end =
  692. FLOOR(addr_to_page(page_start, (void *)shadow.end), ARCH_PAGE_SIZE);
  693. rt_size_t init_mpr_size = init_mpr_align_end - init_mpr_align_start;
  694. rt_size_t init_mpr_npage = init_mpr_size >> ARCH_PAGE_SHIFT;
  695. init_mpr_cont_start = (void *)reg.start;
  696. rt_size_t init_mpr_cont_end = (rt_size_t)init_mpr_cont_start + init_mpr_size;
  697. early_offset = (rt_size_t)init_mpr_cont_start - init_mpr_align_start;
  698. rt_page_t mpr_cont = (void *)((char *)rt_mpr_start + early_offset);
  699. /* mark init mpr pages as illegal */
  700. rt_page_t head_cont = addr_to_page(mpr_cont, (void *)reg.start);
  701. rt_page_t tail_cont = addr_to_page(mpr_cont, (void *)reg.end);
  702. for (rt_page_t iter = head_cont; iter < tail_cont; iter++)
  703. {
  704. iter->size_bits = ARCH_ADDRESS_WIDTH_BITS;
  705. }
  706. reg.start = init_mpr_cont_end;
  707. _install_page(mpr_cont, reg, _early_page_insert);
  708. pages_alloc_handler = _early_pages_alloc;
  709. /* doing the page table bushiness */
  710. if (rt_aspace_load_page(&rt_kernel_space, (void *)init_mpr_align_start, init_mpr_npage))
  711. {
  712. LOG_E("%s: failed to load pages", __func__);
  713. RT_ASSERT(0);
  714. }
  715. if (rt_hw_mmu_tbl_get() == rt_kernel_space.page_table)
  716. rt_page_cleanup();
  717. }
  718. static int _load_mpr_area(void *head, void *tail)
  719. {
  720. int err = 0;
  721. char *iter = (char *)((rt_ubase_t)head & ~ARCH_PAGE_MASK);
  722. tail = (void *)FLOOR(tail, ARCH_PAGE_SIZE);
  723. while (iter != tail)
  724. {
  725. void *paddr = rt_kmem_v2p(iter);
  726. if (paddr == ARCH_MAP_FAILED)
  727. {
  728. err = rt_aspace_load_page(&rt_kernel_space, iter, 1);
  729. if (err != RT_EOK)
  730. {
  731. LOG_E("%s: failed to load page", __func__);
  732. break;
  733. }
  734. }
  735. iter += ARCH_PAGE_SIZE;
  736. }
  737. return err;
  738. }
  739. int rt_page_install(rt_region_t region)
  740. {
  741. int err = -RT_EINVAL;
  742. rt_region_t shadow;
  743. void *head, *tail;
  744. if (region.end != region.start && !(region.start & ARCH_PAGE_MASK) &&
  745. !(region.end & ARCH_PAGE_MASK))
  746. {
  747. shadow.start = region.start & ~shadow_mask;
  748. shadow.end = FLOOR(region.end, shadow_mask + 1);
  749. head = addr_to_page(page_start, (void *)shadow.start);
  750. tail = addr_to_page(page_start, (void *)shadow.end);
  751. page_nr += ((region.end - region.start) >> ARCH_PAGE_SHIFT);
  752. err = _load_mpr_area(head, tail);
  753. if (err == RT_EOK)
  754. {
  755. _install_page(rt_mpr_start, region, _page_insert);
  756. }
  757. }
  758. return err;
  759. }
  760. void rt_page_cleanup(void)
  761. {
  762. early_offset = 0;
  763. pages_alloc_handler = _pages_alloc;
  764. }