mmu.c 15 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641
  1. /*
  2. * Copyright (c) 2006-2021, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Change Logs:
  7. * Date Author Notes
  8. * 2021-01-30 lizhirui first version
  9. */
  10. #include <rtthread.h>
  11. #include <rthw.h>
  12. #include <board.h>
  13. #include <page.h>
  14. #include <stdlib.h>
  15. #include <string.h>
  16. #include <lwp_mm.h>
  17. #include <cache.h>
  18. #define DBG_TAG "mmu"
  19. #define DBG_LVL DBG_INFO
  20. #include <rtdbg.h>
  21. #include "riscv.h"
  22. #include "riscv_mmu.h"
  23. #include "mmu.h"
  24. void *current_mmu_table = RT_NULL;
  25. volatile rt_ubase_t MMUTable[__SIZE(VPN2_BIT)] __attribute__((aligned(4 * 1024)));
  26. static void rt_hw_cpu_tlb_invalidate()
  27. {
  28. rt_size_t satpv = read_csr(satp);
  29. write_csr(satp, satpv);
  30. mmu_flush_tlb();
  31. }
  32. void *rt_hw_mmu_tbl_get()
  33. {
  34. return current_mmu_table;
  35. }
  36. void rt_hw_mmu_switch(void *mmu_table)
  37. {
  38. current_mmu_table = mmu_table;
  39. RT_ASSERT(__CHECKALIGN(mmu_table, PAGE_OFFSET_BIT));
  40. mmu_set_pagetable((rt_ubase_t)mmu_table);
  41. rt_hw_cpu_dcache_clean_all();
  42. rt_hw_cpu_icache_invalidate_all();
  43. }
  44. int rt_hw_mmu_map_init(rt_mmu_info *mmu_info, void *v_address, rt_size_t size, rt_size_t *vtable, rt_size_t pv_off)
  45. {
  46. size_t l1_off, va_s, va_e;
  47. rt_base_t level;
  48. if ((!mmu_info) || (!vtable))
  49. {
  50. return -1;
  51. }
  52. va_s = (rt_size_t)v_address;
  53. va_e = ((rt_size_t)v_address) + size - 1;
  54. if (va_e < va_s)
  55. {
  56. return -1;
  57. }
  58. // convert address to PPN2 index
  59. va_s = GET_L1(va_s);
  60. va_e = GET_L1(va_e);
  61. if (va_s == 0)
  62. {
  63. return -1;
  64. }
  65. rt_mm_lock();
  66. // vtable initialization check
  67. for (l1_off = va_s; l1_off <= va_e; l1_off++)
  68. {
  69. size_t v = vtable[l1_off];
  70. if (v)
  71. {
  72. rt_mm_unlock();
  73. return -1;
  74. }
  75. }
  76. rt_mm_unlock();
  77. mmu_info->vtable = vtable;
  78. mmu_info->vstart = va_s;
  79. mmu_info->vend = va_e;
  80. mmu_info->pv_off = pv_off;
  81. return 0;
  82. }
  83. void rt_hw_mmu_kernel_map_init(rt_mmu_info *mmu_info, rt_size_t vaddr_start, rt_size_t size)
  84. {
  85. rt_size_t paddr_start = __UMASKVALUE(VPN_TO_PPN(vaddr_start, mmu_info->pv_off), PAGE_OFFSET_MASK);
  86. rt_size_t va_s = GET_L1(vaddr_start);
  87. rt_size_t va_e = GET_L1(vaddr_start + size - 1);
  88. rt_size_t i;
  89. for (i = va_s; i <= va_e; i++)
  90. {
  91. mmu_info->vtable[i] = COMBINEPTE(paddr_start, PAGE_ATTR_RWX | PTE_G | PTE_V);
  92. paddr_start += L1_PAGE_SIZE;
  93. }
  94. rt_hw_cpu_tlb_invalidate();
  95. }
  96. // find a range of free virtual address specified by pages
  97. static size_t find_vaddr(rt_mmu_info *mmu_info, int pages)
  98. {
  99. size_t loop_pages;
  100. size_t va;
  101. size_t find_va = 0;
  102. int n = 0;
  103. size_t i;
  104. if (!pages || !mmu_info)
  105. {
  106. return 0;
  107. }
  108. loop_pages = (mmu_info->vend - mmu_info->vstart) ? (mmu_info->vend - mmu_info->vstart) : 1;
  109. loop_pages <<= (ARCH_INDEX_WIDTH * 2);
  110. va = mmu_info->vstart;
  111. va <<= (ARCH_PAGE_SHIFT + ARCH_INDEX_WIDTH * 2);
  112. for (i = 0; i < loop_pages; i++, va += ARCH_PAGE_SIZE)
  113. {
  114. if (_rt_hw_mmu_v2p(mmu_info, (void *)va))
  115. {
  116. n = 0;
  117. find_va = 0;
  118. continue;
  119. }
  120. if (!find_va)
  121. {
  122. find_va = va;
  123. }
  124. n++;
  125. if (n >= pages)
  126. {
  127. return find_va;
  128. }
  129. }
  130. return 0;
  131. }
  132. // check whether the range of virtual address are free
  133. static int check_vaddr(rt_mmu_info *mmu_info, void *va, rt_size_t pages)
  134. {
  135. rt_size_t loop_va = __UMASKVALUE((rt_size_t)va, PAGE_OFFSET_MASK);
  136. rt_size_t l1_off, l2_off, l3_off;
  137. rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
  138. if (!pages)
  139. {
  140. return -1;
  141. }
  142. if (!mmu_info)
  143. {
  144. return -1;
  145. }
  146. while (pages--)
  147. {
  148. l1_off = GET_L1(loop_va);
  149. l2_off = GET_L2(loop_va);
  150. l3_off = GET_L3(loop_va);
  151. mmu_l1 = ((rt_size_t *)mmu_info->vtable) + l1_off;
  152. if (PTE_USED(*mmu_l1))
  153. {
  154. RT_ASSERT(!PAGE_IS_LEAF(*mmu_l1));
  155. mmu_l2 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l1), mmu_info->pv_off) + l2_off;
  156. if (PTE_USED(*mmu_l2))
  157. {
  158. RT_ASSERT(!PAGE_IS_LEAF(*mmu_l2));
  159. mmu_l3 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l2), mmu_info->pv_off) + l3_off;
  160. if (PTE_USED(*mmu_l3))
  161. {
  162. RT_ASSERT(PAGE_IS_LEAF(*mmu_l3));
  163. return -1;
  164. }
  165. }
  166. }
  167. loop_va += PAGE_SIZE;
  168. }
  169. return 0;
  170. }
  171. static void __rt_hw_mmu_unmap(rt_mmu_info *mmu_info, void *v_addr, rt_size_t npages)
  172. {
  173. rt_size_t loop_va = __UMASKVALUE((rt_size_t)v_addr, PAGE_OFFSET_MASK);
  174. rt_size_t l1_off, l2_off, l3_off;
  175. rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
  176. RT_ASSERT(mmu_info);
  177. while (npages--)
  178. {
  179. l1_off = (rt_size_t)GET_L1(loop_va);
  180. RT_ASSERT((l1_off >= mmu_info->vstart) && (l1_off <= mmu_info->vend));
  181. l2_off = (rt_size_t)GET_L2(loop_va);
  182. l3_off = (rt_size_t)GET_L3(loop_va);
  183. mmu_l1 = ((rt_size_t *)mmu_info->vtable) + l1_off;
  184. RT_ASSERT(PTE_USED(*mmu_l1))
  185. RT_ASSERT(!PAGE_IS_LEAF(*mmu_l1));
  186. mmu_l2 = ((rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l1), mmu_info->pv_off)) + l2_off;
  187. RT_ASSERT(PTE_USED(*mmu_l2));
  188. RT_ASSERT(!PAGE_IS_LEAF(*mmu_l2));
  189. mmu_l3 = ((rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l2), mmu_info->pv_off)) + l3_off;
  190. RT_ASSERT(PTE_USED(*mmu_l3));
  191. RT_ASSERT(PAGE_IS_LEAF(*(mmu_l3)));
  192. *mmu_l3 = 0;
  193. rt_hw_cpu_dcache_clean(mmu_l3, sizeof(*mmu_l3));
  194. // decrease reference from leaf page to l3 page
  195. mmu_l3 -= l3_off;
  196. rt_pages_free(mmu_l3, 0);
  197. int free = rt_page_ref_get(mmu_l3, 0);
  198. if (free == 1)
  199. {
  200. // free l3 page
  201. rt_pages_free(mmu_l3, 0);
  202. *mmu_l2 = 0;
  203. rt_hw_cpu_dcache_clean(mmu_l2, sizeof(*mmu_l2));
  204. // decrease reference from l3 page to l2 page
  205. mmu_l2 -= l2_off;
  206. rt_pages_free(mmu_l2, 0);
  207. free = rt_page_ref_get(mmu_l2, 0);
  208. if (free == 1)
  209. {
  210. // free l3 page
  211. rt_pages_free(mmu_l2, 0);
  212. // reset PTE in l1
  213. *mmu_l1 = 0;
  214. rt_hw_cpu_dcache_clean(mmu_l1, sizeof(*mmu_l1));
  215. }
  216. }
  217. loop_va += PAGE_SIZE;
  218. }
  219. }
  220. static int _mmu_map_one_page(rt_mmu_info *mmu_info, size_t va, size_t pa, size_t attr)
  221. {
  222. rt_size_t l1_off, l2_off, l3_off;
  223. rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
  224. l1_off = GET_L1(va);
  225. l2_off = GET_L2(va);
  226. l3_off = GET_L3(va);
  227. mmu_l1 = ((rt_size_t *)mmu_info->vtable) + l1_off;
  228. if (PTE_USED(*mmu_l1))
  229. {
  230. RT_ASSERT(!PAGE_IS_LEAF(*mmu_l1));
  231. mmu_l2 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l1), mmu_info->pv_off);
  232. }
  233. else
  234. {
  235. mmu_l2 = (rt_size_t *)rt_pages_alloc(0);
  236. if (mmu_l2)
  237. {
  238. rt_memset(mmu_l2, 0, PAGE_SIZE);
  239. rt_hw_cpu_dcache_clean(mmu_l2, PAGE_SIZE);
  240. *mmu_l1 = COMBINEPTE((rt_size_t)VPN_TO_PPN(mmu_l2, mmu_info->pv_off), PAGE_DEFAULT_ATTR_NEXT);
  241. rt_hw_cpu_dcache_clean(mmu_l1, sizeof(*mmu_l1));
  242. }
  243. else
  244. {
  245. return -1;
  246. }
  247. }
  248. if (PTE_USED(*(mmu_l2 + l2_off)))
  249. {
  250. RT_ASSERT(!PAGE_IS_LEAF(*(mmu_l2 + l2_off)));
  251. mmu_l3 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*(mmu_l2 + l2_off)), mmu_info->pv_off);
  252. }
  253. else
  254. {
  255. mmu_l3 = (rt_size_t *)rt_pages_alloc(0);
  256. if (mmu_l3)
  257. {
  258. rt_memset(mmu_l3, 0, PAGE_SIZE);
  259. rt_hw_cpu_dcache_clean(mmu_l3, PAGE_SIZE);
  260. *(mmu_l2 + l2_off) = COMBINEPTE((rt_size_t)VPN_TO_PPN(mmu_l3, mmu_info->pv_off), PAGE_DEFAULT_ATTR_NEXT);
  261. rt_hw_cpu_dcache_clean(mmu_l2, sizeof(*mmu_l2));
  262. // declares a reference to parent page table
  263. rt_page_ref_inc((void *)mmu_l2, 0);
  264. }
  265. else
  266. {
  267. return -1;
  268. }
  269. }
  270. RT_ASSERT(!PTE_USED(*(mmu_l3 + l3_off)));
  271. // declares a reference to parent page table
  272. rt_page_ref_inc((void *)mmu_l3, 0);
  273. *(mmu_l3 + l3_off) = COMBINEPTE((rt_size_t)pa, attr);
  274. rt_hw_cpu_dcache_clean(mmu_l3 + l3_off, sizeof(*(mmu_l3 + l3_off)));
  275. return 0;
  276. }
  277. static int __rt_hw_mmu_map(rt_mmu_info *mmu_info, void *v_addr, void *p_addr, rt_size_t npages, rt_size_t attr)
  278. {
  279. rt_size_t loop_va = __UMASKVALUE((rt_size_t)v_addr, PAGE_OFFSET_MASK);
  280. rt_size_t loop_pa = __UMASKVALUE((rt_size_t)p_addr, PAGE_OFFSET_MASK);
  281. if (!mmu_info)
  282. {
  283. return -1;
  284. }
  285. while (npages--)
  286. {
  287. if (_mmu_map_one_page(mmu_info, loop_va, loop_pa, attr) != 0)
  288. {
  289. __rt_hw_mmu_unmap(mmu_info, v_addr, npages);
  290. return -1;
  291. }
  292. loop_va += PAGE_SIZE;
  293. loop_pa += PAGE_SIZE;
  294. }
  295. return 0;
  296. }
  297. void *_rt_hw_mmu_map(rt_mmu_info *mmu_info, void *v_addr, void *p_addr, rt_size_t size, rt_size_t attr)
  298. {
  299. rt_size_t pa_s, pa_e;
  300. rt_size_t vaddr;
  301. rt_size_t pages;
  302. int ret;
  303. if (!size)
  304. {
  305. return 0;
  306. }
  307. pa_s = (rt_size_t)p_addr;
  308. pa_e = ((rt_size_t)p_addr) + size - 1;
  309. pa_s = GET_PF_ID(pa_s);
  310. pa_e = GET_PF_ID(pa_e);
  311. pages = pa_e - pa_s + 1;
  312. if (v_addr)
  313. {
  314. vaddr = (rt_size_t)v_addr;
  315. pa_s = (rt_size_t)p_addr;
  316. if (GET_PF_OFFSET(vaddr) != GET_PF_OFFSET(pa_s))
  317. {
  318. return 0;
  319. }
  320. vaddr = __UMASKVALUE(vaddr, PAGE_OFFSET_MASK);
  321. if (check_vaddr(mmu_info, (void *)vaddr, pages) != 0)
  322. {
  323. return 0;
  324. }
  325. }
  326. else
  327. {
  328. vaddr = find_vaddr(mmu_info, pages);
  329. }
  330. if (vaddr)
  331. {
  332. ret = __rt_hw_mmu_map(mmu_info, (void *)vaddr, p_addr, pages, attr);
  333. if (ret == 0)
  334. {
  335. rt_hw_cpu_tlb_invalidate();
  336. return (void *)(vaddr | GET_PF_OFFSET((rt_size_t)p_addr));
  337. }
  338. }
  339. return 0;
  340. }
  341. static int __rt_hw_mmu_map_auto(rt_mmu_info *mmu_info, void *v_addr, rt_size_t npages, rt_size_t attr)
  342. {
  343. rt_size_t loop_va = __UMASKVALUE((rt_size_t)v_addr, PAGE_OFFSET_MASK);
  344. rt_size_t loop_pa;
  345. rt_size_t l1_off, l2_off, l3_off;
  346. rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
  347. rt_size_t *ref_cnt;
  348. rt_size_t i;
  349. void *va, *pa;
  350. if (!mmu_info)
  351. {
  352. return -1;
  353. }
  354. while (npages--)
  355. {
  356. loop_pa = (rt_size_t)rt_pages_alloc(0);
  357. if (!loop_pa)
  358. {
  359. goto err;
  360. }
  361. if (__rt_hw_mmu_map(mmu_info, (void *)loop_va, (void *)loop_pa, 1, attr) < 0)
  362. {
  363. goto err;
  364. }
  365. loop_va += PAGE_SIZE;
  366. }
  367. return 0;
  368. err:
  369. va = (void *)__UMASKVALUE((rt_size_t)v_addr, PAGE_OFFSET_MASK);
  370. for (i = 0; i < npages; i++)
  371. {
  372. pa = rt_hw_mmu_v2p(mmu_info, va);
  373. if (pa)
  374. {
  375. rt_pages_free((void *)PPN_TO_VPN(pa, mmu_info->pv_off), 0);
  376. }
  377. va = (void *)((rt_uint8_t *)va + PAGE_SIZE);
  378. }
  379. __rt_hw_mmu_unmap(mmu_info, v_addr, npages);
  380. return -1;
  381. }
  382. void *_rt_hw_mmu_map_auto(rt_mmu_info *mmu_info, void *v_addr, rt_size_t size, rt_size_t attr)
  383. {
  384. rt_size_t vaddr;
  385. rt_size_t offset;
  386. rt_size_t pages;
  387. int ret;
  388. if (!size)
  389. {
  390. return 0;
  391. }
  392. offset = GET_PF_OFFSET((rt_size_t)v_addr);
  393. size += (offset + PAGE_SIZE - 1);
  394. pages = size >> PAGE_OFFSET_BIT;
  395. if (v_addr)
  396. {
  397. vaddr = __UMASKVALUE((rt_size_t)v_addr, PAGE_OFFSET_MASK);
  398. if (check_vaddr(mmu_info, (void *)vaddr, pages) != 0)
  399. {
  400. return 0;
  401. }
  402. }
  403. else
  404. {
  405. vaddr = find_vaddr(mmu_info, pages);
  406. }
  407. if (vaddr)
  408. {
  409. ret = __rt_hw_mmu_map_auto(mmu_info, (void *)vaddr, pages, attr);
  410. if (ret == 0)
  411. {
  412. rt_hw_cpu_tlb_invalidate();
  413. return (void *)(vaddr | offset);
  414. }
  415. }
  416. return 0;
  417. }
  418. void _rt_hw_mmu_unmap(rt_mmu_info *mmu_info, void *v_addr, rt_size_t size)
  419. {
  420. rt_size_t va_s, va_e;
  421. rt_size_t pages;
  422. va_s = ((rt_size_t)v_addr) >> PAGE_OFFSET_BIT;
  423. va_e = (((rt_size_t)v_addr) + size - 1) >> PAGE_OFFSET_BIT;
  424. pages = va_e - va_s + 1;
  425. __rt_hw_mmu_unmap(mmu_info, v_addr, pages);
  426. rt_hw_cpu_tlb_invalidate();
  427. }
  428. void *rt_hw_mmu_map(rt_mmu_info *mmu_info, void *v_addr, void *p_addr, rt_size_t size, rt_size_t attr)
  429. {
  430. void *ret;
  431. rt_base_t level;
  432. rt_mm_lock();
  433. ret = _rt_hw_mmu_map(mmu_info, v_addr, p_addr, size, attr);
  434. rt_mm_unlock();
  435. return ret;
  436. }
  437. void *rt_hw_mmu_map_auto(rt_mmu_info *mmu_info, void *v_addr, rt_size_t size, rt_size_t attr)
  438. {
  439. void *ret;
  440. rt_base_t level;
  441. rt_mm_lock();
  442. ret = _rt_hw_mmu_map_auto(mmu_info, v_addr, size, attr);
  443. rt_mm_unlock();
  444. return ret;
  445. }
  446. void rt_hw_mmu_unmap(rt_mmu_info *mmu_info, void *v_addr, rt_size_t size)
  447. {
  448. rt_base_t level;
  449. rt_mm_lock();
  450. _rt_hw_mmu_unmap(mmu_info, v_addr, size);
  451. rt_mm_unlock();
  452. }
  453. void *_rt_hw_mmu_v2p(rt_mmu_info *mmu_info, void *v_addr)
  454. {
  455. rt_size_t l1_off, l2_off, l3_off;
  456. rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
  457. rt_size_t pa;
  458. l1_off = GET_L1((rt_size_t)v_addr);
  459. l2_off = GET_L2((rt_size_t)v_addr);
  460. l3_off = GET_L3((rt_size_t)v_addr);
  461. if (!mmu_info)
  462. {
  463. return RT_NULL;
  464. }
  465. mmu_l1 = ((rt_size_t *)mmu_info->vtable) + l1_off;
  466. if (PTE_USED(*mmu_l1))
  467. {
  468. if (*mmu_l1 & PTE_XWR_MASK)
  469. return (void *)(GET_PADDR(*mmu_l1) | ((rt_size_t)v_addr & ((1 << 30) - 1)));
  470. mmu_l2 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l1), mmu_info->pv_off);
  471. if (PTE_USED(*(mmu_l2 + l2_off)))
  472. {
  473. if (*(mmu_l2 + l2_off) & PTE_XWR_MASK)
  474. return (void *)(GET_PADDR(*(mmu_l2 + l2_off)) | ((rt_size_t)v_addr & ((1 << 21) - 1)));
  475. mmu_l3 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*(mmu_l2 + l2_off)), mmu_info->pv_off);
  476. if (PTE_USED(*(mmu_l3 + l3_off)))
  477. {
  478. return (void *)(GET_PADDR(*(mmu_l3 + l3_off)) | GET_PF_OFFSET((rt_size_t)v_addr));
  479. }
  480. }
  481. }
  482. return RT_NULL;
  483. }
  484. void *rt_hw_mmu_v2p(rt_mmu_info *mmu_info, void *v_addr)
  485. {
  486. void *ret;
  487. rt_base_t level;
  488. rt_mm_lock();
  489. ret = _rt_hw_mmu_v2p(mmu_info, v_addr);
  490. rt_mm_unlock();
  491. return ret;
  492. }
  493. /**
  494. * @brief setup Page Table for kernel space. It's a fixed map
  495. * and all mappings cannot be changed after initialization.
  496. *
  497. * Memory region in struct mem_desc must be page aligned,
  498. * otherwise is a failure and no report will be
  499. * returned.
  500. *
  501. * @param mmu_info
  502. * @param mdesc
  503. * @param desc_nr
  504. */
  505. void rt_hw_mmu_setup(rt_mmu_info *mmu_info, struct mem_desc *mdesc, int desc_nr)
  506. {
  507. void *err;
  508. for (size_t i = 0; i < desc_nr; i++)
  509. {
  510. size_t attr;
  511. switch (mdesc->attr)
  512. {
  513. case NORMAL_MEM:
  514. attr = MMU_MAP_K_RWCB;
  515. break;
  516. case NORMAL_NOCACHE_MEM:
  517. attr = MMU_MAP_K_RWCB;
  518. break;
  519. case DEVICE_MEM:
  520. attr = MMU_MAP_K_DEVICE;
  521. break;
  522. default:
  523. attr = MMU_MAP_K_DEVICE;
  524. }
  525. err = _rt_hw_mmu_map(mmu_info, (void *)mdesc->vaddr_start, (void *)mdesc->paddr_start,
  526. mdesc->vaddr_end - mdesc->vaddr_start + 1, attr);
  527. mdesc++;
  528. }
  529. rt_hw_mmu_switch((void *)MMUTable);
  530. }