mmu.c 12 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505
  1. /*
  2. * Copyright (c) 2006-2021, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Change Logs:
  7. * Date Author Notes
  8. * 2021-01-30 lizhirui first version
  9. */
  10. #include "rtconfig.h"
  11. #include <rtthread.h>
  12. #include <stddef.h>
  13. #include <stdint.h>
  14. #include <cache.h>
  15. #include <mm_aspace.h>
  16. #include <mm_page.h>
  17. #include <mmu.h>
  18. #include <riscv_mmu.h>
  19. #include <tlb.h>
  20. #ifdef RT_USING_SMART
  21. #include <ioremap.h>
  22. #include <lwp_user_mm.h>
  23. #include <tlb.h>
  24. #endif
  25. #define DBG_TAG "MMU"
  26. #define DBG_LVL DBG_LOG
  27. #include <rtdbg.h>
  28. #ifndef RT_USING_SMART
  29. #define PV_OFFSET 0
  30. #define USER_VADDR_START 0
  31. #endif
  32. static size_t _unmap_area(struct rt_aspace *aspace, void *v_addr, size_t size);
  33. void rt_hw_aspace_switch(rt_aspace_t aspace)
  34. {
  35. uintptr_t page_table = (uintptr_t)_rt_kmem_v2p(aspace->page_table);
  36. write_csr(satp, (((size_t)SATP_MODE) << SATP_MODE_OFFSET) |
  37. ((rt_ubase_t)page_table >> PAGE_OFFSET_BIT));
  38. rt_hw_tlb_invalidate_all_local();
  39. }
  40. static void *current_mmu_table = RT_NULL;
  41. volatile __attribute__((aligned(4 * 1024)))
  42. rt_ubase_t MMUTable[__SIZE(VPN2_BIT)];
  43. void *rt_hw_mmu_tbl_get()
  44. {
  45. return current_mmu_table;
  46. }
  47. static int _map_one_page(struct rt_aspace *aspace, void *va, void *pa,
  48. size_t attr)
  49. {
  50. rt_size_t l1_off, l2_off, l3_off;
  51. rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
  52. l1_off = GET_L1((size_t)va);
  53. l2_off = GET_L2((size_t)va);
  54. l3_off = GET_L3((size_t)va);
  55. mmu_l1 = ((rt_size_t *)aspace->page_table) + l1_off;
  56. if (PTE_USED(*mmu_l1))
  57. {
  58. mmu_l2 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l1), PV_OFFSET);
  59. }
  60. else
  61. {
  62. mmu_l2 = (rt_size_t *)rt_pages_alloc(0);
  63. if (mmu_l2)
  64. {
  65. rt_memset(mmu_l2, 0, PAGE_SIZE);
  66. rt_hw_cpu_dcache_clean(mmu_l2, PAGE_SIZE);
  67. *mmu_l1 = COMBINEPTE((rt_size_t)VPN_TO_PPN(mmu_l2, PV_OFFSET),
  68. PAGE_DEFAULT_ATTR_NEXT);
  69. rt_hw_cpu_dcache_clean(mmu_l1, sizeof(*mmu_l1));
  70. }
  71. else
  72. {
  73. return -1;
  74. }
  75. }
  76. if (PTE_USED(*(mmu_l2 + l2_off)))
  77. {
  78. RT_ASSERT(!PAGE_IS_LEAF(*(mmu_l2 + l2_off)));
  79. mmu_l3 =
  80. (rt_size_t *)PPN_TO_VPN(GET_PADDR(*(mmu_l2 + l2_off)), PV_OFFSET);
  81. }
  82. else
  83. {
  84. mmu_l3 = (rt_size_t *)rt_pages_alloc(0);
  85. if (mmu_l3)
  86. {
  87. rt_memset(mmu_l3, 0, PAGE_SIZE);
  88. rt_hw_cpu_dcache_clean(mmu_l3, PAGE_SIZE);
  89. *(mmu_l2 + l2_off) =
  90. COMBINEPTE((rt_size_t)VPN_TO_PPN(mmu_l3, PV_OFFSET),
  91. PAGE_DEFAULT_ATTR_NEXT);
  92. rt_hw_cpu_dcache_clean(mmu_l2, sizeof(*mmu_l2));
  93. // declares a reference to parent page table
  94. rt_page_ref_inc((void *)mmu_l2, 0);
  95. }
  96. else
  97. {
  98. return -1;
  99. }
  100. }
  101. RT_ASSERT(!PTE_USED(*(mmu_l3 + l3_off)));
  102. // declares a reference to parent page table
  103. rt_page_ref_inc((void *)mmu_l3, 0);
  104. *(mmu_l3 + l3_off) = COMBINEPTE((rt_size_t)pa, attr);
  105. rt_hw_cpu_dcache_clean(mmu_l3 + l3_off, sizeof(*(mmu_l3 + l3_off)));
  106. return 0;
  107. }
  108. /** rt_hw_mmu_map will never override existed page table entry */
  109. void *rt_hw_mmu_map(struct rt_aspace *aspace, void *v_addr, void *p_addr,
  110. size_t size, size_t attr)
  111. {
  112. int ret = -1;
  113. void *unmap_va = v_addr;
  114. size_t npages = size >> ARCH_PAGE_SHIFT;
  115. // TODO trying with HUGEPAGE here
  116. while (npages--)
  117. {
  118. ret = _map_one_page(aspace, v_addr, p_addr, attr);
  119. if (ret != 0)
  120. {
  121. /* error, undo map */
  122. while (unmap_va != v_addr)
  123. {
  124. MM_PGTBL_LOCK(aspace);
  125. _unmap_area(aspace, unmap_va, ARCH_PAGE_SIZE);
  126. MM_PGTBL_UNLOCK(aspace);
  127. unmap_va += ARCH_PAGE_SIZE;
  128. }
  129. break;
  130. }
  131. v_addr += ARCH_PAGE_SIZE;
  132. p_addr += ARCH_PAGE_SIZE;
  133. }
  134. if (ret == 0)
  135. {
  136. return unmap_va;
  137. }
  138. return NULL;
  139. }
  140. static void _unmap_pte(rt_size_t *pentry, rt_size_t *lvl_entry[], int level)
  141. {
  142. int loop_flag = 1;
  143. while (loop_flag)
  144. {
  145. loop_flag = 0;
  146. *pentry = 0;
  147. rt_hw_cpu_dcache_clean(pentry, sizeof(*pentry));
  148. // we don't handle level 0, which is maintained by caller
  149. if (level > 0)
  150. {
  151. void *page = (void *)((rt_ubase_t)pentry & ~ARCH_PAGE_MASK);
  152. // decrease reference from child page to parent
  153. rt_pages_free(page, 0);
  154. int free = rt_page_ref_get(page, 0);
  155. if (free == 1)
  156. {
  157. rt_pages_free(page, 0);
  158. pentry = lvl_entry[--level];
  159. loop_flag = 1;
  160. }
  161. }
  162. }
  163. }
  164. static size_t _unmap_area(struct rt_aspace *aspace, void *v_addr, size_t size)
  165. {
  166. rt_size_t loop_va = __UMASKVALUE((rt_size_t)v_addr, PAGE_OFFSET_MASK);
  167. size_t unmapped = 0;
  168. int i = 0;
  169. rt_size_t lvl_off[3];
  170. rt_size_t *lvl_entry[3];
  171. lvl_off[0] = (rt_size_t)GET_L1(loop_va);
  172. lvl_off[1] = (rt_size_t)GET_L2(loop_va);
  173. lvl_off[2] = (rt_size_t)GET_L3(loop_va);
  174. unmapped = 1 << (ARCH_PAGE_SHIFT + ARCH_INDEX_WIDTH * 2ul);
  175. rt_size_t *pentry;
  176. lvl_entry[i] = ((rt_size_t *)aspace->page_table + lvl_off[i]);
  177. pentry = lvl_entry[i];
  178. // find leaf page table entry
  179. while (PTE_USED(*pentry) && !PAGE_IS_LEAF(*pentry))
  180. {
  181. i += 1;
  182. lvl_entry[i] = ((rt_size_t *)PPN_TO_VPN(GET_PADDR(*pentry), PV_OFFSET) +
  183. lvl_off[i]);
  184. pentry = lvl_entry[i];
  185. unmapped >>= ARCH_INDEX_WIDTH;
  186. }
  187. // clear PTE & setup its
  188. if (PTE_USED(*pentry))
  189. {
  190. _unmap_pte(pentry, lvl_entry, i);
  191. }
  192. return unmapped;
  193. }
  194. /** unmap is different from map that it can handle multiple pages */
  195. void rt_hw_mmu_unmap(struct rt_aspace *aspace, void *v_addr, size_t size)
  196. {
  197. // caller guarantee that v_addr & size are page aligned
  198. if (!aspace->page_table)
  199. {
  200. return;
  201. }
  202. size_t unmapped = 0;
  203. while (size > 0)
  204. {
  205. MM_PGTBL_LOCK(aspace);
  206. unmapped = _unmap_area(aspace, v_addr, size);
  207. MM_PGTBL_UNLOCK(aspace);
  208. // when unmapped == 0, region not exist in pgtbl
  209. if (!unmapped || unmapped > size)
  210. break;
  211. size -= unmapped;
  212. v_addr += unmapped;
  213. }
  214. }
  215. #ifdef RT_USING_SMART
  216. static inline void _init_region(void *vaddr, size_t size)
  217. {
  218. rt_ioremap_start = vaddr;
  219. rt_ioremap_size = size;
  220. rt_mpr_start = rt_ioremap_start - rt_mpr_size;
  221. rt_kprintf("rt_ioremap_start: %p, rt_mpr_start: %p\n", rt_ioremap_start, rt_mpr_start);
  222. }
  223. #else
  224. static inline void _init_region(void *vaddr, size_t size)
  225. {
  226. rt_mpr_start = vaddr - rt_mpr_size;
  227. }
  228. #endif
  229. int rt_hw_mmu_map_init(rt_aspace_t aspace, void *v_address, rt_size_t size,
  230. rt_size_t *vtable, rt_size_t pv_off)
  231. {
  232. size_t l1_off, va_s, va_e;
  233. rt_base_t level;
  234. if ((!aspace) || (!vtable))
  235. {
  236. return -1;
  237. }
  238. va_s = (rt_size_t)v_address;
  239. va_e = ((rt_size_t)v_address) + size - 1;
  240. if (va_e < va_s)
  241. {
  242. return -1;
  243. }
  244. // convert address to PPN2 index
  245. va_s = GET_L1(va_s);
  246. va_e = GET_L1(va_e);
  247. if (va_s == 0)
  248. {
  249. return -1;
  250. }
  251. // vtable initialization check
  252. for (l1_off = va_s; l1_off <= va_e; l1_off++)
  253. {
  254. size_t v = vtable[l1_off];
  255. if (v)
  256. {
  257. return -1;
  258. }
  259. }
  260. rt_aspace_init(&rt_kernel_space, (void *)0x1000, USER_VADDR_START - 0x1000,
  261. vtable);
  262. _init_region(v_address, size);
  263. return 0;
  264. }
  265. const static int max_level =
  266. (ARCH_VADDR_WIDTH - ARCH_PAGE_SHIFT) / ARCH_INDEX_WIDTH;
  267. static inline uintptr_t _get_level_size(int level)
  268. {
  269. return 1ul << (ARCH_PAGE_SHIFT + (max_level - level) * ARCH_INDEX_WIDTH);
  270. }
  271. static rt_size_t *_query(struct rt_aspace *aspace, void *vaddr, int *level)
  272. {
  273. rt_size_t l1_off, l2_off, l3_off;
  274. rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
  275. rt_size_t pa;
  276. l1_off = GET_L1((rt_size_t)vaddr);
  277. l2_off = GET_L2((rt_size_t)vaddr);
  278. l3_off = GET_L3((rt_size_t)vaddr);
  279. if (!aspace)
  280. {
  281. LOG_W("%s: no aspace", __func__);
  282. return RT_NULL;
  283. }
  284. mmu_l1 = ((rt_size_t *)aspace->page_table) + l1_off;
  285. if (PTE_USED(*mmu_l1))
  286. {
  287. if (*mmu_l1 & PTE_XWR_MASK)
  288. {
  289. *level = 1;
  290. return mmu_l1;
  291. }
  292. mmu_l2 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l1), PV_OFFSET);
  293. if (PTE_USED(*(mmu_l2 + l2_off)))
  294. {
  295. if (*(mmu_l2 + l2_off) & PTE_XWR_MASK)
  296. {
  297. *level = 2;
  298. return mmu_l2 + l2_off;
  299. }
  300. mmu_l3 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*(mmu_l2 + l2_off)),
  301. PV_OFFSET);
  302. if (PTE_USED(*(mmu_l3 + l3_off)))
  303. {
  304. *level = 3;
  305. return mmu_l3 + l3_off;
  306. }
  307. }
  308. }
  309. return RT_NULL;
  310. }
  311. void *rt_hw_mmu_v2p(struct rt_aspace *aspace, void *vaddr)
  312. {
  313. int level;
  314. uintptr_t *pte = _query(aspace, vaddr, &level);
  315. uintptr_t paddr;
  316. if (pte)
  317. {
  318. paddr = GET_PADDR(*pte);
  319. paddr |= ((intptr_t)vaddr & (_get_level_size(level) - 1));
  320. }
  321. else
  322. {
  323. paddr = 0;
  324. }
  325. return (void *)paddr;
  326. }
  327. static int _noncache(uintptr_t *pte)
  328. {
  329. return 0;
  330. }
  331. static int _cache(uintptr_t *pte)
  332. {
  333. return 0;
  334. }
  335. static int (*control_handler[MMU_CNTL_DUMMY_END])(uintptr_t *pte) = {
  336. [MMU_CNTL_CACHE] = _cache,
  337. [MMU_CNTL_NONCACHE] = _noncache,
  338. };
  339. int rt_hw_mmu_control(struct rt_aspace *aspace, void *vaddr, size_t size,
  340. enum rt_mmu_cntl cmd)
  341. {
  342. int level;
  343. int err = -RT_EINVAL;
  344. void *vend = vaddr + size;
  345. int (*handler)(uintptr_t * pte);
  346. if (cmd >= 0 && cmd < MMU_CNTL_DUMMY_END)
  347. {
  348. handler = control_handler[cmd];
  349. while (vaddr < vend)
  350. {
  351. uintptr_t *pte = _query(aspace, vaddr, &level);
  352. void *range_end = vaddr + _get_level_size(level);
  353. RT_ASSERT(range_end < vend);
  354. if (pte)
  355. {
  356. err = handler(pte);
  357. RT_ASSERT(err == RT_EOK);
  358. }
  359. vaddr = range_end;
  360. }
  361. }
  362. else
  363. {
  364. err = -RT_ENOSYS;
  365. }
  366. return err;
  367. }
  368. /**
  369. * @brief setup Page Table for kernel space. It's a fixed map
  370. * and all mappings cannot be changed after initialization.
  371. *
  372. * Memory region in struct mem_desc must be page aligned,
  373. * otherwise is a failure and no report will be
  374. * returned.
  375. *
  376. * @param mmu_info
  377. * @param mdesc
  378. * @param desc_nr
  379. */
  380. void rt_hw_mmu_setup(rt_aspace_t aspace, struct mem_desc *mdesc, int desc_nr)
  381. {
  382. void *err;
  383. for (size_t i = 0; i < desc_nr; i++)
  384. {
  385. size_t attr;
  386. switch (mdesc->attr)
  387. {
  388. case NORMAL_MEM:
  389. attr = MMU_MAP_K_RWCB;
  390. break;
  391. case NORMAL_NOCACHE_MEM:
  392. attr = MMU_MAP_K_RWCB;
  393. break;
  394. case DEVICE_MEM:
  395. attr = MMU_MAP_K_DEVICE;
  396. break;
  397. default:
  398. attr = MMU_MAP_K_DEVICE;
  399. }
  400. struct rt_mm_va_hint hint = {.flags = MMF_MAP_FIXED,
  401. .limit_start = aspace->start,
  402. .limit_range_size = aspace->size,
  403. .map_size = mdesc->vaddr_end -
  404. mdesc->vaddr_start + 1,
  405. .prefer = (void *)mdesc->vaddr_start};
  406. rt_aspace_map_phy_static(aspace, &mdesc->varea, &hint, attr,
  407. mdesc->paddr_start >> MM_PAGE_SHIFT, &err);
  408. mdesc++;
  409. }
  410. rt_hw_aspace_switch(&rt_kernel_space);
  411. rt_page_cleanup();
  412. }
  413. void rt_hw_mmu_kernel_map_init(rt_aspace_t aspace, rt_size_t vaddr_start, rt_size_t size)
  414. {
  415. rt_size_t paddr_start =
  416. __UMASKVALUE(VPN_TO_PPN(vaddr_start, PV_OFFSET), PAGE_OFFSET_MASK);
  417. rt_size_t va_s = GET_L1(vaddr_start);
  418. rt_size_t va_e = GET_L1(vaddr_start + size - 1);
  419. rt_size_t i;
  420. for (i = va_s; i <= va_e; i++)
  421. {
  422. MMUTable[i] =
  423. COMBINEPTE(paddr_start, PAGE_ATTR_RWX | PTE_G | PTE_V | PTE_CACHE |
  424. PTE_SHARE | PTE_BUF | PTE_A | PTE_D);
  425. paddr_start += L1_PAGE_SIZE;
  426. }
  427. rt_hw_tlb_invalidate_all_local();
  428. }