mmu.c 15 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589
  1. /*
  2. * Copyright (c) 2006-2021, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Change Logs:
  7. * Date Author Notes
  8. * 2021-01-30 lizhirui first version
  9. * 2022-12-13 WangXiaoyao Port to new mm
  10. * 2023-10-12 Shell Add permission control API
  11. */
  12. #include <rtthread.h>
  13. #include <stddef.h>
  14. #include <stdint.h>
  15. #define DBG_TAG "hw.mmu"
  16. #define DBG_LVL DBG_WARNING
  17. #include <rtdbg.h>
  18. #include <cache.h>
  19. #include <mm_aspace.h>
  20. #include <mm_page.h>
  21. #include <mmu.h>
  22. #include <riscv_mmu.h>
  23. #include <tlb.h>
  24. #ifdef RT_USING_SMART
  25. #include <board.h>
  26. #include <ioremap.h>
  27. #include <lwp_user_mm.h>
  28. #endif
  29. #ifndef RT_USING_SMART
  30. #define USER_VADDR_START 0
  31. #endif
  32. static size_t _unmap_area(struct rt_aspace *aspace, void *v_addr, size_t size);
  33. static void *current_mmu_table = RT_NULL;
  34. volatile __attribute__((aligned(4 * 1024)))
  35. rt_ubase_t MMUTable[__SIZE(VPN2_BIT)];
  36. static rt_uint8_t ASID_BITS = 0;
  37. static rt_uint32_t next_asid;
  38. static rt_uint64_t global_asid_generation;
  39. #define ASID_MASK ((1 << ASID_BITS) - 1)
  40. #define ASID_FIRST_GENERATION (1 << ASID_BITS)
  41. #define MAX_ASID ASID_FIRST_GENERATION
  42. static void _asid_init()
  43. {
  44. unsigned int satp_reg = read_csr(satp);
  45. satp_reg |= (((rt_uint64_t)0xffff) << PPN_BITS);
  46. write_csr(satp, satp_reg);
  47. unsigned short valid_asid_bit = ((read_csr(satp) >> PPN_BITS) & 0xffff);
  48. // The maximal value of ASIDLEN, is 9 for Sv32 or 16 for Sv39, Sv48, and Sv57
  49. for (unsigned i = 0; i < 16; i++)
  50. {
  51. if (!(valid_asid_bit & 0x1))
  52. {
  53. break;
  54. }
  55. valid_asid_bit >>= 1;
  56. ASID_BITS++;
  57. }
  58. global_asid_generation = ASID_FIRST_GENERATION;
  59. next_asid = 1;
  60. }
  61. static rt_uint64_t _asid_check_switch(rt_aspace_t aspace)
  62. {
  63. if ((aspace->asid ^ global_asid_generation) >> ASID_BITS) // not same generation
  64. {
  65. if (next_asid != MAX_ASID)
  66. {
  67. aspace->asid = global_asid_generation | next_asid;
  68. next_asid++;
  69. }
  70. else
  71. {
  72. // scroll to next generation
  73. global_asid_generation += ASID_FIRST_GENERATION;
  74. next_asid = 1;
  75. rt_hw_tlb_invalidate_all_local();
  76. aspace->asid = global_asid_generation | next_asid;
  77. next_asid++;
  78. }
  79. }
  80. return aspace->asid & ASID_MASK;
  81. }
  82. void rt_hw_aspace_switch(rt_aspace_t aspace)
  83. {
  84. uintptr_t page_table = (uintptr_t)rt_kmem_v2p(aspace->page_table);
  85. current_mmu_table = aspace->page_table;
  86. rt_uint64_t asid = _asid_check_switch(aspace);
  87. write_csr(satp, (((size_t)SATP_MODE) << SATP_MODE_OFFSET) |
  88. (asid << PPN_BITS) |
  89. ((rt_ubase_t)page_table >> PAGE_OFFSET_BIT));
  90. asm volatile("sfence.vma x0,%0"::"r"(asid):"memory");
  91. }
  92. void *rt_hw_mmu_tbl_get()
  93. {
  94. return current_mmu_table;
  95. }
  96. static int _map_one_page(struct rt_aspace *aspace, void *va, void *pa,
  97. size_t attr)
  98. {
  99. rt_size_t l1_off, l2_off, l3_off;
  100. rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
  101. l1_off = GET_L1((size_t)va);
  102. l2_off = GET_L2((size_t)va);
  103. l3_off = GET_L3((size_t)va);
  104. mmu_l1 = ((rt_size_t *)aspace->page_table) + l1_off;
  105. if (PTE_USED(*mmu_l1))
  106. {
  107. mmu_l2 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l1), PV_OFFSET);
  108. }
  109. else
  110. {
  111. mmu_l2 = (rt_size_t *)rt_pages_alloc(0);
  112. if (mmu_l2)
  113. {
  114. rt_memset(mmu_l2, 0, PAGE_SIZE);
  115. rt_hw_cpu_dcache_clean(mmu_l2, PAGE_SIZE);
  116. *mmu_l1 = COMBINEPTE((rt_size_t)VPN_TO_PPN(mmu_l2, PV_OFFSET),
  117. PAGE_DEFAULT_ATTR_NEXT);
  118. rt_hw_cpu_dcache_clean(mmu_l1, sizeof(*mmu_l1));
  119. }
  120. else
  121. {
  122. return -1;
  123. }
  124. }
  125. if (PTE_USED(*(mmu_l2 + l2_off)))
  126. {
  127. RT_ASSERT(!PAGE_IS_LEAF(*(mmu_l2 + l2_off)));
  128. mmu_l3 =
  129. (rt_size_t *)PPN_TO_VPN(GET_PADDR(*(mmu_l2 + l2_off)), PV_OFFSET);
  130. }
  131. else
  132. {
  133. mmu_l3 = (rt_size_t *)rt_pages_alloc(0);
  134. if (mmu_l3)
  135. {
  136. rt_memset(mmu_l3, 0, PAGE_SIZE);
  137. rt_hw_cpu_dcache_clean(mmu_l3, PAGE_SIZE);
  138. *(mmu_l2 + l2_off) =
  139. COMBINEPTE((rt_size_t)VPN_TO_PPN(mmu_l3, PV_OFFSET),
  140. PAGE_DEFAULT_ATTR_NEXT);
  141. rt_hw_cpu_dcache_clean(mmu_l2, sizeof(*mmu_l2));
  142. // declares a reference to parent page table
  143. rt_page_ref_inc((void *)mmu_l2, 0);
  144. }
  145. else
  146. {
  147. return -1;
  148. }
  149. }
  150. RT_ASSERT(!PTE_USED(*(mmu_l3 + l3_off)));
  151. // declares a reference to parent page table
  152. rt_page_ref_inc((void *)mmu_l3, 0);
  153. *(mmu_l3 + l3_off) = COMBINEPTE((rt_size_t)pa, attr);
  154. rt_hw_cpu_dcache_clean(mmu_l3 + l3_off, sizeof(*(mmu_l3 + l3_off)));
  155. return 0;
  156. }
  157. /** rt_hw_mmu_map will never override existed page table entry */
  158. void *rt_hw_mmu_map(struct rt_aspace *aspace, void *v_addr, void *p_addr,
  159. size_t size, size_t attr)
  160. {
  161. int ret = -1;
  162. void *unmap_va = v_addr;
  163. size_t npages = size >> ARCH_PAGE_SHIFT;
  164. // TODO trying with HUGEPAGE here
  165. while (npages--)
  166. {
  167. MM_PGTBL_LOCK(aspace);
  168. ret = _map_one_page(aspace, v_addr, p_addr, attr);
  169. MM_PGTBL_UNLOCK(aspace);
  170. if (ret != 0)
  171. {
  172. /* error, undo map */
  173. while (unmap_va != v_addr)
  174. {
  175. MM_PGTBL_LOCK(aspace);
  176. _unmap_area(aspace, unmap_va, ARCH_PAGE_SIZE);
  177. MM_PGTBL_UNLOCK(aspace);
  178. unmap_va += ARCH_PAGE_SIZE;
  179. }
  180. break;
  181. }
  182. v_addr += ARCH_PAGE_SIZE;
  183. p_addr += ARCH_PAGE_SIZE;
  184. }
  185. if (ret == 0)
  186. {
  187. return unmap_va;
  188. }
  189. return NULL;
  190. }
  191. static void _unmap_pte(rt_size_t *pentry, rt_size_t *lvl_entry[], int level)
  192. {
  193. int loop_flag = 1;
  194. while (loop_flag)
  195. {
  196. loop_flag = 0;
  197. *pentry = 0;
  198. rt_hw_cpu_dcache_clean(pentry, sizeof(*pentry));
  199. // we don't handle level 0, which is maintained by caller
  200. if (level > 0)
  201. {
  202. void *page = (void *)((rt_ubase_t)pentry & ~ARCH_PAGE_MASK);
  203. // decrease reference from child page to parent
  204. rt_pages_free(page, 0);
  205. int free = rt_page_ref_get(page, 0);
  206. if (free == 1)
  207. {
  208. rt_pages_free(page, 0);
  209. pentry = lvl_entry[--level];
  210. loop_flag = 1;
  211. }
  212. }
  213. }
  214. }
  215. static size_t _unmap_area(struct rt_aspace *aspace, void *v_addr, size_t size)
  216. {
  217. rt_size_t loop_va = __UMASKVALUE((rt_size_t)v_addr, PAGE_OFFSET_MASK);
  218. size_t unmapped = 0;
  219. int i = 0;
  220. rt_size_t lvl_off[3];
  221. rt_size_t *lvl_entry[3];
  222. lvl_off[0] = (rt_size_t)GET_L1(loop_va);
  223. lvl_off[1] = (rt_size_t)GET_L2(loop_va);
  224. lvl_off[2] = (rt_size_t)GET_L3(loop_va);
  225. unmapped = 1 << (ARCH_PAGE_SHIFT + ARCH_INDEX_WIDTH * 2ul);
  226. rt_size_t *pentry;
  227. lvl_entry[i] = ((rt_size_t *)aspace->page_table + lvl_off[i]);
  228. pentry = lvl_entry[i];
  229. // find leaf page table entry
  230. while (PTE_USED(*pentry) && !PAGE_IS_LEAF(*pentry))
  231. {
  232. i += 1;
  233. lvl_entry[i] = ((rt_size_t *)PPN_TO_VPN(GET_PADDR(*pentry), PV_OFFSET) +
  234. lvl_off[i]);
  235. pentry = lvl_entry[i];
  236. unmapped >>= ARCH_INDEX_WIDTH;
  237. }
  238. // clear PTE & setup its
  239. if (PTE_USED(*pentry))
  240. {
  241. _unmap_pte(pentry, lvl_entry, i);
  242. }
  243. return unmapped;
  244. }
  245. /** unmap is different from map that it can handle multiple pages */
  246. void rt_hw_mmu_unmap(struct rt_aspace *aspace, void *v_addr, size_t size)
  247. {
  248. // caller guarantee that v_addr & size are page aligned
  249. if (!aspace->page_table)
  250. {
  251. return;
  252. }
  253. size_t unmapped = 0;
  254. while (size > 0)
  255. {
  256. MM_PGTBL_LOCK(aspace);
  257. unmapped = _unmap_area(aspace, v_addr, size);
  258. MM_PGTBL_UNLOCK(aspace);
  259. // when unmapped == 0, region not exist in pgtbl
  260. if (!unmapped || unmapped > size)
  261. break;
  262. size -= unmapped;
  263. v_addr += unmapped;
  264. }
  265. }
  266. #ifdef RT_USING_SMART
  267. static inline void _init_region(void *vaddr, size_t size)
  268. {
  269. rt_ioremap_start = vaddr;
  270. rt_ioremap_size = size;
  271. rt_mpr_start = rt_ioremap_start - rt_mpr_size;
  272. LOG_D("rt_ioremap_start: %p, rt_mpr_start: %p", rt_ioremap_start, rt_mpr_start);
  273. }
  274. #else
  275. static inline void _init_region(void *vaddr, size_t size)
  276. {
  277. rt_mpr_start = vaddr - rt_mpr_size;
  278. }
  279. #endif
  280. int rt_hw_mmu_map_init(rt_aspace_t aspace, void *v_address, rt_size_t size,
  281. rt_size_t *vtable, rt_size_t pv_off)
  282. {
  283. size_t l1_off, va_s, va_e;
  284. rt_base_t level;
  285. if ((!aspace) || (!vtable))
  286. {
  287. return -1;
  288. }
  289. va_s = (rt_size_t)v_address;
  290. va_e = ((rt_size_t)v_address) + size - 1;
  291. if (va_e < va_s)
  292. {
  293. return -1;
  294. }
  295. // convert address to PPN2 index
  296. va_s = GET_L1(va_s);
  297. va_e = GET_L1(va_e);
  298. if (va_s == 0)
  299. {
  300. return -1;
  301. }
  302. // vtable initialization check
  303. for (l1_off = va_s; l1_off <= va_e; l1_off++)
  304. {
  305. size_t v = vtable[l1_off];
  306. if (v)
  307. {
  308. return -1;
  309. }
  310. }
  311. rt_aspace_init(&rt_kernel_space, (void *)0x1000, USER_VADDR_START - 0x1000,
  312. vtable);
  313. _init_region(v_address, size);
  314. return 0;
  315. }
  316. const static int max_level =
  317. (ARCH_VADDR_WIDTH - ARCH_PAGE_SHIFT) / ARCH_INDEX_WIDTH;
  318. static inline uintptr_t _get_level_size(int level)
  319. {
  320. return 1ul << (ARCH_PAGE_SHIFT + (max_level - level) * ARCH_INDEX_WIDTH);
  321. }
  322. static rt_size_t *_query(struct rt_aspace *aspace, void *vaddr, int *level)
  323. {
  324. rt_size_t l1_off, l2_off, l3_off;
  325. rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
  326. rt_size_t pa;
  327. l1_off = GET_L1((rt_size_t)vaddr);
  328. l2_off = GET_L2((rt_size_t)vaddr);
  329. l3_off = GET_L3((rt_size_t)vaddr);
  330. if (!aspace)
  331. {
  332. LOG_W("%s: no aspace", __func__);
  333. return RT_NULL;
  334. }
  335. mmu_l1 = ((rt_size_t *)aspace->page_table) + l1_off;
  336. if (PTE_USED(*mmu_l1))
  337. {
  338. if (*mmu_l1 & PTE_XWR_MASK)
  339. {
  340. *level = 1;
  341. return mmu_l1;
  342. }
  343. mmu_l2 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l1), PV_OFFSET);
  344. if (PTE_USED(*(mmu_l2 + l2_off)))
  345. {
  346. if (*(mmu_l2 + l2_off) & PTE_XWR_MASK)
  347. {
  348. *level = 2;
  349. return mmu_l2 + l2_off;
  350. }
  351. mmu_l3 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*(mmu_l2 + l2_off)),
  352. PV_OFFSET);
  353. if (PTE_USED(*(mmu_l3 + l3_off)))
  354. {
  355. *level = 3;
  356. return mmu_l3 + l3_off;
  357. }
  358. }
  359. }
  360. return RT_NULL;
  361. }
  362. void *rt_hw_mmu_v2p(struct rt_aspace *aspace, void *vaddr)
  363. {
  364. int level;
  365. uintptr_t *pte = _query(aspace, vaddr, &level);
  366. uintptr_t paddr;
  367. if (pte)
  368. {
  369. paddr = GET_PADDR(*pte);
  370. paddr |= ((intptr_t)vaddr & (_get_level_size(level) - 1));
  371. }
  372. else
  373. {
  374. LOG_I("%s: failed at %p", __func__, vaddr);
  375. paddr = (uintptr_t)ARCH_MAP_FAILED;
  376. }
  377. return (void *)paddr;
  378. }
  379. static int _noncache(uintptr_t *pte)
  380. {
  381. return 0;
  382. }
  383. static int _cache(uintptr_t *pte)
  384. {
  385. return 0;
  386. }
  387. static int (*control_handler[MMU_CNTL_DUMMY_END])(uintptr_t *pte) = {
  388. [MMU_CNTL_CACHE] = _cache,
  389. [MMU_CNTL_NONCACHE] = _noncache,
  390. };
  391. int rt_hw_mmu_control(struct rt_aspace *aspace, void *vaddr, size_t size,
  392. enum rt_mmu_cntl cmd)
  393. {
  394. int level;
  395. int err = -RT_EINVAL;
  396. void *vend = vaddr + size;
  397. int (*handler)(uintptr_t * pte);
  398. if (cmd >= 0 && cmd < MMU_CNTL_DUMMY_END)
  399. {
  400. handler = control_handler[cmd];
  401. while (vaddr < vend)
  402. {
  403. uintptr_t *pte = _query(aspace, vaddr, &level);
  404. void *range_end = vaddr + _get_level_size(level);
  405. RT_ASSERT(range_end <= vend);
  406. if (pte)
  407. {
  408. err = handler(pte);
  409. RT_ASSERT(err == RT_EOK);
  410. }
  411. vaddr = range_end;
  412. }
  413. }
  414. else
  415. {
  416. err = -RT_ENOSYS;
  417. }
  418. return err;
  419. }
  420. /**
  421. * @brief setup Page Table for kernel space. It's a fixed map
  422. * and all mappings cannot be changed after initialization.
  423. *
  424. * Memory region in struct mem_desc must be page aligned,
  425. * otherwise is a failure and no report will be
  426. * returned.
  427. *
  428. * @param aspace
  429. * @param mdesc
  430. * @param desc_nr
  431. */
  432. void rt_hw_mmu_setup(rt_aspace_t aspace, struct mem_desc *mdesc, int desc_nr)
  433. {
  434. void *err;
  435. for (size_t i = 0; i < desc_nr; i++)
  436. {
  437. size_t attr;
  438. switch (mdesc->attr)
  439. {
  440. case NORMAL_MEM:
  441. attr = MMU_MAP_K_RWCB;
  442. break;
  443. case NORMAL_NOCACHE_MEM:
  444. attr = MMU_MAP_K_RWCB;
  445. break;
  446. case DEVICE_MEM:
  447. attr = MMU_MAP_K_DEVICE;
  448. break;
  449. default:
  450. attr = MMU_MAP_K_DEVICE;
  451. }
  452. struct rt_mm_va_hint hint = {.flags = MMF_MAP_FIXED,
  453. .limit_start = aspace->start,
  454. .limit_range_size = aspace->size,
  455. .map_size = mdesc->vaddr_end -
  456. mdesc->vaddr_start + 1,
  457. .prefer = (void *)mdesc->vaddr_start};
  458. if (mdesc->paddr_start == (rt_size_t)ARCH_MAP_FAILED)
  459. mdesc->paddr_start = mdesc->vaddr_start + PV_OFFSET;
  460. rt_aspace_map_phy_static(aspace, &mdesc->varea, &hint, attr,
  461. mdesc->paddr_start >> MM_PAGE_SHIFT, &err);
  462. mdesc++;
  463. }
  464. _asid_init();
  465. rt_hw_aspace_switch(&rt_kernel_space);
  466. rt_page_cleanup();
  467. }
  468. void rt_hw_mmu_kernel_map_init(rt_aspace_t aspace, rt_size_t vaddr_start, rt_size_t size)
  469. {
  470. rt_size_t paddr_start =
  471. __UMASKVALUE(VPN_TO_PPN(vaddr_start, PV_OFFSET), PAGE_OFFSET_MASK);
  472. rt_size_t va_s = GET_L1(vaddr_start);
  473. rt_size_t va_e = GET_L1(vaddr_start + size - 1);
  474. rt_size_t i;
  475. for (i = va_s; i <= va_e; i++)
  476. {
  477. MMUTable[i] =
  478. COMBINEPTE(paddr_start, PAGE_ATTR_RWX | PTE_G | PTE_V | PTE_CACHE |
  479. PTE_SHARE | PTE_BUF | PTE_A | PTE_D);
  480. paddr_start += L1_PAGE_SIZE;
  481. }
  482. rt_hw_tlb_invalidate_all_local();
  483. }
  484. void *rt_hw_mmu_pgtbl_create(void)
  485. {
  486. size_t *mmu_table;
  487. mmu_table = (rt_ubase_t *)rt_pages_alloc_ext(0, PAGE_ANY_AVAILABLE);
  488. if (!mmu_table)
  489. {
  490. return RT_NULL;
  491. }
  492. rt_memcpy(mmu_table, rt_kernel_space.page_table, ARCH_PAGE_SIZE);
  493. rt_hw_cpu_dcache_ops(RT_HW_CACHE_FLUSH, mmu_table, ARCH_PAGE_SIZE);
  494. return mmu_table;
  495. }
  496. void rt_hw_mmu_pgtbl_delete(void *pgtbl)
  497. {
  498. rt_pages_free(pgtbl, 0);
  499. }