lwp_user_mm.c 15 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623
  1. /*
  2. * Copyright (c) 2006-2021, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Change Logs:
  7. * Date Author Notes
  8. * 2019-10-28 Jesven first version
  9. * 2021-02-06 lizhirui fixed fixed vtable size problem
  10. * 2021-02-12 lizhirui add 64-bit support for lwp_brk
  11. * 2021-02-19 lizhirui add riscv64 support for lwp_user_accessable and lwp_get_from_user
  12. * 2021-06-07 lizhirui modify user space bound check
  13. */
  14. #include <rtthread.h>
  15. #include <rthw.h>
  16. #ifdef ARCH_MM_MMU
  17. #include <mmu.h>
  18. #include <page.h>
  19. #include <lwp_mm_area.h>
  20. #include <lwp_user_mm.h>
  21. #include <lwp_arch.h>
  22. #include <lwp_mm.h>
  23. int lwp_user_space_init(struct rt_lwp *lwp)
  24. {
  25. return arch_user_space_init(lwp);
  26. }
  27. #ifdef LWP_ENABLE_ASID
  28. void rt_hw_mmu_switch(void *mtable, unsigned int pid, unsigned int asid);
  29. #else
  30. void rt_hw_mmu_switch(void *mtable);
  31. #endif
  32. void *rt_hw_mmu_tbl_get(void);
  33. void lwp_mmu_switch(struct rt_thread *thread)
  34. {
  35. struct rt_lwp *l = RT_NULL;
  36. void *pre_mmu_table = RT_NULL, *new_mmu_table = RT_NULL;
  37. if (thread->lwp)
  38. {
  39. l = (struct rt_lwp *)thread->lwp;
  40. new_mmu_table = (void *)((char *)l->mmu_info.vtable + l->mmu_info.pv_off);
  41. }
  42. else
  43. {
  44. new_mmu_table = arch_kernel_mmu_table_get();
  45. }
  46. pre_mmu_table = rt_hw_mmu_tbl_get();
  47. if (pre_mmu_table != new_mmu_table)
  48. {
  49. #ifdef LWP_ENABLE_ASID
  50. rt_hw_mmu_switch(new_mmu_table, l ? l->pid : 0, arch_get_asid(l));
  51. #else
  52. rt_hw_mmu_switch(new_mmu_table);
  53. #endif
  54. }
  55. }
  56. static void unmap_range(struct rt_lwp *lwp, void *addr, size_t size, int pa_need_free)
  57. {
  58. void *va = RT_NULL, *pa = RT_NULL;
  59. int i = 0;
  60. for (va = addr, i = 0; i < size; va = (void *)((char *)va + ARCH_PAGE_SIZE), i += ARCH_PAGE_SIZE)
  61. {
  62. pa = rt_hw_mmu_v2p(&lwp->mmu_info, va);
  63. if (pa)
  64. {
  65. rt_hw_mmu_unmap(&lwp->mmu_info, va, ARCH_PAGE_SIZE);
  66. if (pa_need_free)
  67. {
  68. rt_pages_free((void *)((char *)pa - PV_OFFSET), 0);
  69. }
  70. }
  71. }
  72. }
  73. void lwp_unmap_user_space(struct rt_lwp *lwp)
  74. {
  75. struct lwp_avl_struct *node = RT_NULL;
  76. while ((node = lwp_map_find_first(lwp->map_area)) != 0)
  77. {
  78. struct rt_mm_area_struct *ma = (struct rt_mm_area_struct *)node->data;
  79. int pa_need_free = 0;
  80. RT_ASSERT(ma->type < MM_AREA_TYPE_UNKNOW);
  81. switch (ma->type)
  82. {
  83. case MM_AREA_TYPE_DATA:
  84. case MM_AREA_TYPE_TEXT:
  85. pa_need_free = 1;
  86. break;
  87. case MM_AREA_TYPE_SHM:
  88. lwp_shm_ref_dec(lwp, (void *)ma->addr);
  89. break;
  90. }
  91. unmap_range(lwp, (void *)ma->addr, ma->size, pa_need_free);
  92. lwp_map_area_remove(&lwp->map_area, ma->addr);
  93. }
  94. arch_user_space_vtable_free(lwp);
  95. }
  96. static void *_lwp_map_user(struct rt_lwp *lwp, void *map_va, size_t map_size, int text)
  97. {
  98. void *va = RT_NULL;
  99. int ret = 0;
  100. rt_mmu_info *m_info = &lwp->mmu_info;
  101. int area_type;
  102. va = rt_hw_mmu_map_auto(m_info, map_va, map_size, MMU_MAP_U_RWCB);
  103. if (!va)
  104. {
  105. return 0;
  106. }
  107. area_type = text ? MM_AREA_TYPE_TEXT : MM_AREA_TYPE_DATA;
  108. ret = lwp_map_area_insert(&lwp->map_area, (size_t)va, map_size, area_type);
  109. if (ret != 0)
  110. {
  111. unmap_range(lwp, va, map_size, 1);
  112. return 0;
  113. }
  114. return va;
  115. }
  116. int lwp_unmap_user(struct rt_lwp *lwp, void *va)
  117. {
  118. struct lwp_avl_struct *ma_avl_node = RT_NULL;
  119. struct rt_mm_area_struct *ma = RT_NULL;
  120. int pa_need_free = 0;
  121. rt_mm_lock();
  122. va = (void *)((size_t)va & ~ARCH_PAGE_MASK);
  123. ma_avl_node = lwp_map_find(lwp->map_area, (size_t)va);
  124. if (!ma_avl_node)
  125. {
  126. rt_mm_unlock();
  127. return -1;
  128. }
  129. ma = (struct rt_mm_area_struct *)ma_avl_node->data;
  130. RT_ASSERT(ma->type < MM_AREA_TYPE_UNKNOW);
  131. if ((ma->type == MM_AREA_TYPE_DATA) || (ma->type == MM_AREA_TYPE_TEXT))
  132. {
  133. pa_need_free = 1;
  134. }
  135. unmap_range(lwp, (void *)ma->addr, ma->size, pa_need_free);
  136. lwp_map_area_remove(&lwp->map_area, (size_t)va);
  137. rt_mm_unlock();
  138. return 0;
  139. }
  140. int lwp_dup_user(struct lwp_avl_struct *ptree, void *arg)
  141. {
  142. struct rt_lwp *self_lwp = lwp_self();
  143. struct rt_lwp *new_lwp = (struct rt_lwp *)arg;
  144. struct rt_mm_area_struct *ma = (struct rt_mm_area_struct *)ptree->data;
  145. void *pa = RT_NULL;
  146. void *va = RT_NULL;
  147. switch (ma->type)
  148. {
  149. case MM_AREA_TYPE_PHY:
  150. pa = rt_hw_mmu_v2p(&self_lwp->mmu_info, (void *)ma->addr);
  151. va = lwp_map_user_type(new_lwp, (void *)ma->addr, pa, ma->size, 0, MM_AREA_TYPE_PHY);
  152. break;
  153. case MM_AREA_TYPE_PHY_CACHED:
  154. pa = rt_hw_mmu_v2p(&self_lwp->mmu_info, (void *)ma->addr);
  155. va = lwp_map_user_type(new_lwp, (void *)ma->addr, pa, ma->size, 0, MM_AREA_TYPE_PHY_CACHED);
  156. break;
  157. case MM_AREA_TYPE_SHM:
  158. va = (void *)ma->addr;
  159. if (lwp_shm_ref_inc(self_lwp, va) > 0)
  160. {
  161. pa = rt_hw_mmu_v2p(&self_lwp->mmu_info, va);
  162. va = lwp_map_user_type(new_lwp, va, pa, ma->size, 1, MM_AREA_TYPE_SHM);
  163. }
  164. break;
  165. case MM_AREA_TYPE_DATA:
  166. va = lwp_map_user(new_lwp, (void *)ma->addr, ma->size, 0);
  167. if (va == (void *)ma->addr)
  168. {
  169. lwp_data_put(&new_lwp->mmu_info, va, va, ma->size);
  170. }
  171. break;
  172. case MM_AREA_TYPE_TEXT:
  173. {
  174. char *addr = (char *)ma->addr;
  175. size_t size = ma->size;
  176. while (size)
  177. {
  178. pa = rt_hw_mmu_v2p(&self_lwp->mmu_info, (void *)addr);
  179. rt_page_ref_inc((char *)pa - self_lwp->mmu_info.pv_off, 0);
  180. va = lwp_map_user_type(new_lwp, addr, pa, ARCH_PAGE_SIZE, 1, MM_AREA_TYPE_TEXT);
  181. if (va != addr)
  182. {
  183. return -1;
  184. }
  185. addr += ARCH_PAGE_SIZE;
  186. size -= ARCH_PAGE_SIZE;
  187. }
  188. va = (void *)ma->addr;
  189. }
  190. break;
  191. default:
  192. RT_ASSERT(0);
  193. break;
  194. }
  195. if (va != (void *)ma->addr)
  196. {
  197. return -1;
  198. }
  199. return 0;
  200. }
  201. int lwp_unmap_user_phy(struct rt_lwp *lwp, void *va)
  202. {
  203. return lwp_unmap_user(lwp, va);
  204. }
  205. int lwp_unmap_user_type(struct rt_lwp *lwp, void *va)
  206. {
  207. return lwp_unmap_user(lwp, va);
  208. }
  209. void *lwp_map_user(struct rt_lwp *lwp, void *map_va, size_t map_size, int text)
  210. {
  211. void *ret = RT_NULL;
  212. size_t offset = 0;
  213. if (!map_size)
  214. {
  215. return 0;
  216. }
  217. offset = (size_t)map_va & ARCH_PAGE_MASK;
  218. map_size += (offset + ARCH_PAGE_SIZE - 1);
  219. map_size &= ~ARCH_PAGE_MASK;
  220. map_va = (void *)((size_t)map_va & ~ARCH_PAGE_MASK);
  221. rt_mm_lock();
  222. ret = _lwp_map_user(lwp, map_va, map_size, text);
  223. rt_mm_unlock();
  224. if (ret)
  225. {
  226. ret = (void *)((char *)ret + offset);
  227. }
  228. return ret;
  229. }
  230. static void *_lwp_map_user_type(struct rt_lwp *lwp, void *map_va, void *map_pa, size_t map_size, int cached, int type)
  231. {
  232. void *va = RT_NULL;
  233. rt_mmu_info *m_info = &lwp->mmu_info;
  234. size_t attr = 0;
  235. int ret = 0;
  236. if (cached)
  237. {
  238. attr = MMU_MAP_U_RWCB;
  239. if (type == MM_AREA_TYPE_PHY)
  240. {
  241. type = MM_AREA_TYPE_PHY_CACHED;
  242. }
  243. }
  244. else
  245. {
  246. attr = MMU_MAP_U_RW;
  247. }
  248. va = rt_hw_mmu_map(m_info, map_va, map_pa, map_size, attr);
  249. if (va)
  250. {
  251. ret = lwp_map_area_insert(&lwp->map_area, (size_t)va, map_size, type);
  252. if (ret != 0)
  253. {
  254. unmap_range(lwp, va, map_size, 0);
  255. return 0;
  256. }
  257. }
  258. return va;
  259. }
  260. void *lwp_map_user_type(struct rt_lwp *lwp, void *map_va, void *map_pa, size_t map_size, int cached, int type)
  261. {
  262. void *ret = RT_NULL;
  263. size_t offset = 0;
  264. if (!map_size)
  265. {
  266. return 0;
  267. }
  268. if (map_va)
  269. {
  270. if (((size_t)map_va & ARCH_PAGE_MASK) != ((size_t)map_pa & ARCH_PAGE_MASK))
  271. {
  272. return 0;
  273. }
  274. }
  275. offset = (size_t)map_pa & ARCH_PAGE_MASK;
  276. map_size += (offset + ARCH_PAGE_SIZE - 1);
  277. map_size &= ~ARCH_PAGE_MASK;
  278. map_pa = (void *)((size_t)map_pa & ~ARCH_PAGE_MASK);
  279. rt_mm_lock();
  280. ret = _lwp_map_user_type(lwp, map_va, map_pa, map_size, cached, type);
  281. rt_mm_unlock();
  282. if (ret)
  283. {
  284. ret = (void *)((char *)ret + offset);
  285. }
  286. return ret;
  287. }
  288. void *lwp_map_user_phy(struct rt_lwp *lwp, void *map_va, void *map_pa, size_t map_size, int cached)
  289. {
  290. return lwp_map_user_type(lwp, map_va, map_pa, map_size, cached, MM_AREA_TYPE_PHY);
  291. }
  292. rt_base_t lwp_brk(void *addr)
  293. {
  294. rt_base_t ret = -1;
  295. struct rt_lwp *lwp = RT_NULL;
  296. rt_mm_lock();
  297. lwp = rt_thread_self()->lwp;
  298. if ((size_t)addr <= lwp->end_heap)
  299. {
  300. ret = (rt_base_t)lwp->end_heap;
  301. }
  302. else
  303. {
  304. size_t size = 0;
  305. void *va = RT_NULL;
  306. if ((size_t)addr <= USER_HEAP_VEND)
  307. {
  308. size = (((size_t)addr - lwp->end_heap) + ARCH_PAGE_SIZE - 1) & ~ARCH_PAGE_MASK;
  309. va = lwp_map_user(lwp, (void *)lwp->end_heap, size, 0);
  310. }
  311. if (va)
  312. {
  313. lwp->end_heap += size;
  314. ret = lwp->end_heap;
  315. }
  316. }
  317. rt_mm_unlock();
  318. return ret;
  319. }
  320. #define MAP_ANONYMOUS 0x20
  321. void* lwp_mmap2(void *addr, size_t length, int prot,
  322. int flags, int fd, off_t pgoffset)
  323. {
  324. void *ret = (void *)-1;
  325. if (fd == -1)
  326. {
  327. rt_mm_lock();
  328. ret = lwp_map_user(lwp_self(), addr, length, 0);
  329. rt_mm_unlock();
  330. if (ret)
  331. {
  332. if ((flags & MAP_ANONYMOUS) != 0)
  333. {
  334. rt_memset(ret, 0, length);
  335. }
  336. }
  337. else
  338. {
  339. ret = (void *)-1;
  340. }
  341. }
  342. else
  343. {
  344. struct dfs_fd *d;
  345. d = fd_get(fd);
  346. if (d && d->vnode->type == FT_DEVICE)
  347. {
  348. struct dfs_mmap2_args mmap2;
  349. mmap2.addr = addr;
  350. mmap2.length = length;
  351. mmap2.prot = prot;
  352. mmap2.flags = flags;
  353. mmap2.pgoffset = pgoffset;
  354. mmap2.ret = (void*) -1;
  355. if (dfs_file_mmap2(d, &mmap2) == 0)
  356. {
  357. ret = mmap2.ret;
  358. }
  359. }
  360. }
  361. return ret;
  362. }
  363. int lwp_munmap(void *addr)
  364. {
  365. int ret = 0;
  366. rt_mm_lock();
  367. ret = lwp_unmap_user(lwp_self(), addr);
  368. rt_mm_unlock();
  369. return ret;
  370. }
  371. size_t lwp_get_from_user(void *dst, void *src, size_t size)
  372. {
  373. struct rt_lwp *lwp = RT_NULL;
  374. rt_mmu_info *m_info = RT_NULL;
  375. /* check src */
  376. if (src < (void *)USER_VADDR_START)
  377. {
  378. return 0;
  379. }
  380. if (src >= (void *)USER_VADDR_TOP)
  381. {
  382. return 0;
  383. }
  384. if ((void *)((char *)src + size) > (void *)USER_VADDR_TOP)
  385. {
  386. return 0;
  387. }
  388. lwp = lwp_self();
  389. if (!lwp)
  390. {
  391. return 0;
  392. }
  393. m_info = &lwp->mmu_info;
  394. return lwp_data_get(m_info, dst, src, size);
  395. }
  396. size_t lwp_put_to_user(void *dst, void *src, size_t size)
  397. {
  398. struct rt_lwp *lwp = RT_NULL;
  399. rt_mmu_info *m_info = RT_NULL;
  400. /* check dst */
  401. if (dst < (void *)USER_VADDR_START)
  402. {
  403. return 0;
  404. }
  405. if (dst >= (void *)USER_VADDR_TOP)
  406. {
  407. return 0;
  408. }
  409. if ((void *)((char *)dst + size) > (void *)USER_VADDR_TOP)
  410. {
  411. return 0;
  412. }
  413. lwp = lwp_self();
  414. if (!lwp)
  415. {
  416. return 0;
  417. }
  418. m_info = &lwp->mmu_info;
  419. return lwp_data_put(m_info, dst, src, size);
  420. }
  421. int lwp_user_accessable(void *addr, size_t size)
  422. {
  423. void *addr_start = RT_NULL, *addr_end = RT_NULL, *next_page = RT_NULL;
  424. void *tmp_addr = RT_NULL;
  425. struct rt_lwp *lwp = lwp_self();
  426. rt_mmu_info *mmu_info = RT_NULL;
  427. if (!lwp)
  428. {
  429. return 0;
  430. }
  431. if (!size || !addr)
  432. {
  433. return 0;
  434. }
  435. addr_start = addr;
  436. addr_end = (void *)((char *)addr + size);
  437. #ifdef ARCH_RISCV64
  438. if(addr_start < (void *)USER_VADDR_START)
  439. {
  440. return 0;
  441. }
  442. #else
  443. if (addr_start >= (void *)USER_VADDR_TOP)
  444. {
  445. return 0;
  446. }
  447. if (addr_end > (void *)USER_VADDR_TOP)
  448. {
  449. return 0;
  450. }
  451. #endif
  452. mmu_info = &lwp->mmu_info;
  453. next_page = (void *)(((size_t)addr_start + ARCH_PAGE_SIZE) & ~(ARCH_PAGE_SIZE - 1));
  454. do
  455. {
  456. size_t len = (char *)next_page - (char *)addr_start;
  457. if (size < len)
  458. {
  459. len = size;
  460. }
  461. tmp_addr = rt_hw_mmu_v2p(mmu_info, addr_start);
  462. if (!tmp_addr)
  463. {
  464. return 0;
  465. }
  466. addr_start = (void *)((char *)addr_start + len);
  467. size -= len;
  468. next_page = (void *)((char *)next_page + ARCH_PAGE_SIZE);
  469. } while (addr_start < addr_end);
  470. return 1;
  471. }
  472. /* src is in mmu_info space, dst is in current thread space */
  473. size_t lwp_data_get(rt_mmu_info *mmu_info, void *dst, void *src, size_t size)
  474. {
  475. size_t copy_len = 0;
  476. void *addr_start = RT_NULL, *addr_end = RT_NULL, *next_page = RT_NULL;
  477. void *tmp_dst = RT_NULL, *tmp_src = RT_NULL;
  478. if (!size || !dst)
  479. {
  480. return 0;
  481. }
  482. tmp_dst = dst;
  483. addr_start = src;
  484. addr_end = (void *)((char *)src + size);
  485. next_page = (void *)(((size_t)addr_start + ARCH_PAGE_SIZE) & ~(ARCH_PAGE_SIZE - 1));
  486. do
  487. {
  488. size_t len = (char *)next_page - (char *)addr_start;
  489. if (size < len)
  490. {
  491. len = size;
  492. }
  493. tmp_src = rt_hw_mmu_v2p(mmu_info, addr_start);
  494. if (!tmp_src)
  495. {
  496. break;
  497. }
  498. tmp_src = (void *)((char *)tmp_src - PV_OFFSET);
  499. rt_memcpy(tmp_dst, tmp_src, len);
  500. tmp_dst = (void *)((char *)tmp_dst + len);
  501. addr_start = (void *)((char *)addr_start + len);
  502. size -= len;
  503. next_page = (void *)((char *)next_page + ARCH_PAGE_SIZE);
  504. copy_len += len;
  505. } while (addr_start < addr_end);
  506. return copy_len;
  507. }
  508. /* dst is in mmu_info space, src is in current thread space */
  509. size_t lwp_data_put(rt_mmu_info *mmu_info, void *dst, void *src, size_t size)
  510. {
  511. size_t copy_len = 0;
  512. void *addr_start = RT_NULL, *addr_end = RT_NULL, *next_page = RT_NULL;
  513. void *tmp_dst = RT_NULL, *tmp_src = RT_NULL;
  514. if (!size || !dst)
  515. {
  516. return 0;
  517. }
  518. tmp_src = src;
  519. addr_start = dst;
  520. addr_end = (void *)((char *)dst + size);
  521. next_page = (void *)(((size_t)addr_start + ARCH_PAGE_SIZE) & ~(ARCH_PAGE_SIZE - 1));
  522. do
  523. {
  524. size_t len = (char *)next_page - (char *)addr_start;
  525. if (size < len)
  526. {
  527. len = size;
  528. }
  529. tmp_dst = rt_hw_mmu_v2p(mmu_info, addr_start);
  530. if (!tmp_dst)
  531. {
  532. break;
  533. }
  534. tmp_dst = (void *)((char *)tmp_dst - PV_OFFSET);
  535. rt_memcpy(tmp_dst, tmp_src, len);
  536. tmp_src = (void *)((char *)tmp_src + len);
  537. addr_start = (void *)((char *)addr_start + len);
  538. size -= len;
  539. next_page = (void *)((char *)next_page + ARCH_PAGE_SIZE);
  540. copy_len += len;
  541. } while (addr_start < addr_end);
  542. return copy_len;
  543. }
  544. void lwp_data_cache_flush(rt_mmu_info *mmu_info, void *vaddr, size_t size)
  545. {
  546. void *paddr = RT_NULL;
  547. paddr = rt_hw_mmu_v2p(mmu_info, vaddr);
  548. paddr = (void *)((char *)paddr - PV_OFFSET);
  549. rt_hw_cpu_dcache_ops(RT_HW_CACHE_FLUSH, paddr, size);
  550. }
  551. #endif