lwp_arch.c 6.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263
  1. /*
  2. * Copyright (c) 2006-2023, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Change Logs:
  7. * Date Author Notes
  8. * 2021-05-18 Jesven first version
  9. * 2023-07-16 Shell Move part of the codes to C from asm in signal handling
  10. * 2023-10-16 Shell Support a new backtrace framework
  11. * 2023-08-03 Shell Support of syscall restart (SA_RESTART)
  12. */
  13. #include <armv8.h>
  14. #include <rthw.h>
  15. #include <rtthread.h>
  16. #include <stdlib.h>
  17. #include <string.h>
  18. #include <lwp_internal.h>
  19. #ifdef ARCH_MM_MMU
  20. #define DBG_TAG "lwp.arch"
  21. #define DBG_LVL DBG_INFO
  22. #include <rtdbg.h>
  23. #include <lwp_arch.h>
  24. #include <lwp_user_mm.h>
  25. extern size_t MMUTable[];
  26. int arch_user_space_init(struct rt_lwp *lwp)
  27. {
  28. size_t *mmu_table;
  29. mmu_table = rt_hw_mmu_pgtbl_create();
  30. if (mmu_table)
  31. {
  32. lwp->end_heap = USER_HEAP_VADDR;
  33. lwp->aspace = rt_aspace_create(
  34. (void *)USER_VADDR_START, USER_VADDR_TOP - USER_VADDR_START, mmu_table);
  35. if (!lwp->aspace)
  36. {
  37. return -RT_ERROR;
  38. }
  39. }
  40. else
  41. {
  42. return -RT_ENOMEM;
  43. }
  44. return 0;
  45. }
  46. void *arch_kernel_mmu_table_get(void)
  47. {
  48. return (void *)NULL;
  49. }
  50. void arch_user_space_free(struct rt_lwp *lwp)
  51. {
  52. if (lwp)
  53. {
  54. RT_ASSERT(lwp->aspace);
  55. void *pgtbl = lwp->aspace->page_table;
  56. rt_aspace_delete(lwp->aspace);
  57. /* must be freed after aspace delete, pgtbl is required for unmap */
  58. rt_pages_free(pgtbl, 0);
  59. lwp->aspace = NULL;
  60. }
  61. else
  62. {
  63. LOG_W("%s: NULL lwp as parameter", __func__);
  64. RT_ASSERT(0);
  65. }
  66. }
  67. int arch_expand_user_stack(void *addr)
  68. {
  69. int ret = 0;
  70. size_t stack_addr = (size_t)addr;
  71. stack_addr &= ~ARCH_PAGE_MASK;
  72. if ((stack_addr >= (size_t)USER_STACK_VSTART) &&
  73. (stack_addr < (size_t)USER_STACK_VEND))
  74. {
  75. void *map =
  76. lwp_map_user(lwp_self(), (void *)stack_addr, ARCH_PAGE_SIZE, 0);
  77. if (map || lwp_user_accessable(addr, 1))
  78. {
  79. ret = 1;
  80. }
  81. }
  82. return ret;
  83. }
  84. #endif
  85. int arch_set_thread_context(void (*exit)(void), void *new_thread_stack,
  86. void *user_stack, void **thread_sp)
  87. {
  88. struct rt_hw_exp_stack *syscall_frame;
  89. struct rt_hw_exp_stack *thread_frame;
  90. struct rt_hw_exp_stack *ori_syscall = rt_thread_self()->user_ctx.ctx;
  91. RT_ASSERT(ori_syscall != RT_NULL);
  92. new_thread_stack = (rt_ubase_t*)RT_ALIGN_DOWN((rt_ubase_t)new_thread_stack, 16);
  93. syscall_frame = (void *)((long)new_thread_stack - sizeof(struct rt_hw_exp_stack));
  94. memcpy(syscall_frame, ori_syscall, sizeof(*syscall_frame));
  95. syscall_frame->sp_el0 = (long)user_stack;
  96. syscall_frame->x0 = 0;
  97. thread_frame = (void *)rt_hw_stack_init(exit, RT_NULL, (void *)syscall_frame, RT_NULL);
  98. *thread_sp = thread_frame;
  99. return 0;
  100. }
  101. #define ALGIN_BYTES (16)
  102. /* the layout is part of ABI, dont change it */
  103. struct signal_ucontext
  104. {
  105. rt_int64_t sigreturn;
  106. lwp_sigset_t save_sigmask;
  107. siginfo_t si;
  108. rt_align(ALGIN_BYTES)
  109. struct rt_hw_exp_stack frame;
  110. };
  111. RT_STATIC_ASSERT(abi_offset_compatible, offsetof(struct signal_ucontext, si) == UCTX_ABI_OFFSET_TO_SI);
  112. void *arch_signal_ucontext_get_frame(struct signal_ucontext *uctx)
  113. {
  114. return &uctx->frame;
  115. }
  116. /* internal used only */
  117. void arch_syscall_prepare_signal(rt_base_t rc, struct rt_hw_exp_stack *exp_frame)
  118. {
  119. long x0 = exp_frame->x0;
  120. exp_frame->x0 = rc;
  121. exp_frame->x7 = x0;
  122. return ;
  123. }
  124. void arch_syscall_restart(void *sp, void *ksp);
  125. void arch_syscall_set_errno(void *eframe, int expected, int code)
  126. {
  127. struct rt_hw_exp_stack *exp_frame = eframe;
  128. if (exp_frame->x0 == -expected)
  129. exp_frame->x0 = -code;
  130. return ;
  131. }
  132. void arch_signal_check_erestart(void *eframe, void *ksp)
  133. {
  134. struct rt_hw_exp_stack *exp_frame = eframe;
  135. long rc = exp_frame->x0;
  136. long sys_id = exp_frame->x8;
  137. (void)sys_id;
  138. if (rc == -ERESTART)
  139. {
  140. LOG_D("%s(rc=%ld,sys_id=%ld,pid=%d)", __func__, rc, sys_id, lwp_self()->pid);
  141. LOG_D("%s: restart rc = %ld", lwp_get_syscall_name(sys_id), rc);
  142. exp_frame->x0 = exp_frame->x7;
  143. arch_syscall_restart(eframe, ksp);
  144. }
  145. return ;
  146. }
  147. static void arch_signal_post_action(struct signal_ucontext *new_sp, rt_base_t kernel_sp)
  148. {
  149. arch_signal_check_erestart(&new_sp->frame, (void *)kernel_sp);
  150. return ;
  151. }
  152. void *arch_signal_ucontext_restore(rt_base_t user_sp, rt_base_t kernel_sp)
  153. {
  154. struct signal_ucontext *new_sp;
  155. new_sp = (void *)user_sp;
  156. if (lwp_user_accessable(new_sp, sizeof(*new_sp)))
  157. {
  158. lwp_thread_signal_mask(rt_thread_self(), LWP_SIG_MASK_CMD_SET_MASK, &new_sp->save_sigmask, RT_NULL);
  159. arch_signal_post_action(new_sp, kernel_sp);
  160. }
  161. else
  162. {
  163. LOG_I("User frame corrupted during signal handling\nexiting...");
  164. sys_exit_group(EXIT_FAILURE);
  165. }
  166. return (char *)&new_sp->frame + sizeof(struct rt_hw_exp_stack);
  167. }
  168. void *arch_signal_ucontext_save(rt_base_t user_sp, siginfo_t *psiginfo,
  169. struct rt_hw_exp_stack *exp_frame,
  170. lwp_sigset_t *save_sig_mask)
  171. {
  172. struct signal_ucontext *new_sp;
  173. new_sp = (void *)((user_sp - sizeof(struct signal_ucontext)) & ~0xf);
  174. if (lwp_user_accessable(new_sp, sizeof(*new_sp)))
  175. {
  176. /* push psiginfo */
  177. if (psiginfo)
  178. {
  179. lwp_memcpy(&new_sp->si, psiginfo, sizeof(*psiginfo));
  180. }
  181. /* exp frame is already aligned as AAPCS64 required */
  182. lwp_memcpy(&new_sp->frame, exp_frame, sizeof(*exp_frame));
  183. /* copy the save_sig_mask */
  184. lwp_memcpy(&new_sp->save_sigmask, save_sig_mask, sizeof(lwp_sigset_t));
  185. /* copy lwp_sigreturn */
  186. const size_t lwp_sigreturn_bytes = 8;
  187. extern void lwp_sigreturn(void);
  188. /* -> ensure that the sigreturn start at the outer most boundary */
  189. lwp_memcpy(&new_sp->sigreturn, &lwp_sigreturn, lwp_sigreturn_bytes);
  190. }
  191. else
  192. {
  193. LOG_I("%s: User stack overflow", __func__);
  194. sys_exit_group(EXIT_FAILURE);
  195. }
  196. return new_sp;
  197. }
  198. int arch_backtrace_uthread(rt_thread_t thread)
  199. {
  200. struct rt_hw_backtrace_frame frame;
  201. struct rt_hw_exp_stack *stack;
  202. if (thread && thread->lwp)
  203. {
  204. stack = thread->user_ctx.ctx;
  205. if ((long)stack > (unsigned long)thread->stack_addr
  206. && (long)stack < (unsigned long)thread->stack_addr + thread->stack_size)
  207. {
  208. frame.pc = stack->pc;
  209. frame.fp = stack->x29;
  210. lwp_backtrace_frame(thread, &frame);
  211. return 0;
  212. }
  213. else
  214. return -1;
  215. }
  216. return -1;
  217. }