start_gcc.S 9.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353
  1. /*
  2. * Copyright (c) 2006-2022, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Change Logs:
  7. * Date Author Notes
  8. * 2013-07-05 Bernard the first version
  9. */
  10. #include <rtconfig.h>
  11. #ifdef RT_USING_VMM
  12. #include <vmm.h>
  13. .equ orig_irq_isr, LINUX_VECTOR_POS+0x18
  14. #else
  15. #undef RT_VMM_USING_DOMAIN
  16. #endif
  17. .equ Mode_USR, 0x10
  18. .equ Mode_FIQ, 0x11
  19. .equ Mode_IRQ, 0x12
  20. .equ Mode_SVC, 0x13
  21. .equ Mode_ABT, 0x17
  22. .equ Mode_UND, 0x1B
  23. .equ Mode_SYS, 0x1F
  24. .equ I_Bit, 0x80 @ when I bit is set, IRQ is disabled
  25. .equ F_Bit, 0x40 @ when F bit is set, FIQ is disabled
  26. #ifndef RT_USING_VMM
  27. .equ UND_Stack_Size, 0x00000000
  28. .equ SVC_Stack_Size, 0x00000100
  29. .equ ABT_Stack_Size, 0x00000000
  30. .equ RT_FIQ_STACK_PGSZ, 0x00000000
  31. .equ RT_IRQ_STACK_PGSZ, 0x00000100
  32. .equ USR_Stack_Size, 0x00000100
  33. #define ISR_Stack_Size (UND_Stack_Size + SVC_Stack_Size + ABT_Stack_Size + \
  34. RT_FIQ_STACK_PGSZ + RT_IRQ_STACK_PGSZ)
  35. #else
  36. #define ISR_Stack_Size (RT_FIQ_STACK_PGSZ + RT_IRQ_STACK_PGSZ)
  37. #endif
  38. .section .data.share.isr
  39. /* stack */
  40. .globl stack_start
  41. .globl stack_top
  42. .align 3
  43. stack_start:
  44. .rept ISR_Stack_Size
  45. .byte 0
  46. .endr
  47. stack_top:
  48. .text
  49. /* reset entry */
  50. .globl _reset
  51. _reset:
  52. #ifdef RT_USING_VMM
  53. /* save all the parameter and variable registers */
  54. stmfd sp!, {r0-r12, lr}
  55. #endif
  56. /* set the cpu to SVC32 mode and disable interrupt */
  57. mrs r0, cpsr
  58. bic r0, r0, #0x1f
  59. orr r0, r0, #0x13
  60. msr cpsr_c, r0
  61. /* setup stack */
  62. bl stack_setup
  63. /* clear .bss */
  64. mov r0,#0 /* get a zero */
  65. ldr r1,=__bss_start /* bss start */
  66. ldr r2,=__bss_end /* bss end */
  67. bss_loop:
  68. cmp r1,r2 /* check if data to clear */
  69. strlo r0,[r1],#4 /* clear 4 bytes */
  70. blo bss_loop /* loop until done */
  71. #ifdef RT_USING_VMM
  72. /* clear .bss.share */
  73. mov r0,#0 /* get a zero */
  74. ldr r1,=__bss_share_start /* bss start */
  75. ldr r2,=__bss_share_end /* bss end */
  76. bss_share_loop:
  77. cmp r1,r2 /* check if data to clear */
  78. strlo r0,[r1],#4 /* clear 4 bytes */
  79. blo bss_share_loop /* loop until done */
  80. #endif
  81. /* call C++ constructors of global objects */
  82. ldr r0, =__ctors_start__
  83. ldr r1, =__ctors_end__
  84. ctor_loop:
  85. cmp r0, r1
  86. beq ctor_end
  87. ldr r2, [r0], #4
  88. stmfd sp!, {r0-r1}
  89. mov lr, pc
  90. bx r2
  91. ldmfd sp!, {r0-r1}
  92. b ctor_loop
  93. ctor_end:
  94. /* start RT-Thread Kernel */
  95. #ifdef RT_USING_VMM
  96. /* restore the parameter */
  97. ldmfd sp!, {r0-r3}
  98. bl vmm_entry
  99. ldmfd sp!, {r4-r12, pc}
  100. #else
  101. ldr pc, _rtthread_startup
  102. _rtthread_startup:
  103. .word rtthread_startup
  104. #endif
  105. stack_setup:
  106. ldr r0, =stack_top
  107. #ifdef RT_USING_VMM
  108. @ Linux use stmia to save r0, lr and spsr. To align to 8 byte boundary,
  109. @ just allocate 16 bytes for it.
  110. sub r0, r0, #16
  111. #endif
  112. #ifndef RT_USING_VMM
  113. @ Set the startup stack for svc
  114. mov sp, r0
  115. #endif
  116. #ifndef RT_USING_VMM
  117. @ Enter Undefined Instruction Mode and set its Stack Pointer
  118. msr cpsr_c, #Mode_UND|I_Bit|F_Bit
  119. mov sp, r0
  120. sub r0, r0, #UND_Stack_Size
  121. @ Enter Abort Mode and set its Stack Pointer
  122. msr cpsr_c, #Mode_ABT|I_Bit|F_Bit
  123. mov sp, r0
  124. sub r0, r0, #ABT_Stack_Size
  125. #endif
  126. @ Enter FIQ Mode and set its Stack Pointer
  127. msr cpsr_c, #Mode_FIQ|I_Bit|F_Bit
  128. mov sp, r0
  129. sub r0, r0, #RT_FIQ_STACK_PGSZ
  130. @ Enter IRQ Mode and set its Stack Pointer
  131. msr cpsr_c, #Mode_IRQ|I_Bit|F_Bit
  132. mov sp, r0
  133. sub r0, r0, #RT_IRQ_STACK_PGSZ
  134. /* come back to SVC mode */
  135. msr cpsr_c, #Mode_SVC|I_Bit|F_Bit
  136. bx lr
  137. /* exception handlers: undef, swi, padt, dabt, resv, irq, fiq */
  138. .section .text.isr, "ax"
  139. .align 5
  140. .globl vector_fiq
  141. vector_fiq:
  142. stmfd sp!,{r0-r7,lr}
  143. bl rt_hw_trap_fiq
  144. ldmfd sp!,{r0-r7,lr}
  145. subs pc, lr, #4
  146. .globl rt_interrupt_enter
  147. .globl rt_interrupt_leave
  148. .globl rt_thread_switch_interrupt_flag
  149. .globl rt_interrupt_from_thread
  150. .globl rt_interrupt_to_thread
  151. .globl rt_current_thread
  152. .globl vmm_thread
  153. .globl vmm_virq_check
  154. .align 5
  155. .globl vector_irq
  156. vector_irq:
  157. stmfd sp!, {r0-r12,lr}
  158. #ifdef RT_VMM_USING_DOMAIN
  159. @ save the last domain
  160. mrc p15, 0, r5, c3, c0
  161. @ switch to vmm domain as we are going to call vmm codes
  162. ldr r1, =vmm_domain_val
  163. ldr r4, [r1]
  164. mcr p15, 0, r4, c3, c0
  165. #endif
  166. bl rt_interrupt_enter
  167. bl rt_hw_trap_irq
  168. bl rt_interrupt_leave
  169. #ifdef RT_VMM_USING_DOMAIN
  170. @ restore the last domain. It do some redundant work but simplify the
  171. @ logic. It might be the guest domain so rt_thread_switch_interrupt_flag
  172. @ should lay in .bss.share
  173. mcr p15, 0, r5, c3, c0
  174. #endif
  175. @ if rt_thread_switch_interrupt_flag set, jump to
  176. @ rt_hw_context_switch_interrupt_do and don't return
  177. ldr r0, =rt_thread_switch_interrupt_flag
  178. ldr r1, [r0]
  179. cmp r1, #1
  180. beq rt_hw_context_switch_interrupt_do
  181. #ifndef RT_USING_VMM
  182. ldmfd sp!, {r0-r12,lr}
  183. subs pc, lr, #4
  184. #else
  185. #ifdef RT_VMM_USING_DOMAIN
  186. @ r4 is vmm_domain_val
  187. @ back to vmm domain as we need access rt_current_thread
  188. mcr p15, 0, r4, c3, c0
  189. #endif
  190. /* check whether we need to do IRQ routing
  191. * ensure the int is disabled. Or there will be an infinite loop. */
  192. ldr r0, =rt_current_thread
  193. ldr r0, [r0]
  194. ldr r1, =vmm_thread
  195. cmp r0, r1
  196. beq switch_to_guest
  197. #ifdef RT_VMM_USING_DOMAIN
  198. @ r5 is domain of interrupted context
  199. @ it might be super_domain_val or vmm_domain_val so we need to restore it.
  200. mcr p15, 0, r5, c3, c0
  201. #endif
  202. @ switch back if the interrupted thread is not vmm
  203. ldmfd sp!, {r0-r12,lr}
  204. subs pc, lr, #4
  205. switch_to_guest:
  206. #ifdef RT_VMM_USING_DOMAIN
  207. @ We are going to execute rt-thread code but accessing the content of the
  208. @ guest. So switch to super domain.
  209. ldr r1, =super_domain_val
  210. ldr r0, [r1]
  211. mcr p15, 0, r0, c3, c0
  212. #endif
  213. /* check whether there is a pending interrupt for Guest OS */
  214. bl vmm_virq_check
  215. #ifdef RT_VMM_USING_DOMAIN
  216. @ All done, restore the guest domain.
  217. mcr p15, 0, r5, c3, c0
  218. #endif
  219. cmp r0, #0x0
  220. beq route_irq_to_guest
  221. ldmfd sp!, {r0-r12,lr}
  222. subs pc, lr, #4
  223. route_irq_to_guest:
  224. ldmfd sp!, {r0-r12,lr}
  225. b orig_irq_isr
  226. #endif /* RT_USING_VMM */
  227. rt_hw_context_switch_interrupt_do:
  228. mov r1, #0 @ clear flag
  229. str r1, [r0]
  230. mov r1, sp @ r1 point to {r0-r3} in stack
  231. add sp, sp, #4*4
  232. ldmfd sp!, {r4-r12,lr}@ reload saved registers
  233. mrs r0, spsr @ get cpsr of interrupt thread
  234. sub r2, lr, #4 @ save old task's pc to r2
  235. @ Switch to SVC mode with no interrupt. If the usr mode guest is
  236. @ interrupted, this will just switch to the stack of kernel space.
  237. @ save the registers in kernel space won't trigger data abort.
  238. msr cpsr_c, #I_Bit|F_Bit|Mode_SVC
  239. stmfd sp!, {r2} @ push old task's pc
  240. stmfd sp!, {r4-r12,lr}@ push old task's lr,r12-r4
  241. ldmfd r1, {r1-r4} @ restore r0-r3 of the interrupt thread
  242. stmfd sp!, {r1-r4} @ push old task's r0-r3
  243. stmfd sp!, {r0} @ push old task's cpsr
  244. ldr r4, =rt_interrupt_from_thread
  245. ldr r5, [r4]
  246. str sp, [r5] @ store sp in preempted tasks's TCB
  247. #ifdef RT_VMM_USING_DOMAIN
  248. @ If a thread is wake up by interrupt, it should be RTT thread.
  249. @ Make sure the domain is correct.
  250. ldr r1, =vmm_domain_val
  251. ldr r2, [r1]
  252. mcr p15, 0, r2, c3, c0
  253. #endif
  254. ldr r6, =rt_interrupt_to_thread
  255. ldr r6, [r6]
  256. ldr sp, [r6] @ get new task's stack pointer
  257. ldmfd sp!, {r4} @ pop new task's cpsr to spsr
  258. msr spsr_cxsf, r4
  259. ldmfd sp!, {r0-r12,lr,pc}^ @ pop new task's r0-r12,lr & pc, copy spsr to cpsr
  260. .macro push_svc_reg
  261. sub sp, sp, #17 * 4 @/* Sizeof(struct rt_hw_exp_stack) */
  262. stmia sp, {r0 - r12} @/* Calling r0-r12 */
  263. mov r0, sp
  264. mrs r6, spsr @/* Save CPSR */
  265. str lr, [r0, #15*4] @/* Push PC */
  266. str r6, [r0, #16*4] @/* Push CPSR */
  267. cps #Mode_SVC
  268. str sp, [r0, #13*4] @/* Save calling SP */
  269. str lr, [r0, #14*4] @/* Save calling PC */
  270. .endm
  271. .align 5
  272. .globl vector_swi
  273. vector_swi:
  274. push_svc_reg
  275. bl rt_hw_trap_swi
  276. b .
  277. .align 5
  278. .globl vector_undef
  279. vector_undef:
  280. push_svc_reg
  281. bl rt_hw_trap_undef
  282. b .
  283. .align 5
  284. .globl vector_pabt
  285. vector_pabt:
  286. push_svc_reg
  287. bl rt_hw_trap_pabt
  288. b .
  289. .align 5
  290. .globl vector_dabt
  291. vector_dabt:
  292. push_svc_reg
  293. bl rt_hw_trap_dabt
  294. b .
  295. .align 5
  296. .globl vector_resv
  297. vector_resv:
  298. push_svc_reg
  299. bl rt_hw_trap_resv
  300. b .