start_gcc.S 10 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442
  1. /*
  2. * Copyright (c) 2006-2018, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Change Logs:
  7. * Date Author Notes
  8. * 2013-07-05 Bernard the first version
  9. * 2018-11-22 Jesven in the interrupt context, use rt_scheduler_do_irq_switch checks
  10. * and switches to a new thread
  11. */
  12. #include "rtconfig.h"
  13. .equ Mode_USR, 0x10
  14. .equ Mode_FIQ, 0x11
  15. .equ Mode_IRQ, 0x12
  16. .equ Mode_SVC, 0x13
  17. .equ Mode_ABT, 0x17
  18. .equ Mode_UND, 0x1B
  19. .equ Mode_SYS, 0x1F
  20. .equ I_Bit, 0x80 @ when I bit is set, IRQ is disabled
  21. .equ F_Bit, 0x40 @ when F bit is set, FIQ is disabled
  22. #ifdef RT_USING_FPU
  23. .equ UND_Stack_Size, 0x00000400
  24. #else
  25. .equ UND_Stack_Size, 0x00000000
  26. #endif
  27. .equ SVC_Stack_Size, 0x00000400
  28. .equ ABT_Stack_Size, 0x00000000
  29. .equ RT_FIQ_STACK_PGSZ, 0x00000000
  30. .equ RT_IRQ_STACK_PGSZ, 0x00000800
  31. .equ USR_Stack_Size, 0x00000400
  32. #define ISR_Stack_Size (UND_Stack_Size + SVC_Stack_Size + ABT_Stack_Size + \
  33. RT_FIQ_STACK_PGSZ + RT_IRQ_STACK_PGSZ)
  34. .section .data.share.isr
  35. /* stack */
  36. .globl stack_start
  37. .globl stack_top
  38. .align 3
  39. stack_start:
  40. .rept ISR_Stack_Size
  41. .byte 0
  42. .endr
  43. stack_top:
  44. .text
  45. /* reset entry */
  46. .globl _reset
  47. _reset:
  48. #ifdef ARCH_ARMV8
  49. /* Check for HYP mode */
  50. mrs r0, cpsr_all
  51. and r0, r0, #0x1F
  52. mov r8, #0x1A
  53. cmp r0, r8
  54. beq overHyped
  55. b continue
  56. overHyped: /* Get out of HYP mode */
  57. adr r1, continue
  58. msr ELR_hyp, r1
  59. mrs r1, cpsr_all
  60. and r1, r1, #0x1f ;@ CPSR_MODE_MASK
  61. orr r1, r1, #0x13 ;@ CPSR_MODE_SUPERVISOR
  62. msr SPSR_hyp, r1
  63. eret
  64. continue:
  65. #endif
  66. /* set the cpu to SVC32 mode and disable interrupt */
  67. cps #Mode_SVC
  68. #ifdef RT_USING_FPU
  69. mov r4, #0xfffffff
  70. mcr p15, 0, r4, c1, c0, 2
  71. #endif
  72. /* disable the data alignment check */
  73. mrc p15, 0, r1, c1, c0, 0
  74. bic r1, #(1<<1)
  75. mcr p15, 0, r1, c1, c0, 0
  76. /* setup stack */
  77. bl stack_setup
  78. /* clear .bss */
  79. mov r0,#0 /* get a zero */
  80. ldr r1,=__bss_start /* bss start */
  81. ldr r2,=__bss_end /* bss end */
  82. bss_loop:
  83. cmp r1,r2 /* check if data to clear */
  84. strlo r0,[r1],#4 /* clear 4 bytes */
  85. blo bss_loop /* loop until done */
  86. #ifdef RT_USING_SMP
  87. mrc p15, 0, r1, c1, c0, 1
  88. mov r0, #(1<<6)
  89. orr r1, r0
  90. mcr p15, 0, r1, c1, c0, 1 //enable smp
  91. #endif
  92. /* initialize the mmu table and enable mmu */
  93. ldr r0, =platform_mem_desc
  94. ldr r1, =platform_mem_desc_size
  95. ldr r1, [r1]
  96. bl rt_hw_init_mmu_table
  97. bl rt_hw_mmu_init
  98. /* call C++ constructors of global objects */
  99. ldr r0, =__ctors_start__
  100. ldr r1, =__ctors_end__
  101. ctor_loop:
  102. cmp r0, r1
  103. beq ctor_end
  104. ldr r2, [r0], #4
  105. stmfd sp!, {r0-r1}
  106. mov lr, pc
  107. bx r2
  108. ldmfd sp!, {r0-r1}
  109. b ctor_loop
  110. ctor_end:
  111. /* start RT-Thread Kernel */
  112. ldr pc, _rtthread_startup
  113. _rtthread_startup:
  114. .word rtthread_startup
  115. stack_setup:
  116. ldr r0, =stack_top
  117. @ Set the startup stack for svc
  118. mov sp, r0
  119. @ Enter Undefined Instruction Mode and set its Stack Pointer
  120. msr cpsr_c, #Mode_UND|I_Bit|F_Bit
  121. mov sp, r0
  122. sub r0, r0, #UND_Stack_Size
  123. @ Enter Abort Mode and set its Stack Pointer
  124. msr cpsr_c, #Mode_ABT|I_Bit|F_Bit
  125. mov sp, r0
  126. sub r0, r0, #ABT_Stack_Size
  127. @ Enter FIQ Mode and set its Stack Pointer
  128. msr cpsr_c, #Mode_FIQ|I_Bit|F_Bit
  129. mov sp, r0
  130. sub r0, r0, #RT_FIQ_STACK_PGSZ
  131. @ Enter IRQ Mode and set its Stack Pointer
  132. msr cpsr_c, #Mode_IRQ|I_Bit|F_Bit
  133. mov sp, r0
  134. sub r0, r0, #RT_IRQ_STACK_PGSZ
  135. /* come back to SVC mode */
  136. msr cpsr_c, #Mode_SVC|I_Bit|F_Bit
  137. bx lr
  138. /* exception handlers: undef, swi, padt, dabt, resv, irq, fiq */
  139. .section .text.isr, "ax"
  140. .align 5
  141. .globl vector_fiq
  142. vector_fiq:
  143. stmfd sp!,{r0-r7,lr}
  144. bl rt_hw_trap_fiq
  145. ldmfd sp!,{r0-r7,lr}
  146. subs pc, lr, #4
  147. .globl rt_interrupt_enter
  148. .globl rt_interrupt_leave
  149. .globl rt_thread_switch_interrupt_flag
  150. .globl rt_interrupt_from_thread
  151. .globl rt_interrupt_to_thread
  152. .globl rt_current_thread
  153. .globl vmm_thread
  154. .globl vmm_virq_check
  155. .align 5
  156. .globl vector_irq
  157. vector_irq:
  158. #ifdef RT_USING_SMP
  159. clrex
  160. stmfd sp!, {r0, r1}
  161. cps #Mode_SVC
  162. mov r0, sp /* svc_sp */
  163. mov r1, lr /* svc_lr */
  164. cps #Mode_IRQ
  165. sub lr, #4
  166. stmfd r0!, {r1, lr} /* svc_lr, svc_pc */
  167. stmfd r0!, {r2 - r12}
  168. ldmfd sp!, {r1, r2} /* original r0, r1 */
  169. stmfd r0!, {r1 - r2}
  170. mrs r1, spsr /* original mode */
  171. stmfd r0!, {r1}
  172. #ifdef RT_USING_LWP
  173. stmfd r0, {r13, r14}^ /* usr_sp, usr_lr */
  174. sub r0, #8
  175. #endif
  176. #ifdef RT_USING_FPU
  177. /* fpu context */
  178. vmrs r6, fpexc
  179. tst r6, #(1<<30)
  180. beq 1f
  181. vstmdb r0!, {d0-d15}
  182. vstmdb r0!, {d16-d31}
  183. vmrs r5, fpscr
  184. stmfd r0!, {r5}
  185. 1:
  186. stmfd r0!, {r6}
  187. #endif
  188. /* now irq stack is clean */
  189. /* r0 is task svc_sp */
  190. /* backup r0 -> r8 */
  191. mov r8, r0
  192. bl rt_interrupt_enter
  193. bl rt_hw_trap_irq
  194. bl rt_interrupt_leave
  195. cps #Mode_SVC
  196. mov sp, r8
  197. mov r0, r8
  198. bl rt_scheduler_do_irq_switch
  199. b rt_hw_context_switch_exit
  200. #else
  201. stmfd sp!, {r0-r12,lr}
  202. bl rt_interrupt_enter
  203. bl rt_hw_trap_irq
  204. bl rt_interrupt_leave
  205. @ if rt_thread_switch_interrupt_flag set, jump to
  206. @ rt_hw_context_switch_interrupt_do and don't return
  207. ldr r0, =rt_thread_switch_interrupt_flag
  208. ldr r1, [r0]
  209. cmp r1, #1
  210. beq rt_hw_context_switch_interrupt_do
  211. ldmfd sp!, {r0-r12,lr}
  212. subs pc, lr, #4
  213. rt_hw_context_switch_interrupt_do:
  214. mov r1, #0 @ clear flag
  215. str r1, [r0]
  216. mov r1, sp @ r1 point to {r0-r3} in stack
  217. add sp, sp, #4*4
  218. ldmfd sp!, {r4-r12,lr}@ reload saved registers
  219. mrs r0, spsr @ get cpsr of interrupt thread
  220. sub r2, lr, #4 @ save old task's pc to r2
  221. @ Switch to SVC mode with no interrupt. If the usr mode guest is
  222. @ interrupted, this will just switch to the stack of kernel space.
  223. @ save the registers in kernel space won't trigger data abort.
  224. msr cpsr_c, #I_Bit|F_Bit|Mode_SVC
  225. stmfd sp!, {r2} @ push old task's pc
  226. stmfd sp!, {r4-r12,lr}@ push old task's lr,r12-r4
  227. ldmfd r1, {r1-r4} @ restore r0-r3 of the interrupt thread
  228. stmfd sp!, {r1-r4} @ push old task's r0-r3
  229. stmfd sp!, {r0} @ push old task's cpsr
  230. #ifdef RT_USING_LWP
  231. stmfd sp, {r13, r14}^ @push usr_sp, usr_lr
  232. sub sp, #8
  233. #endif
  234. #ifdef RT_USING_FPU
  235. /* fpu context */
  236. vmrs r6, fpexc
  237. tst r6, #(1<<30)
  238. beq 1f
  239. vstmdb sp!, {d0-d15}
  240. vstmdb sp!, {d16-d31}
  241. vmrs r5, fpscr
  242. stmfd sp!, {r5}
  243. 1:
  244. stmfd sp!, {r6}
  245. #endif
  246. ldr r4, =rt_interrupt_from_thread
  247. ldr r5, [r4]
  248. str sp, [r5] @ store sp in preempted tasks's TCB
  249. ldr r6, =rt_interrupt_to_thread
  250. ldr r6, [r6]
  251. ldr sp, [r6] @ get new task's stack pointer
  252. #ifdef RT_USING_FPU
  253. /* fpu context */
  254. ldmfd sp!, {r6}
  255. vmsr fpexc, r6
  256. tst r6, #(1<<30)
  257. beq 1f
  258. ldmfd sp!, {r5}
  259. vmsr fpscr, r5
  260. vldmia sp!, {d16-d31}
  261. vldmia sp!, {d0-d15}
  262. 1:
  263. #endif
  264. #ifdef RT_USING_LWP
  265. ldmfd sp, {r13, r14}^ @pop usr_sp, usr_lr
  266. add sp, #8
  267. #endif
  268. ldmfd sp!, {r4} @ pop new task's cpsr to spsr
  269. msr spsr_cxsf, r4
  270. ldmfd sp!, {r0-r12,lr,pc}^ @ pop new task's r0-r12,lr & pc, copy spsr to cpsr
  271. #endif
  272. .macro push_svc_reg
  273. sub sp, sp, #17 * 4 @/* Sizeof(struct rt_hw_exp_stack) */
  274. stmia sp, {r0 - r12} @/* Calling r0-r12 */
  275. mov r0, sp
  276. mrs r6, spsr @/* Save CPSR */
  277. str lr, [r0, #15*4] @/* Push PC */
  278. str r6, [r0, #16*4] @/* Push CPSR */
  279. cps #Mode_SVC
  280. str sp, [r0, #13*4] @/* Save calling SP */
  281. str lr, [r0, #14*4] @/* Save calling PC */
  282. .endm
  283. .align 5
  284. .weak vector_swi
  285. vector_swi:
  286. push_svc_reg
  287. bl rt_hw_trap_swi
  288. b .
  289. .align 5
  290. .globl vector_undef
  291. vector_undef:
  292. push_svc_reg
  293. cps #Mode_UND
  294. bl rt_hw_trap_undef
  295. #ifdef RT_USING_FPU
  296. ldr lr, [sp, #15*4]
  297. ldmia sp, {r0 - r12}
  298. add sp, sp, #17 * 4
  299. movs pc, lr
  300. #endif
  301. b .
  302. .align 5
  303. .globl vector_pabt
  304. vector_pabt:
  305. push_svc_reg
  306. bl rt_hw_trap_pabt
  307. b .
  308. .align 5
  309. .globl vector_dabt
  310. vector_dabt:
  311. push_svc_reg
  312. bl rt_hw_trap_dabt
  313. b .
  314. .align 5
  315. .globl vector_resv
  316. vector_resv:
  317. push_svc_reg
  318. bl rt_hw_trap_resv
  319. b .
  320. #ifdef RT_USING_SMP
  321. .global set_secondary_cpu_boot_address
  322. set_secondary_cpu_boot_address:
  323. ldr r0, =secondary_cpu_start
  324. mvn r1, #0 //0xffffffff
  325. ldr r2, =0x10000034
  326. str r1, [r2]
  327. str r0, [r2, #-4]
  328. mov pc, lr
  329. .global secondary_cpu_start
  330. secondary_cpu_start:
  331. #ifdef RT_USING_FPU
  332. mov r4, #0xfffffff
  333. mcr p15, 0, r4, c1, c0, 2
  334. #endif
  335. mrc p15, 0, r1, c1, c0, 1
  336. mov r0, #(1<<6)
  337. orr r1, r0
  338. mcr p15, 0, r1, c1, c0, 1 //enable smp
  339. mrc p15, 0, r0, c1, c0, 0
  340. bic r0, #(1<<13)
  341. mcr p15, 0, r0, c1, c0, 0
  342. #ifdef RT_USING_FPU
  343. cps #Mode_UND
  344. ldr sp, =und_stack_2_limit
  345. #endif
  346. cps #Mode_IRQ
  347. ldr sp, =irq_stack_2_limit
  348. cps #Mode_FIQ
  349. ldr sp, =irq_stack_2_limit
  350. cps #Mode_SVC
  351. ldr sp, =svc_stack_2_limit
  352. /* initialize the mmu table and enable mmu */
  353. bl rt_hw_mmu_init
  354. b secondary_cpu_c_start
  355. #endif
  356. .bss
  357. .align 2 //align to 2~2=4
  358. svc_stack_2:
  359. .space (1 << 10)
  360. svc_stack_2_limit:
  361. irq_stack_2:
  362. .space (1 << 10)
  363. irq_stack_2_limit:
  364. #ifdef RT_USING_FPU
  365. und_stack_2:
  366. .space (1 << 10)
  367. und_stack_2_limit:
  368. #endif