start_gcc.S 15 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669
  1. /*
  2. * Copyright (c) 2006-2018, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Change Logs:
  7. * Date Author Notes
  8. * 2013-07-05 Bernard the first version
  9. * 2018-11-22 Jesven in the interrupt context, use rt_scheduler_do_irq_switch checks
  10. * and switches to a new thread
  11. */
  12. #include "rtconfig.h"
  13. .equ Mode_USR, 0x10
  14. .equ Mode_FIQ, 0x11
  15. .equ Mode_IRQ, 0x12
  16. .equ Mode_SVC, 0x13
  17. .equ Mode_ABT, 0x17
  18. .equ Mode_UND, 0x1B
  19. .equ Mode_SYS, 0x1F
  20. .equ I_Bit, 0x80 /* when I bit is set, IRQ is disabled */
  21. .equ F_Bit, 0x40 /* when F bit is set, FIQ is disabled */
  22. #ifdef RT_USING_USERSPACE
  23. .data
  24. .align 14
  25. init_mtbl:
  26. .space 16*1024
  27. #endif
  28. .text
  29. /* reset entry */
  30. .globl _reset
  31. _reset:
  32. #ifdef ARCH_ARMV8
  33. /* Check for HYP mode */
  34. mrs r0, cpsr_all
  35. and r0, r0, #0x1F
  36. mov r8, #0x1A
  37. cmp r0, r8
  38. beq overHyped
  39. b continue
  40. overHyped: /* Get out of HYP mode */
  41. adr r1, continue
  42. msr ELR_hyp, r1
  43. mrs r1, cpsr_all
  44. and r1, r1, #0x1f /* CPSR_MODE_MASK */
  45. orr r1, r1, #0x13 /* CPSR_MODE_SUPERVISOR */
  46. msr SPSR_hyp, r1
  47. eret
  48. continue:
  49. #endif
  50. #ifdef SOC_BCM283x
  51. /* Suspend the other cpu cores */
  52. mrc p15, 0, r0, c0, c0, 5
  53. ands r0, #3
  54. bne _halt
  55. /* Disable IRQ & FIQ */
  56. cpsid if
  57. /* Check for HYP mode */
  58. mrs r0, cpsr_all
  59. and r0, r0, #0x1F
  60. mov r8, #0x1A
  61. cmp r0, r8
  62. beq overHyped
  63. b continue
  64. overHyped: /* Get out of HYP mode */
  65. adr r1, continue
  66. msr ELR_hyp, r1
  67. mrs r1, cpsr_all
  68. and r1, r1, #0x1f /* CPSR_MODE_MASK */
  69. orr r1, r1, #0x13 /* CPSR_MODE_SUPERVISOR */
  70. msr SPSR_hyp, r1
  71. eret
  72. continue:
  73. /* set the cpu to SVC32 mode and disable interrupt */
  74. mrs r0, cpsr
  75. bic r0, r0, #0x1f
  76. orr r0, r0, #0x13
  77. msr cpsr_c, r0
  78. #endif
  79. /* invalid tlb before enable mmu */
  80. mrc p15, 0, r0, c1, c0, 0
  81. bic r0, #1
  82. mcr p15, 0, r0, c1, c0, 0
  83. dsb
  84. isb
  85. mov r0, #0
  86. mcr p15, 0, r0, c8, c7, 0
  87. mcr p15, 0, r0, c7, c5, 0 /* iciallu */
  88. mcr p15, 0, r0, c7, c5, 6 /* bpiall */
  89. dsb
  90. isb
  91. #ifdef RT_USING_USERSPACE
  92. ldr r5, =PV_OFFSET
  93. mov r7, #0x100000
  94. sub r7, #1
  95. mvn r8, r7
  96. ldr r9, =KERNEL_VADDR_START
  97. ldr r6, =__bss_end
  98. add r6, r7
  99. and r6, r8 /* r6 end vaddr align up to 1M */
  100. sub r6, r9 /* r6 is size */
  101. ldr sp, =svc_stack_n_limit
  102. add sp, r5 /* use paddr */
  103. ldr r0, =init_mtbl
  104. add r0, r5
  105. mov r1, r6
  106. mov r2, r5
  107. bl init_mm_setup
  108. ldr lr, =after_enable_mmu
  109. ldr r0, =init_mtbl
  110. add r0, r5
  111. b enable_mmu
  112. after_enable_mmu:
  113. #endif
  114. #ifndef SOC_BCM283x
  115. /* set the cpu to SVC32 mode and disable interrupt */
  116. cps #Mode_SVC
  117. #endif
  118. #ifdef RT_USING_FPU
  119. mov r4, #0xfffffff
  120. mcr p15, 0, r4, c1, c0, 2
  121. #endif
  122. /* disable the data alignment check */
  123. mrc p15, 0, r1, c1, c0, 0
  124. bic r1, #(1<<1)
  125. mcr p15, 0, r1, c1, c0, 0
  126. /* setup stack */
  127. bl stack_setup
  128. /* clear .bss */
  129. mov r0,#0 /* get a zero */
  130. ldr r1,=__bss_start /* bss start */
  131. ldr r2,=__bss_end /* bss end */
  132. bss_loop:
  133. cmp r1,r2 /* check if data to clear */
  134. strlo r0,[r1],#4 /* clear 4 bytes */
  135. blo bss_loop /* loop until done */
  136. #ifdef RT_USING_SMP
  137. mrc p15, 0, r1, c1, c0, 1
  138. mov r0, #(1<<6)
  139. orr r1, r0
  140. mcr p15, 0, r1, c1, c0, 1 /* enable smp */
  141. #endif
  142. /* initialize the mmu table and enable mmu */
  143. ldr r0, =platform_mem_desc
  144. ldr r1, =platform_mem_desc_size
  145. ldr r1, [r1]
  146. bl rt_hw_init_mmu_table
  147. #ifdef RT_USING_USERSPACE
  148. ldr r0, =MMUTable /* vaddr */
  149. add r0, r5 /* to paddr */
  150. bl rt_hw_mmu_switch
  151. #else
  152. bl rt_hw_mmu_init
  153. #endif
  154. /* call C++ constructors of global objects */
  155. ldr r0, =__ctors_start__
  156. ldr r1, =__ctors_end__
  157. ctor_loop:
  158. cmp r0, r1
  159. beq ctor_end
  160. ldr r2, [r0], #4
  161. stmfd sp!, {r0-r1}
  162. mov lr, pc
  163. bx r2
  164. ldmfd sp!, {r0-r1}
  165. b ctor_loop
  166. ctor_end:
  167. /* start RT-Thread Kernel */
  168. ldr pc, _rtthread_startup
  169. _rtthread_startup:
  170. .word rtthread_startup
  171. stack_setup:
  172. #ifdef RT_USING_SMP
  173. /* cpu id */
  174. mrc p15, 0, r0, c0, c0, 5
  175. and r0, r0, #0xf
  176. add r0, r0, #1
  177. #else
  178. mov r0, #1
  179. #endif
  180. cps #Mode_UND
  181. ldr r1, =und_stack_n
  182. add sp, r1, r0, asl #12
  183. cps #Mode_IRQ
  184. ldr r1, =irq_stack_n
  185. add sp, r1, r0, asl #12
  186. cps #Mode_FIQ
  187. ldr r1, =irq_stack_n
  188. add sp, r1, r0, asl #12
  189. cps #Mode_ABT
  190. ldr r1, =abt_stack_n
  191. add sp, r1, r0, asl #12
  192. cps #Mode_SVC
  193. ldr r1, =svc_stack_n
  194. add sp, r1, r0, asl #12
  195. bx lr
  196. #ifdef RT_USING_USERSPACE
  197. .align 2
  198. .global enable_mmu
  199. enable_mmu:
  200. orr r0, #0x18
  201. mcr p15, 0, r0, c2, c0, 0 /* ttbr0 */
  202. mov r0, #(1 << 5) /* PD1=1 */
  203. mcr p15, 0, r0, c2, c0, 2 /* ttbcr */
  204. mov r0, #1
  205. mcr p15, 0, r0, c3, c0, 0 /* dacr */
  206. /* invalid tlb before enable mmu */
  207. mov r0, #0
  208. mcr p15, 0, r0, c8, c7, 0
  209. mcr p15, 0, r0, c7, c5, 0 /* iciallu */
  210. mcr p15, 0, r0, c7, c5, 6 /* bpiall */
  211. mrc p15, 0, r0, c1, c0, 0
  212. orr r0, #((1 << 12) | (1 << 11)) /* instruction cache, branch prediction */
  213. orr r0, #((1 << 2) | (1 << 0)) /* data cache, mmu enable */
  214. mcr p15, 0, r0, c1, c0, 0
  215. dsb
  216. isb
  217. mov pc, lr
  218. .global rt_hw_set_process_id
  219. rt_hw_set_process_id:
  220. LSL r0, r0, #8
  221. MCR p15, 0, r0, c13, c0, 1
  222. mov pc, lr
  223. .global rt_hw_mmu_switch
  224. rt_hw_mmu_switch:
  225. orr r0, #0x18
  226. mcr p15, 0, r0, c2, c0, 0 /* ttbr0 */
  227. /* invalid tlb */
  228. mov r0, #0
  229. mcr p15, 0, r0, c8, c7, 0
  230. mcr p15, 0, r0, c7, c5, 0 /* iciallu */
  231. mcr p15, 0, r0, c7, c5, 6 /* bpiall */
  232. dsb
  233. isb
  234. mov pc, lr
  235. .global rt_hw_mmu_tbl_get
  236. rt_hw_mmu_tbl_get:
  237. mrc p15, 0, r0, c2, c0, 0 /* ttbr0 */
  238. bic r0, #0x18
  239. mov pc, lr
  240. #endif
  241. _halt:
  242. wfe
  243. b _halt
  244. /* exception handlers: undef, swi, padt, dabt, resv, irq, fiq */
  245. .section .text.isr, "ax"
  246. .align 5
  247. .globl vector_fiq
  248. vector_fiq:
  249. stmfd sp!,{r0-r7,lr}
  250. bl rt_hw_trap_fiq
  251. ldmfd sp!,{r0-r7,lr}
  252. subs pc, lr, #4
  253. .globl rt_interrupt_enter
  254. .globl rt_interrupt_leave
  255. .globl rt_thread_switch_interrupt_flag
  256. .globl rt_interrupt_from_thread
  257. .globl rt_interrupt_to_thread
  258. .globl rt_current_thread
  259. .globl vmm_thread
  260. .globl vmm_virq_check
  261. .align 5
  262. .globl vector_irq
  263. vector_irq:
  264. #ifdef RT_USING_SMP
  265. clrex
  266. stmfd sp!, {r0, r1}
  267. cps #Mode_SVC
  268. mov r0, sp /* svc_sp */
  269. mov r1, lr /* svc_lr */
  270. cps #Mode_IRQ
  271. sub lr, #4
  272. stmfd r0!, {r1, lr} /* svc_lr, svc_pc */
  273. stmfd r0!, {r2 - r12}
  274. ldmfd sp!, {r1, r2} /* original r0, r1 */
  275. stmfd r0!, {r1 - r2}
  276. mrs r1, spsr /* original mode */
  277. stmfd r0!, {r1}
  278. #ifdef RT_USING_LWP
  279. stmfd r0, {r13, r14}^ /* usr_sp, usr_lr */
  280. sub r0, #8
  281. #endif
  282. #ifdef RT_USING_FPU
  283. /* fpu context */
  284. vmrs r6, fpexc
  285. tst r6, #(1<<30)
  286. beq 1f
  287. vstmdb r0!, {d0-d15}
  288. vstmdb r0!, {d16-d31}
  289. vmrs r5, fpscr
  290. stmfd r0!, {r5}
  291. 1:
  292. stmfd r0!, {r6}
  293. #endif
  294. /* now irq stack is clean */
  295. /* r0 is task svc_sp */
  296. /* backup r0 -> r8 */
  297. mov r8, r0
  298. cps #Mode_SVC
  299. mov sp, r8
  300. bl rt_interrupt_enter
  301. bl rt_hw_trap_irq
  302. bl rt_interrupt_leave
  303. mov r0, r8
  304. bl rt_scheduler_do_irq_switch
  305. b rt_hw_context_switch_exit
  306. #else
  307. stmfd sp!, {r0-r12,lr}
  308. bl rt_interrupt_enter
  309. bl rt_hw_trap_irq
  310. bl rt_interrupt_leave
  311. /* if rt_thread_switch_interrupt_flag set, jump to
  312. * rt_hw_context_switch_interrupt_do and don't return */
  313. ldr r0, =rt_thread_switch_interrupt_flag
  314. ldr r1, [r0]
  315. cmp r1, #1
  316. beq rt_hw_context_switch_interrupt_do
  317. #ifdef RT_USING_LWP
  318. ldmfd sp!, {r0-r12,lr}
  319. cps #Mode_SVC
  320. push {r0-r12}
  321. mov r7, lr
  322. cps #Mode_IRQ
  323. mrs r4, spsr
  324. sub r5, lr, #4
  325. cps #Mode_SVC
  326. and r6, r4, #0x1f
  327. cmp r6, #0x10
  328. bne 1f
  329. msr spsr_csxf, r4
  330. mov lr, r5
  331. pop {r0-r12}
  332. b ret_to_user
  333. 1:
  334. mov lr, r7
  335. cps #Mode_IRQ
  336. msr spsr_csxf, r4
  337. mov lr, r5
  338. cps #Mode_SVC
  339. pop {r0-r12}
  340. cps #Mode_IRQ
  341. movs pc, lr
  342. #else
  343. ldmfd sp!, {r0-r12,lr}
  344. subs pc, lr, #4
  345. #endif
  346. rt_hw_context_switch_interrupt_do:
  347. mov r1, #0 /* clear flag */
  348. str r1, [r0]
  349. mov r1, sp /* r1 point to {r0-r3} in stack */
  350. add sp, sp, #4*4
  351. ldmfd sp!, {r4-r12,lr} /* reload saved registers */
  352. mrs r0, spsr /* get cpsr of interrupt thread */
  353. sub r2, lr, #4 /* save old task's pc to r2 */
  354. /* Switch to SVC mode with no interrupt. If the usr mode guest is
  355. * interrupted, this will just switch to the stack of kernel space.
  356. * save the registers in kernel space won't trigger data abort. */
  357. msr cpsr_c, #I_Bit|F_Bit|Mode_SVC
  358. stmfd sp!, {r2} /* push old task's pc */
  359. stmfd sp!, {r4-r12,lr} /* push old task's lr,r12-r4 */
  360. ldmfd r1, {r1-r4} /* restore r0-r3 of the interrupt thread */
  361. stmfd sp!, {r1-r4} /* push old task's r0-r3 */
  362. stmfd sp!, {r0} /* push old task's cpsr */
  363. #ifdef RT_USING_LWP
  364. stmfd sp, {r13, r14}^ /*push usr_sp, usr_lr */
  365. sub sp, #8
  366. #endif
  367. #ifdef RT_USING_FPU
  368. /* fpu context */
  369. vmrs r6, fpexc
  370. tst r6, #(1<<30)
  371. beq 1f
  372. vstmdb sp!, {d0-d15}
  373. vstmdb sp!, {d16-d31}
  374. vmrs r5, fpscr
  375. stmfd sp!, {r5}
  376. 1:
  377. stmfd sp!, {r6}
  378. #endif
  379. ldr r4, =rt_interrupt_from_thread
  380. ldr r5, [r4]
  381. str sp, [r5] /* store sp in preempted tasks's TCB */
  382. ldr r6, =rt_interrupt_to_thread
  383. ldr r6, [r6]
  384. ldr sp, [r6] /* get new task's stack pointer */
  385. bl rt_thread_self
  386. #ifdef RT_USING_USERSPACE
  387. mov r4, r0
  388. bl lwp_mmu_switch
  389. mov r0, r4
  390. bl lwp_user_setting_restore
  391. #endif
  392. #ifdef RT_USING_FPU
  393. /* fpu context */
  394. ldmfd sp!, {r6}
  395. vmsr fpexc, r6
  396. tst r6, #(1<<30)
  397. beq 1f
  398. ldmfd sp!, {r5}
  399. vmsr fpscr, r5
  400. vldmia sp!, {d16-d31}
  401. vldmia sp!, {d0-d15}
  402. 1:
  403. #endif
  404. #ifdef RT_USING_LWP
  405. ldmfd sp, {r13, r14}^ /*pop usr_sp, usr_lr */
  406. add sp, #8
  407. #endif
  408. ldmfd sp!, {r4} /* pop new task's cpsr to spsr */
  409. msr spsr_cxsf, r4
  410. #ifdef RT_USING_LWP
  411. and r4, #0x1f
  412. cmp r4, #0x10
  413. bne 1f
  414. ldmfd sp!, {r0-r12,lr}
  415. ldmfd sp!, {lr}
  416. b ret_to_user
  417. 1:
  418. #endif
  419. /* pop new task's r0-r12,lr & pc, copy spsr to cpsr */
  420. ldmfd sp!, {r0-r12,lr,pc}^
  421. #endif
  422. .macro push_svc_reg
  423. sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
  424. stmia sp, {r0 - r12} /* Calling r0-r12 */
  425. mov r0, sp
  426. add sp, sp, #17 * 4
  427. mrs r6, spsr /* Save CPSR */
  428. str lr, [r0, #15*4] /* Push PC */
  429. str r6, [r0, #16*4] /* Push CPSR */
  430. and r1, r6, #0x1f
  431. cmp r1, #0x10
  432. cps #Mode_SYS
  433. streq sp, [r0, #13*4] /* Save calling SP */
  434. streq lr, [r0, #14*4] /* Save calling PC */
  435. cps #Mode_SVC
  436. strne sp, [r0, #13*4] /* Save calling SP */
  437. strne lr, [r0, #14*4] /* Save calling PC */
  438. .endm
  439. .align 5
  440. .weak vector_swi
  441. vector_swi:
  442. push_svc_reg
  443. bl rt_hw_trap_swi
  444. b .
  445. .align 5
  446. .globl vector_undef
  447. vector_undef:
  448. push_svc_reg
  449. bl rt_hw_trap_undef
  450. cps #Mode_UND
  451. #ifdef RT_USING_FPU
  452. sub sp, sp, #17 * 4
  453. ldr lr, [sp, #15*4]
  454. ldmia sp, {r0 - r12}
  455. add sp, sp, #17 * 4
  456. movs pc, lr
  457. #endif
  458. b .
  459. .align 5
  460. .globl vector_pabt
  461. vector_pabt:
  462. push_svc_reg
  463. #ifdef RT_USING_USERSPACE
  464. /* cp Mode_ABT stack to SVC */
  465. sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
  466. mov lr, r0
  467. ldmia lr, {r0 - r12}
  468. stmia sp, {r0 - r12}
  469. add r1, lr, #13 * 4
  470. add r2, sp, #13 * 4
  471. ldmia r1, {r4 - r7}
  472. stmia r2, {r4 - r7}
  473. mov r0, sp
  474. bl rt_hw_trap_pabt
  475. /* return to user */
  476. ldr lr, [sp, #16*4] /* orign spsr */
  477. msr spsr_cxsf, lr
  478. ldr lr, [sp, #15*4] /* orign pc */
  479. ldmia sp, {r0 - r12}
  480. add sp, #17 * 4
  481. b ret_to_user
  482. #else
  483. bl rt_hw_trap_pabt
  484. b .
  485. #endif
  486. .align 5
  487. .globl vector_dabt
  488. vector_dabt:
  489. push_svc_reg
  490. #ifdef RT_USING_USERSPACE
  491. /* cp Mode_ABT stack to SVC */
  492. sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
  493. mov lr, r0
  494. ldmia lr, {r0 - r12}
  495. stmia sp, {r0 - r12}
  496. add r1, lr, #13 * 4
  497. add r2, sp, #13 * 4
  498. ldmia r1, {r4 - r7}
  499. stmia r2, {r4 - r7}
  500. mov r0, sp
  501. bl rt_hw_trap_dabt
  502. /* return to user */
  503. ldr lr, [sp, #16*4] /* orign spsr */
  504. msr spsr_cxsf, lr
  505. ldr lr, [sp, #15*4] /* orign pc */
  506. ldmia sp, {r0 - r12}
  507. add sp, #17 * 4
  508. b ret_to_user
  509. #else
  510. bl rt_hw_trap_dabt
  511. b .
  512. #endif
  513. .align 5
  514. .globl vector_resv
  515. vector_resv:
  516. push_svc_reg
  517. bl rt_hw_trap_resv
  518. b .
  519. #ifdef RT_USING_SMP
  520. .global rt_hw_clz
  521. rt_hw_clz:
  522. clz r0, r0
  523. bx lr
  524. .global rt_secondary_cpu_entry
  525. rt_secondary_cpu_entry:
  526. #ifdef RT_USING_USERSPACE
  527. ldr r5, =PV_OFFSET
  528. ldr lr, =after_enable_mmu_n
  529. ldr r0, =init_mtbl
  530. add r0, r5
  531. b enable_mmu
  532. after_enable_mmu_n:
  533. ldr r0, =MMUTable
  534. add r0, r5
  535. bl rt_hw_mmu_switch
  536. #endif
  537. #ifdef RT_USING_FPU
  538. mov r4, #0xfffffff
  539. mcr p15, 0, r4, c1, c0, 2
  540. #endif
  541. mrc p15, 0, r1, c1, c0, 1
  542. mov r0, #(1<<6)
  543. orr r1, r0
  544. mcr p15, 0, r1, c1, c0, 1 /* enable smp */
  545. mrc p15, 0, r0, c1, c0, 0
  546. bic r0, #(1<<13)
  547. mcr p15, 0, r0, c1, c0, 0
  548. bl stack_setup
  549. /* initialize the mmu table and enable mmu */
  550. #ifndef RT_USING_USERSPACE
  551. bl rt_hw_mmu_init
  552. #endif
  553. b rt_hw_secondary_cpu_bsp_start
  554. #endif
  555. #ifndef RT_CPUS_NR
  556. #define RT_CPUS_NR 1
  557. #endif
  558. .bss
  559. .align 3 /* align to 2~3=8 */
  560. svc_stack_n:
  561. .space (RT_CPUS_NR << 12)
  562. svc_stack_n_limit:
  563. irq_stack_n:
  564. .space (RT_CPUS_NR << 12)
  565. und_stack_n:
  566. .space (RT_CPUS_NR << 12)
  567. abt_stack_n:
  568. .space (RT_CPUS_NR << 12)