start_gcc.S 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707
  1. /*
  2. * Copyright (c) 2006-2018, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Change Logs:
  7. * Date Author Notes
  8. * 2013-07-05 Bernard the first version
  9. * 2018-11-22 Jesven in the interrupt context, use rt_scheduler_do_irq_switch checks
  10. * and switches to a new thread
  11. */
  12. #include "rtconfig.h"
  13. .equ Mode_USR, 0x10
  14. .equ Mode_FIQ, 0x11
  15. .equ Mode_IRQ, 0x12
  16. .equ Mode_SVC, 0x13
  17. .equ Mode_ABT, 0x17
  18. .equ Mode_UND, 0x1B
  19. .equ Mode_SYS, 0x1F
  20. .equ I_Bit, 0x80 /* when I bit is set, IRQ is disabled */
  21. .equ F_Bit, 0x40 /* when F bit is set, FIQ is disabled */
  22. .equ UND_Stack_Size, 0x00000400
  23. .equ SVC_Stack_Size, 0x00000400
  24. .equ ABT_Stack_Size, 0x00000400
  25. .equ RT_FIQ_STACK_PGSZ, 0x00000000
  26. .equ RT_IRQ_STACK_PGSZ, 0x00000800
  27. .equ USR_Stack_Size, 0x00000400
  28. #define ISR_Stack_Size (UND_Stack_Size + SVC_Stack_Size + ABT_Stack_Size + \
  29. RT_FIQ_STACK_PGSZ + RT_IRQ_STACK_PGSZ)
  30. .section .data.share.isr
  31. /* stack */
  32. .globl stack_start
  33. .globl stack_top
  34. stack_start:
  35. .rept ISR_Stack_Size
  36. .byte 0
  37. .endr
  38. stack_top:
  39. #ifdef RT_USING_USERSPACE
  40. .data
  41. .align 14
  42. init_mtbl:
  43. .space 16*1024
  44. #endif
  45. .text
  46. /* reset entry */
  47. .globl _reset
  48. _reset:
  49. #ifdef ARCH_ARMV8
  50. /* Check for HYP mode */
  51. mrs r0, cpsr_all
  52. and r0, r0, #0x1F
  53. mov r8, #0x1A
  54. cmp r0, r8
  55. beq overHyped
  56. b continue
  57. overHyped: /* Get out of HYP mode */
  58. adr r1, continue
  59. msr ELR_hyp, r1
  60. mrs r1, cpsr_all
  61. and r1, r1, #0x1f /* CPSR_MODE_MASK */
  62. orr r1, r1, #0x13 /* CPSR_MODE_SUPERVISOR */
  63. msr SPSR_hyp, r1
  64. eret
  65. continue:
  66. #endif
  67. #ifdef SOC_BCM283x
  68. /* Suspend the other cpu cores */
  69. mrc p15, 0, r0, c0, c0, 5
  70. ands r0, #3
  71. bne _halt
  72. /* Disable IRQ & FIQ */
  73. cpsid if
  74. /* Check for HYP mode */
  75. mrs r0, cpsr_all
  76. and r0, r0, #0x1F
  77. mov r8, #0x1A
  78. cmp r0, r8
  79. beq overHyped
  80. b continue
  81. overHyped: /* Get out of HYP mode */
  82. adr r1, continue
  83. msr ELR_hyp, r1
  84. mrs r1, cpsr_all
  85. and r1, r1, #0x1f /* CPSR_MODE_MASK */
  86. orr r1, r1, #0x13 /* CPSR_MODE_SUPERVISOR */
  87. msr SPSR_hyp, r1
  88. eret
  89. continue:
  90. /* set the cpu to SVC32 mode and disable interrupt */
  91. mrs r0, cpsr
  92. bic r0, r0, #0x1f
  93. orr r0, r0, #0x13
  94. msr cpsr_c, r0
  95. #endif
  96. /* invalid tlb before enable mmu */
  97. mrc p15, 0, r0, c1, c0, 0
  98. bic r0, #1
  99. mcr p15, 0, r0, c1, c0, 0
  100. dsb
  101. isb
  102. mov r0, #0
  103. mcr p15, 0, r0, c8, c7, 0
  104. mcr p15, 0, r0, c7, c5, 0 /* iciallu */
  105. mcr p15, 0, r0, c7, c5, 6 /* bpiall */
  106. dsb
  107. isb
  108. #ifdef RT_USING_USERSPACE
  109. ldr r5, =PV_OFFSET
  110. mov r7, #0x100000
  111. sub r7, #1
  112. mvn r8, r7
  113. ldr r9, =KERNEL_VADDR_START
  114. ldr r6, =__bss_end
  115. add r6, r7
  116. and r6, r8 /* r6 end vaddr align up to 1M */
  117. sub r6, r9 /* r6 is size */
  118. ldr sp, =stack_top
  119. add sp, r5 /* use paddr */
  120. ldr r0, =init_mtbl
  121. add r0, r5
  122. mov r1, r6
  123. mov r2, r5
  124. bl init_mm_setup
  125. ldr lr, =after_enable_mmu
  126. ldr r0, =init_mtbl
  127. add r0, r5
  128. b enable_mmu
  129. after_enable_mmu:
  130. #endif
  131. #ifndef SOC_BCM283x
  132. /* set the cpu to SVC32 mode and disable interrupt */
  133. cps #Mode_SVC
  134. #endif
  135. #ifdef RT_USING_FPU
  136. mov r4, #0xfffffff
  137. mcr p15, 0, r4, c1, c0, 2
  138. #endif
  139. /* disable the data alignment check */
  140. mrc p15, 0, r1, c1, c0, 0
  141. bic r1, #(1<<1)
  142. mcr p15, 0, r1, c1, c0, 0
  143. /* setup stack */
  144. bl stack_setup
  145. /* clear .bss */
  146. mov r0,#0 /* get a zero */
  147. ldr r1,=__bss_start /* bss start */
  148. ldr r2,=__bss_end /* bss end */
  149. bss_loop:
  150. cmp r1,r2 /* check if data to clear */
  151. strlo r0,[r1],#4 /* clear 4 bytes */
  152. blo bss_loop /* loop until done */
  153. #ifdef RT_USING_SMP
  154. mrc p15, 0, r1, c1, c0, 1
  155. mov r0, #(1<<6)
  156. orr r1, r0
  157. mcr p15, 0, r1, c1, c0, 1 /* enable smp */
  158. #endif
  159. /* initialize the mmu table and enable mmu */
  160. ldr r0, =platform_mem_desc
  161. ldr r1, =platform_mem_desc_size
  162. ldr r1, [r1]
  163. bl rt_hw_init_mmu_table
  164. #ifdef RT_USING_USERSPACE
  165. ldr r0, =MMUTable /* vaddr */
  166. add r0, r5 /* to paddr */
  167. bl switch_mmu
  168. #else
  169. bl rt_hw_mmu_init
  170. #endif
  171. /* call C++ constructors of global objects */
  172. ldr r0, =__ctors_start__
  173. ldr r1, =__ctors_end__
  174. ctor_loop:
  175. cmp r0, r1
  176. beq ctor_end
  177. ldr r2, [r0], #4
  178. stmfd sp!, {r0-r1}
  179. mov lr, pc
  180. bx r2
  181. ldmfd sp!, {r0-r1}
  182. b ctor_loop
  183. ctor_end:
  184. /* start RT-Thread Kernel */
  185. ldr pc, _rtthread_startup
  186. _rtthread_startup:
  187. .word rtthread_startup
  188. stack_setup:
  189. ldr r0, =stack_top
  190. /* Set the startup stack for svc */
  191. mov sp, r0
  192. /* Enter Undefined Instruction Mode and set its Stack Pointer */
  193. msr cpsr_c, #Mode_UND|I_Bit|F_Bit
  194. mov sp, r0
  195. sub r0, r0, #UND_Stack_Size
  196. /* Enter Abort Mode and set its Stack Pointer */
  197. msr cpsr_c, #Mode_ABT|I_Bit|F_Bit
  198. mov sp, r0
  199. sub r0, r0, #ABT_Stack_Size
  200. /* Enter FIQ Mode and set its Stack Pointer */
  201. msr cpsr_c, #Mode_FIQ|I_Bit|F_Bit
  202. mov sp, r0
  203. sub r0, r0, #RT_FIQ_STACK_PGSZ
  204. /* Enter IRQ Mode and set its Stack Pointer */
  205. msr cpsr_c, #Mode_IRQ|I_Bit|F_Bit
  206. mov sp, r0
  207. sub r0, r0, #RT_IRQ_STACK_PGSZ
  208. /* come back to SVC mode */
  209. msr cpsr_c, #Mode_SVC|I_Bit|F_Bit
  210. bx lr
  211. #ifdef RT_USING_USERSPACE
  212. .align 2
  213. .global enable_mmu
  214. enable_mmu:
  215. orr r0, #0x18
  216. mcr p15, 0, r0, c2, c0, 0 /* ttbr0 */
  217. mov r0, #(1 << 5) /* PD1=1 */
  218. mcr p15, 0, r0, c2, c0, 2 /* ttbcr */
  219. mov r0, #1
  220. mcr p15, 0, r0, c3, c0, 0 /* dacr */
  221. /* invalid tlb before enable mmu */
  222. mov r0, #0
  223. mcr p15, 0, r0, c8, c7, 0
  224. mcr p15, 0, r0, c7, c5, 0 /* iciallu */
  225. mcr p15, 0, r0, c7, c5, 6 /* bpiall */
  226. mrc p15, 0, r0, c1, c0, 0
  227. orr r0, #((1 << 12) | (1 << 11)) /* instruction cache, branch prediction */
  228. orr r0, #((1 << 2) | (1 << 0)) /* data cache, mmu enable */
  229. mcr p15, 0, r0, c1, c0, 0
  230. dsb
  231. isb
  232. mov pc, lr
  233. .global set_process_id
  234. set_process_id:
  235. MCR p15, 0, r0, c13, c0, 1
  236. mov pc, lr
  237. .global switch_mmu
  238. switch_mmu:
  239. orr r0, #0x18
  240. mcr p15, 0, r0, c2, c0, 0 /* ttbr0 */
  241. /* invalid tlb */
  242. mov r0, #0
  243. mcr p15, 0, r0, c8, c7, 0
  244. mcr p15, 0, r0, c7, c5, 0 /* iciallu */
  245. mcr p15, 0, r0, c7, c5, 6 /* bpiall */
  246. dsb
  247. isb
  248. mov pc, lr
  249. .global mmu_table_get
  250. mmu_table_get:
  251. mrc p15, 0, r0, c2, c0, 0 /* ttbr0 */
  252. bic r0, #0x18
  253. mov pc, lr
  254. #endif
  255. _halt:
  256. wfe
  257. b _halt
  258. /* exception handlers: undef, swi, padt, dabt, resv, irq, fiq */
  259. .section .text.isr, "ax"
  260. .align 5
  261. .globl vector_fiq
  262. vector_fiq:
  263. stmfd sp!,{r0-r7,lr}
  264. bl rt_hw_trap_fiq
  265. ldmfd sp!,{r0-r7,lr}
  266. subs pc, lr, #4
  267. .globl rt_interrupt_enter
  268. .globl rt_interrupt_leave
  269. .globl rt_thread_switch_interrupt_flag
  270. .globl rt_interrupt_from_thread
  271. .globl rt_interrupt_to_thread
  272. .globl rt_current_thread
  273. .globl vmm_thread
  274. .globl vmm_virq_check
  275. .align 5
  276. .globl vector_irq
  277. vector_irq:
  278. #ifdef RT_USING_SMP
  279. clrex
  280. stmfd sp!, {r0, r1}
  281. cps #Mode_SVC
  282. mov r0, sp /* svc_sp */
  283. mov r1, lr /* svc_lr */
  284. cps #Mode_IRQ
  285. sub lr, #4
  286. stmfd r0!, {r1, lr} /* svc_lr, svc_pc */
  287. stmfd r0!, {r2 - r12}
  288. ldmfd sp!, {r1, r2} /* original r0, r1 */
  289. stmfd r0!, {r1 - r2}
  290. mrs r1, spsr /* original mode */
  291. stmfd r0!, {r1}
  292. #ifdef RT_USING_LWP
  293. stmfd r0, {r13, r14}^ /* usr_sp, usr_lr */
  294. sub r0, #8
  295. #endif
  296. #ifdef RT_USING_FPU
  297. /* fpu context */
  298. vmrs r6, fpexc
  299. tst r6, #(1<<30)
  300. beq 1f
  301. vstmdb r0!, {d0-d15}
  302. vstmdb r0!, {d16-d31}
  303. vmrs r5, fpscr
  304. stmfd r0!, {r5}
  305. 1:
  306. stmfd r0!, {r6}
  307. #endif
  308. /* now irq stack is clean */
  309. /* r0 is task svc_sp */
  310. /* backup r0 -> r8 */
  311. mov r8, r0
  312. cps #Mode_SVC
  313. mov sp, r8
  314. bl rt_interrupt_enter
  315. bl rt_hw_trap_irq
  316. bl rt_interrupt_leave
  317. mov r0, r8
  318. bl rt_scheduler_do_irq_switch
  319. b rt_hw_context_switch_exit
  320. #else
  321. stmfd sp!, {r0-r12,lr}
  322. bl rt_interrupt_enter
  323. bl rt_hw_trap_irq
  324. bl rt_interrupt_leave
  325. /* if rt_thread_switch_interrupt_flag set, jump to
  326. * rt_hw_context_switch_interrupt_do and don't return */
  327. ldr r0, =rt_thread_switch_interrupt_flag
  328. ldr r1, [r0]
  329. cmp r1, #1
  330. beq rt_hw_context_switch_interrupt_do
  331. #ifdef RT_USING_LWP
  332. ldmfd sp!, {r0-r12,lr}
  333. cps #Mode_SVC
  334. push {r0-r12}
  335. mov r7, lr
  336. cps #Mode_IRQ
  337. mrs r4, spsr
  338. sub r5, lr, #4
  339. cps #Mode_SVC
  340. bl lwp_check_exit
  341. and r6, r4, #0x1f
  342. cmp r6, #0x10
  343. bne 1f
  344. msr spsr_csxf, r4
  345. mov lr, r5
  346. pop {r0-r12}
  347. b ret_to_user
  348. 1:
  349. mov lr, r7
  350. cps #Mode_IRQ
  351. msr spsr_csxf, r4
  352. mov lr, r5
  353. cps #Mode_SVC
  354. pop {r0-r12}
  355. cps #Mode_IRQ
  356. movs pc, lr
  357. #else
  358. ldmfd sp!, {r0-r12,lr}
  359. subs pc, lr, #4
  360. #endif
  361. rt_hw_context_switch_interrupt_do:
  362. mov r1, #0 /* clear flag */
  363. str r1, [r0]
  364. mov r1, sp /* r1 point to {r0-r3} in stack */
  365. add sp, sp, #4*4
  366. ldmfd sp!, {r4-r12,lr} /* reload saved registers */
  367. mrs r0, spsr /* get cpsr of interrupt thread */
  368. sub r2, lr, #4 /* save old task's pc to r2 */
  369. /* Switch to SVC mode with no interrupt. If the usr mode guest is
  370. * interrupted, this will just switch to the stack of kernel space.
  371. * save the registers in kernel space won't trigger data abort. */
  372. msr cpsr_c, #I_Bit|F_Bit|Mode_SVC
  373. stmfd sp!, {r2} /* push old task's pc */
  374. stmfd sp!, {r4-r12,lr} /* push old task's lr,r12-r4 */
  375. ldmfd r1, {r1-r4} /* restore r0-r3 of the interrupt thread */
  376. stmfd sp!, {r1-r4} /* push old task's r0-r3 */
  377. stmfd sp!, {r0} /* push old task's cpsr */
  378. #ifdef RT_USING_LWP
  379. stmfd sp, {r13, r14}^ /*push usr_sp, usr_lr */
  380. sub sp, #8
  381. #endif
  382. #ifdef RT_USING_FPU
  383. /* fpu context */
  384. vmrs r6, fpexc
  385. tst r6, #(1<<30)
  386. beq 1f
  387. vstmdb sp!, {d0-d15}
  388. vstmdb sp!, {d16-d31}
  389. vmrs r5, fpscr
  390. stmfd sp!, {r5}
  391. 1:
  392. stmfd sp!, {r6}
  393. #endif
  394. ldr r4, =rt_interrupt_from_thread
  395. ldr r5, [r4]
  396. str sp, [r5] /* store sp in preempted tasks's TCB */
  397. ldr r6, =rt_interrupt_to_thread
  398. ldr r6, [r6]
  399. ldr sp, [r6] /* get new task's stack pointer */
  400. bl rt_thread_self
  401. #ifdef RT_USING_USERSPACE
  402. mov r4, r0
  403. bl lwp_mmu_switch
  404. mov r0, r4
  405. bl lwp_user_setting_restore
  406. #endif
  407. #ifdef RT_USING_FPU
  408. /* fpu context */
  409. ldmfd sp!, {r6}
  410. vmsr fpexc, r6
  411. tst r6, #(1<<30)
  412. beq 1f
  413. ldmfd sp!, {r5}
  414. vmsr fpscr, r5
  415. vldmia sp!, {d16-d31}
  416. vldmia sp!, {d0-d15}
  417. 1:
  418. #endif
  419. #ifdef RT_USING_LWP
  420. ldmfd sp, {r13, r14}^ /*pop usr_sp, usr_lr */
  421. add sp, #8
  422. #endif
  423. ldmfd sp!, {r4} /* pop new task's cpsr to spsr */
  424. msr spsr_cxsf, r4
  425. #ifdef RT_USING_GDBSERVER
  426. bl lwp_check_debug
  427. #endif
  428. #ifdef RT_USING_LWP
  429. bl lwp_check_exit
  430. #endif
  431. #ifdef RT_USING_LWP
  432. and r4, #0x1f
  433. cmp r4, #0x10
  434. bne 1f
  435. ldmfd sp!, {r0-r12,lr}
  436. ldmfd sp!, {lr}
  437. b ret_to_user
  438. 1:
  439. #endif
  440. /* pop new task's r0-r12,lr & pc, copy spsr to cpsr */
  441. ldmfd sp!, {r0-r12,lr,pc}^
  442. #endif
  443. .macro push_svc_reg
  444. sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
  445. stmia sp, {r0 - r12} /* Calling r0-r12 */
  446. mov r0, sp
  447. add sp, sp, #17 * 4
  448. mrs r6, spsr /* Save CPSR */
  449. str lr, [r0, #15*4] /* Push PC */
  450. str r6, [r0, #16*4] /* Push CPSR */
  451. and r1, r6, #0x1f
  452. cmp r1, #0x10
  453. cps #Mode_SYS
  454. streq sp, [r0, #13*4] /* Save calling SP */
  455. streq lr, [r0, #14*4] /* Save calling PC */
  456. cps #Mode_SVC
  457. strne sp, [r0, #13*4] /* Save calling SP */
  458. strne lr, [r0, #14*4] /* Save calling PC */
  459. .endm
  460. .align 5
  461. .weak vector_swi
  462. vector_swi:
  463. push_svc_reg
  464. bl rt_hw_trap_swi
  465. b .
  466. .align 5
  467. .globl vector_undef
  468. vector_undef:
  469. push_svc_reg
  470. bl rt_hw_trap_undef
  471. cps #Mode_UND
  472. #ifdef RT_USING_FPU
  473. sub sp, sp, #17 * 4
  474. ldr lr, [sp, #15*4]
  475. ldmia sp, {r0 - r12}
  476. add sp, sp, #17 * 4
  477. movs pc, lr
  478. #endif
  479. b .
  480. .align 5
  481. .globl vector_pabt
  482. vector_pabt:
  483. push_svc_reg
  484. #ifdef RT_USING_USERSPACE
  485. /* cp Mode_ABT stack to SVC */
  486. sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
  487. mov lr, r0
  488. ldmia lr, {r0 - r12}
  489. stmia sp, {r0 - r12}
  490. add r1, lr, #13 * 4
  491. add r2, sp, #13 * 4
  492. ldmia r1, {r4 - r7}
  493. stmia r2, {r4 - r7}
  494. mov r0, sp
  495. bl rt_hw_trap_pabt
  496. /* return to user */
  497. ldr lr, [sp, #16*4] /* orign spsr */
  498. msr spsr_cxsf, lr
  499. ldr lr, [sp, #15*4] /* orign pc */
  500. ldmia sp, {r0 - r12}
  501. add sp, #17 * 4
  502. b ret_to_user
  503. #else
  504. bl rt_hw_trap_pabt
  505. b .
  506. #endif
  507. .align 5
  508. .globl vector_dabt
  509. vector_dabt:
  510. push_svc_reg
  511. #ifdef RT_USING_USERSPACE
  512. /* cp Mode_ABT stack to SVC */
  513. sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
  514. mov lr, r0
  515. ldmia lr, {r0 - r12}
  516. stmia sp, {r0 - r12}
  517. add r1, lr, #13 * 4
  518. add r2, sp, #13 * 4
  519. ldmia r1, {r4 - r7}
  520. stmia r2, {r4 - r7}
  521. mov r0, sp
  522. bl rt_hw_trap_dabt
  523. /* return to user */
  524. ldr lr, [sp, #16*4] /* orign spsr */
  525. msr spsr_cxsf, lr
  526. ldr lr, [sp, #15*4] /* orign pc */
  527. ldmia sp, {r0 - r12}
  528. add sp, #17 * 4
  529. b ret_to_user
  530. #else
  531. bl rt_hw_trap_dabt
  532. b .
  533. #endif
  534. .align 5
  535. .globl vector_resv
  536. vector_resv:
  537. push_svc_reg
  538. bl rt_hw_trap_resv
  539. b .
  540. #ifdef RT_USING_SMP
  541. .global rt_clz
  542. rt_clz:
  543. clz r0, r0
  544. bx lr
  545. .global rt_secondary_cpu_entry
  546. rt_secondary_cpu_entry:
  547. #ifdef RT_USING_USERSPACE
  548. ldr r5, =PV_OFFSET
  549. ldr lr, =after_enable_mmu2
  550. ldr r0, =init_mtbl
  551. add r0, r5
  552. b enable_mmu
  553. after_enable_mmu2:
  554. ldr r0, =MMUTable
  555. add r0, r5
  556. bl switch_mmu
  557. #endif
  558. #ifdef RT_USING_FPU
  559. mov r4, #0xfffffff
  560. mcr p15, 0, r4, c1, c0, 2
  561. #endif
  562. mrc p15, 0, r1, c1, c0, 1
  563. mov r0, #(1<<6)
  564. orr r1, r0
  565. mcr p15, 0, r1, c1, c0, 1 /* enable smp */
  566. mrc p15, 0, r0, c1, c0, 0
  567. bic r0, #(1<<13)
  568. mcr p15, 0, r0, c1, c0, 0
  569. cps #Mode_UND
  570. ldr sp, =und_stack_2_limit
  571. cps #Mode_IRQ
  572. ldr sp, =irq_stack_2_limit
  573. cps #Mode_FIQ
  574. ldr sp, =irq_stack_2_limit
  575. cps #Mode_SVC
  576. ldr sp, =svc_stack_2_limit
  577. cps #Mode_ABT
  578. ldr sp, =abt_stack_2_limit
  579. /* initialize the mmu table and enable mmu */
  580. #ifndef RT_USING_USERSPACE
  581. bl rt_hw_mmu_init
  582. #endif
  583. b rt_hw_secondary_cpu_bsp_start
  584. #endif
  585. .bss
  586. .align 2 /* align to 2~2=4 */
  587. svc_stack_2:
  588. .space (1 << 10)
  589. svc_stack_2_limit:
  590. irq_stack_2:
  591. .space (1 << 10)
  592. irq_stack_2_limit:
  593. und_stack_2:
  594. .space (1 << 10)
  595. und_stack_2_limit:
  596. abt_stack_2:
  597. .space (1 << 10)
  598. abt_stack_2_limit: