start_gcc.S 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695
  1. /*
  2. * Copyright (c) 2006-2018, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Change Logs:
  7. * Date Author Notes
  8. * 2013-07-05 Bernard the first version
  9. * 2018-11-22 Jesven in the interrupt context, use rt_scheduler_do_irq_switch checks
  10. * and switches to a new thread
  11. * 2024-01-16 huanghe restructure this code section following the aarch64 architectural style
  12. */
  13. #include "rtconfig.h"
  14. #define ARM_CPU_STACK_SIZE_OFFSET 12
  15. #define ARM_CPU_STACK_SIZE (1<<ARM_CPU_STACK_SIZE_OFFSET)
  16. .equ Mode_USR, 0x10
  17. .equ Mode_FIQ, 0x11
  18. .equ Mode_IRQ, 0x12
  19. .equ Mode_SVC, 0x13
  20. .equ Mode_ABT, 0x17
  21. .equ Mode_UND, 0x1B
  22. .equ Mode_SYS, 0x1F
  23. .equ I_Bit, 0x80 /* when I bit is set, IRQ is disabled */
  24. .equ F_Bit, 0x40 /* when F bit is set, FIQ is disabled */
  25. /*Load the physical address of a symbol into a register.
  26. Through pv_off calculates the offset of the physical address */
  27. .macro get_phy, reg, symbol, _pvoff
  28. ldr \reg, =\symbol
  29. add \reg, \_pvoff
  30. .endm
  31. /*Calculate the offset between the physical address and the virtual address of the "_reset".*/
  32. .macro get_pvoff, tmp, out
  33. ldr \tmp, =_reset
  34. adr \out, _reset
  35. sub \out, \out, \tmp
  36. .endm
  37. pv_off .req r11 /* Used to store the offset between physical address and the virtual address */
  38. cpu_id .req r10 /* Used to store the cpu id */
  39. /* reset entry */
  40. .globl _reset
  41. _reset:
  42. /* Calculate the offset between the physical address and the virtual address */
  43. get_pvoff r0, pv_off
  44. /* exit hyp mode */
  45. bl init_cpu_mode
  46. /* clear bss section */
  47. bl init_kernel_bss
  48. /* Initializes the assembly environment stack */
  49. bl init_cpu_stack_early
  50. /* init mmu */
  51. b init_mmu_page_table_early
  52. init_cpu_stack_early:
  53. cps #Mode_SVC
  54. get_phy r0, svc_stack_top, pv_off
  55. mov sp, r0
  56. #ifdef RT_USING_FPU
  57. mov r4, #0xfffffff
  58. mcr p15, 0, r4, c1, c0, 2
  59. #endif
  60. mov pc, lr
  61. init_kernel_bss:
  62. /* enable I cache + branch prediction */
  63. mrc p15, 0, r0, c1, c0, 0
  64. orr r0, r0, #(1<<12)
  65. orr r0, r0, #(1<<11)
  66. mcr p15, 0, r0, c1, c0, 0
  67. mov r0,#0 /* get a zero */
  68. get_phy r1, __bss_start, pv_off
  69. get_phy r2, __bss_end, pv_off
  70. bss_loop:
  71. cmp r1,r2 /* check if data to clear */
  72. strlo r0,[r1],#4 /* clear 4 bytes */
  73. blo bss_loop /* loop until done */
  74. mov pc, lr
  75. init_cpu_mode:
  76. #ifdef ARCH_ARMV8
  77. /* Check for HYP mode */
  78. mrs r0, cpsr_all
  79. and r0, r0, #0x1F
  80. mov r8, #0x1A
  81. cmp r0, r8
  82. beq overHyped
  83. b continue_exit
  84. overHyped: /* Get out of HYP mode */
  85. mov r9, lr
  86. /* HYP mode has a dedicated register, called ELR_hyp,
  87. to store the exception return address.
  88. The lr register needs to be temporarily saved,
  89. otherwise "mov pc lr" cannot be used after switching modes. */
  90. adr r1, continue_exit
  91. msr ELR_hyp, r1
  92. mrs r1, cpsr_all
  93. and r1, r1, #0xFFFFFFE0 /* CPSR_MODE_MASK */
  94. orr r1, r1, #0x13 /* CPSR_MODE_SUPERVISOR */
  95. msr SPSR_hyp, r1
  96. eret
  97. continue_exit:
  98. mov lr ,r9
  99. #endif
  100. #ifdef SOC_BCM283x
  101. /* Suspend the other cpu cores */
  102. mrc p15, 0, r0, c0, c0, 5
  103. ands r0, #3
  104. bne _halt
  105. /* Disable IRQ & FIQ */
  106. cpsid if
  107. /* Check for HYP mode */
  108. mrs r0, cpsr_all
  109. and r0, r0, #0x1F
  110. mov r8, #0x1A
  111. cmp r0, r8
  112. beq overHyped
  113. b continue_exit
  114. overHyped: /* Get out of HYP mode */
  115. mov r9, lr
  116. /* HYP mode has a dedicated register, called ELR_hyp,
  117. to store the exception return address.
  118. The lr register needs to be temporarily saved,
  119. otherwise "mov pc lr" cannot be used after switching modes. */
  120. adr r1, continue_exit
  121. msr ELR_hyp, r1
  122. mrs r1, cpsr_all
  123. and r1, r1, #0xFFFFFFE0 /* CPSR_MODE_MASK */
  124. orr r1, r1, #0x13 /* CPSR_MODE_SUPERVISOR */
  125. msr SPSR_hyp, r1
  126. eret
  127. continue_exit:
  128. mov lr ,r9
  129. /* set the cpu to SVC32 mode and disable interrupt */
  130. mrs r0, cpsr
  131. bic r0, r0, #0x1f
  132. orr r0, r0, #0x13
  133. msr cpsr_c, r0
  134. #endif
  135. /* invalid tlb before enable mmu */
  136. mrc p15, 0, r0, c1, c0, 0
  137. bic r0, #1
  138. mcr p15, 0, r0, c1, c0, 0
  139. dsb
  140. isb
  141. mov r0, #0
  142. mcr p15, 0, r0, c8, c7, 0
  143. mcr p15, 0, r0, c7, c5, 0 /* iciallu */
  144. mcr p15, 0, r0, c7, c5, 6 /* bpiall */
  145. dsb
  146. isb
  147. mov pc, lr
  148. init_mmu_page_table_early:
  149. get_phy r0, init_mtbl, pv_off
  150. mov r1, pv_off
  151. bl rt_hw_mem_setup_early
  152. /* get cpu id */
  153. bl rt_hw_cpu_id_early
  154. mov cpu_id ,r0
  155. /* enable_mmu_page_table_early is changed to master_core_startup */
  156. ldr lr, =master_core_startup
  157. cmp cpu_id, #0
  158. beq enable_mmu_page_table_early
  159. #ifdef RT_USING_SMP
  160. #ifdef RT_SMP_AUTO_BOOT
  161. /* if cpu id > 0, stop or wait */
  162. ldr r0, =secondary_cpu_entry
  163. mov r1, #0
  164. str r1, [r0] /* clean secondary_cpu_entry */
  165. #endif
  166. #endif
  167. secondary_loop:
  168. @ cpu core 1 goes into sleep until core 0 wakeup it
  169. wfe
  170. #ifdef RT_SMP_AUTO_BOOT
  171. ldr r1, =secondary_cpu_entry
  172. ldr r0, [r1]
  173. cmp r0, #0
  174. blxne r0 /* if(secondary_cpu_entry) secondary_cpu_entry(); */
  175. #endif /* RT_SMP_AUTO_BOOT */
  176. b secondary_loop
  177. enable_mmu_page_table_early:
  178. /* init TTBR0 */
  179. get_phy r0, init_mtbl, pv_off
  180. mcr p15, #0, r0, c2, c0, #0
  181. dmb
  182. ldr r0,=#0x55555555
  183. mcr p15, #0, r0, c3, c0, #0
  184. /* disable ttbr1 */
  185. mov r0, #(1 << 5) /* PD1=1 */
  186. mcr p15, 0, r0, c2, c0, 2 /* ttbcr */
  187. /* init stack for cpu mod */
  188. cps #Mode_UND
  189. ldr r1,=und_stack_top
  190. sub sp, r1, cpu_id, asl #ARM_CPU_STACK_SIZE_OFFSET
  191. cps #Mode_IRQ
  192. ldr r1, =irq_stack_top
  193. sub sp, r1, cpu_id, asl #ARM_CPU_STACK_SIZE_OFFSET
  194. cps #Mode_FIQ
  195. ldr r1, =irq_stack_top
  196. sub sp, r1, cpu_id, asl #ARM_CPU_STACK_SIZE_OFFSET
  197. cps #Mode_ABT
  198. ldr r1, =abt_stack_top
  199. sub sp, r1, cpu_id, asl #ARM_CPU_STACK_SIZE_OFFSET
  200. cps #Mode_SVC
  201. ldr r1, =svc_stack_top
  202. sub sp, r1, cpu_id, asl #ARM_CPU_STACK_SIZE_OFFSET
  203. /* invalid tlb before enable mmu */
  204. mov r0, #0
  205. mcr p15, 0, r0, c8, c7, 0
  206. mcr p15, 0, r0, c7, c5, 0 /* iciallu */
  207. mcr p15, 0, r0, c7, c5, 6 /* bpiall */
  208. mrc p15, 0, r0, c1, c0, 0
  209. bic r0, r0, #0x7 /* clear bit1~3 */
  210. orr r0, #((1 << 12) | (1 << 11)) /* instruction cache, branch prediction */
  211. orr r0, #((1 << 2) | (1 << 0)) /* data cache, mmu enable */
  212. mcr p15, 0, r0, c1, c0, 0
  213. dsb
  214. isb
  215. mov pc, lr
  216. master_core_startup :
  217. mov r0 ,pv_off
  218. bl rt_kmem_pvoff_set
  219. ldr lr, =rtthread_startup
  220. mov pc, lr
  221. .global rt_hw_mmu_tbl_get
  222. rt_hw_mmu_tbl_get:
  223. mrc p15, 0, r0, c2, c0, 0 /* ttbr0 */
  224. bic r0, #0x18
  225. mov pc, lr
  226. .weak rt_hw_cpu_id_early
  227. rt_hw_cpu_id_early:
  228. mrc p15, 0, r0, c0, c0, 5
  229. and r0, r0, #0xf
  230. mov pc, lr
  231. #ifdef RT_USING_SMP
  232. .global rt_secondary_cpu_entry
  233. rt_secondary_cpu_entry:
  234. ldr r0, =_reset
  235. adr pv_off, _reset
  236. sub pv_off, pv_off, r0
  237. bl init_cpu_stack_early
  238. /* init mmu */
  239. bl rt_hw_cpu_id_early
  240. mov cpu_id ,r0
  241. ldr lr ,= rt_hw_secondary_cpu_bsp_start
  242. b enable_mmu_page_table_early
  243. #endif
  244. /* exception handlers: undef, swi, padt, dabt, resv, irq, fiq */
  245. .section .text.isr, "ax"
  246. .align 5
  247. .globl vector_fiq
  248. vector_fiq:
  249. stmfd sp!,{r0-r7,lr}
  250. bl rt_hw_trap_fiq
  251. ldmfd sp!,{r0-r7,lr}
  252. subs pc, lr, #4
  253. .globl rt_interrupt_enter
  254. .globl rt_interrupt_leave
  255. .globl rt_thread_switch_interrupt_flag
  256. .globl rt_interrupt_from_thread
  257. .globl rt_interrupt_to_thread
  258. .globl rt_current_thread
  259. .globl vmm_thread
  260. .globl vmm_virq_check
  261. .align 5
  262. .globl vector_irq
  263. vector_irq:
  264. #ifdef RT_USING_SMP
  265. stmfd sp!, {r0, r1}
  266. cps #Mode_SVC
  267. mov r0, sp /* svc_sp */
  268. mov r1, lr /* svc_lr */
  269. cps #Mode_IRQ
  270. sub lr, #4
  271. stmfd r0!, {r1, lr} /* svc_lr, svc_pc */
  272. stmfd r0!, {r2 - r12}
  273. ldmfd sp!, {r1, r2} /* original r0, r1 */
  274. stmfd r0!, {r1 - r2}
  275. mrs r1, spsr /* original mode */
  276. stmfd r0!, {r1}
  277. #ifdef RT_USING_SMART
  278. stmfd r0, {r13, r14}^ /* usr_sp, usr_lr */
  279. sub r0, #8
  280. #endif
  281. #ifdef RT_USING_FPU
  282. /* fpu context */
  283. vmrs r6, fpexc
  284. tst r6, #(1<<30)
  285. beq 1f
  286. vstmdb r0!, {d0-d15}
  287. vstmdb r0!, {d16-d31}
  288. vmrs r5, fpscr
  289. stmfd r0!, {r5}
  290. 1:
  291. stmfd r0!, {r6}
  292. #endif
  293. /* now irq stack is clean */
  294. /* r0 is task svc_sp */
  295. /* backup r0 -> r8 */
  296. mov r8, r0
  297. cps #Mode_SVC
  298. mov sp, r8
  299. bl rt_interrupt_enter
  300. bl rt_hw_trap_irq
  301. bl rt_interrupt_leave
  302. mov r0, r8
  303. bl rt_scheduler_do_irq_switch
  304. b rt_hw_context_switch_exit
  305. #else
  306. stmfd sp!, {r0-r12,lr}
  307. bl rt_interrupt_enter
  308. bl rt_hw_trap_irq
  309. bl rt_interrupt_leave
  310. /* if rt_thread_switch_interrupt_flag set, jump to
  311. * rt_hw_context_switch_interrupt_do and don't return */
  312. ldr r0, =rt_thread_switch_interrupt_flag
  313. ldr r1, [r0]
  314. cmp r1, #1
  315. beq rt_hw_context_switch_interrupt_do
  316. #ifdef RT_USING_SMART
  317. ldmfd sp!, {r0-r12,lr}
  318. cps #Mode_SVC
  319. push {r0-r12}
  320. mov r7, lr
  321. cps #Mode_IRQ
  322. mrs r4, spsr
  323. sub r5, lr, #4
  324. cps #Mode_SVC
  325. and r6, r4, #0x1f
  326. cmp r6, #0x10
  327. bne 1f
  328. msr spsr_csxf, r4
  329. mov lr, r5
  330. pop {r0-r12}
  331. b arch_ret_to_user
  332. 1:
  333. mov lr, r7
  334. cps #Mode_IRQ
  335. msr spsr_csxf, r4
  336. mov lr, r5
  337. cps #Mode_SVC
  338. pop {r0-r12}
  339. cps #Mode_IRQ
  340. movs pc, lr
  341. #else
  342. ldmfd sp!, {r0-r12,lr}
  343. subs pc, lr, #4
  344. #endif
  345. rt_hw_context_switch_interrupt_do:
  346. mov r1, #0 /* clear flag */
  347. str r1, [r0]
  348. mov r1, sp /* r1 point to {r0-r3} in stack */
  349. add sp, sp, #4*4
  350. ldmfd sp!, {r4-r12,lr} /* reload saved registers */
  351. mrs r0, spsr /* get cpsr of interrupt thread */
  352. sub r2, lr, #4 /* save old task's pc to r2 */
  353. /* Switch to SVC mode with no interrupt. If the usr mode guest is
  354. * interrupted, this will just switch to the stack of kernel space.
  355. * save the registers in kernel space won't trigger data abort. */
  356. msr cpsr_c, #I_Bit|F_Bit|Mode_SVC
  357. stmfd sp!, {r2} /* push old task's pc */
  358. stmfd sp!, {r4-r12,lr} /* push old task's lr,r12-r4 */
  359. ldmfd r1, {r1-r4} /* restore r0-r3 of the interrupt thread */
  360. stmfd sp!, {r1-r4} /* push old task's r0-r3 */
  361. stmfd sp!, {r0} /* push old task's cpsr */
  362. #ifdef RT_USING_SMART
  363. stmfd sp, {r13, r14}^ /*push usr_sp, usr_lr */
  364. sub sp, #8
  365. #endif
  366. #ifdef RT_USING_FPU
  367. /* fpu context */
  368. vmrs r6, fpexc
  369. tst r6, #(1<<30)
  370. beq 1f
  371. vstmdb sp!, {d0-d15}
  372. vstmdb sp!, {d16-d31}
  373. vmrs r5, fpscr
  374. stmfd sp!, {r5}
  375. 1:
  376. stmfd sp!, {r6}
  377. #endif
  378. ldr r4, =rt_interrupt_from_thread
  379. ldr r5, [r4]
  380. str sp, [r5] /* store sp in preempted tasks's TCB */
  381. ldr r6, =rt_interrupt_to_thread
  382. ldr r6, [r6]
  383. ldr sp, [r6] /* get new task's stack pointer */
  384. #ifdef RT_USING_SMART
  385. bl rt_thread_self
  386. mov r4, r0
  387. bl lwp_aspace_switch
  388. mov r0, r4
  389. bl lwp_user_setting_restore
  390. #endif
  391. #ifdef RT_USING_FPU
  392. /* fpu context */
  393. ldmfd sp!, {r6}
  394. vmsr fpexc, r6
  395. tst r6, #(1<<30)
  396. beq 1f
  397. ldmfd sp!, {r5}
  398. vmsr fpscr, r5
  399. vldmia sp!, {d16-d31}
  400. vldmia sp!, {d0-d15}
  401. 1:
  402. #endif
  403. #ifdef RT_USING_SMART
  404. ldmfd sp, {r13, r14}^ /*pop usr_sp, usr_lr */
  405. add sp, #8
  406. #endif
  407. ldmfd sp!, {r4} /* pop new task's cpsr to spsr */
  408. msr spsr_cxsf, r4
  409. #ifdef RT_USING_SMART
  410. and r4, #0x1f
  411. cmp r4, #0x10
  412. bne 1f
  413. ldmfd sp!, {r0-r12,lr}
  414. ldmfd sp!, {lr}
  415. b arch_ret_to_user
  416. 1:
  417. #endif
  418. /* pop new task's r0-r12,lr & pc, copy spsr to cpsr */
  419. ldmfd sp!, {r0-r12,lr,pc}^
  420. #endif
  421. .macro push_svc_reg
  422. sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
  423. stmia sp, {r0 - r12} /* Calling r0-r12 */
  424. mov r0, sp
  425. add sp, sp, #17 * 4
  426. mrs r6, spsr /* Save CPSR */
  427. str lr, [r0, #15*4] /* Push PC */
  428. str r6, [r0, #16*4] /* Push CPSR */
  429. and r1, r6, #0x1f
  430. cmp r1, #0x10
  431. cps #Mode_SYS
  432. streq sp, [r0, #13*4] /* Save calling SP */
  433. streq lr, [r0, #14*4] /* Save calling PC */
  434. cps #Mode_SVC
  435. strne sp, [r0, #13*4] /* Save calling SP */
  436. strne lr, [r0, #14*4] /* Save calling PC */
  437. .endm
  438. .align 5
  439. .weak vector_swi
  440. vector_swi:
  441. push_svc_reg
  442. bl rt_hw_trap_swi
  443. b .
  444. .align 5
  445. .globl vector_undef
  446. vector_undef:
  447. push_svc_reg
  448. bl rt_hw_trap_undef
  449. #ifdef RT_USING_FPU
  450. cps #Mode_UND
  451. sub sp, sp, #17 * 4
  452. ldr lr, [sp, #15*4]
  453. ldmia sp, {r0 - r12}
  454. add sp, sp, #17 * 4
  455. movs pc, lr
  456. #endif
  457. b .
  458. .align 5
  459. .globl vector_pabt
  460. vector_pabt:
  461. push_svc_reg
  462. #ifdef RT_USING_SMART
  463. /* cp Mode_ABT stack to SVC */
  464. sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
  465. mov lr, r0
  466. ldmia lr, {r0 - r12}
  467. stmia sp, {r0 - r12}
  468. add r1, lr, #13 * 4
  469. add r2, sp, #13 * 4
  470. ldmia r1, {r4 - r7}
  471. stmia r2, {r4 - r7}
  472. mov r0, sp
  473. bl rt_hw_trap_pabt
  474. /* return to user */
  475. ldr lr, [sp, #16*4] /* orign spsr */
  476. msr spsr_cxsf, lr
  477. ldr lr, [sp, #15*4] /* orign pc */
  478. ldmia sp, {r0 - r12}
  479. add sp, #17 * 4
  480. b arch_ret_to_user
  481. #else
  482. bl rt_hw_trap_pabt
  483. b .
  484. #endif
  485. .align 5
  486. .globl vector_dabt
  487. vector_dabt:
  488. push_svc_reg
  489. #ifdef RT_USING_SMART
  490. /* cp Mode_ABT stack to SVC */
  491. sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
  492. mov lr, r0
  493. ldmia lr, {r0 - r12}
  494. stmia sp, {r0 - r12}
  495. add r1, lr, #13 * 4
  496. add r2, sp, #13 * 4
  497. ldmia r1, {r4 - r7}
  498. stmia r2, {r4 - r7}
  499. mov r0, sp
  500. bl rt_hw_trap_dabt
  501. /* return to user */
  502. ldr lr, [sp, #16*4] /* orign spsr */
  503. msr spsr_cxsf, lr
  504. ldr lr, [sp, #15*4] /* orign pc */
  505. ldmia sp, {r0 - r12}
  506. add sp, #17 * 4
  507. b arch_ret_to_user
  508. #else
  509. bl rt_hw_trap_dabt
  510. b .
  511. #endif
  512. .align 5
  513. .globl vector_resv
  514. vector_resv:
  515. push_svc_reg
  516. bl rt_hw_trap_resv
  517. b .
  518. .global rt_hw_clz
  519. rt_hw_clz:
  520. clz r0, r0
  521. bx lr
  522. #include "asm-generic.h"
  523. START_POINT(_thread_start)
  524. mov r10, lr
  525. blx r1
  526. blx r10
  527. b . /* never here */
  528. START_POINT_END(_thread_start)
  529. .data
  530. .align 14
  531. init_mtbl:
  532. .space (4*4096) /* The L1 translation table therefore contains 4096 32-bit (word-sized) entries. */
  533. .global rt_hw_mmu_switch
  534. rt_hw_mmu_switch:
  535. orr r0, #0x18
  536. mcr p15, 0, r0, c2, c0, 0 // ttbr0
  537. //invalid tlb
  538. mov r0, #0
  539. mcr p15, 0, r0, c8, c7, 0
  540. mcr p15, 0, r0, c7, c5, 0 //iciallu
  541. mcr p15, 0, r0, c7, c5, 6 //bpiall
  542. dsb
  543. isb
  544. mov pc, lr
  545. .global rt_hw_set_process_id
  546. rt_hw_set_process_id:
  547. LSL r0, r0, #8
  548. MCR p15, 0, r0, c13, c0, 1
  549. mov pc, lr
  550. .bss
  551. .align 3 /* align to 2~3=8 */
  552. .cpus_stack:
  553. svc_stack_n:
  554. #if defined(RT_USING_SMP) && (RT_CPUS_NR > 1)
  555. .space ((RT_CPUS_NR - 1) * ARM_CPU_STACK_SIZE)
  556. #endif
  557. .space (ARM_CPU_STACK_SIZE)
  558. svc_stack_top:
  559. irq_stack_n:
  560. #if defined(RT_USING_SMP) && (RT_CPUS_NR > 1)
  561. .space ((RT_CPUS_NR - 1) * ARM_CPU_STACK_SIZE)
  562. #endif
  563. .space (ARM_CPU_STACK_SIZE)
  564. irq_stack_top:
  565. und_stack_n:
  566. #if defined(RT_USING_SMP) && (RT_CPUS_NR > 1)
  567. .space ((RT_CPUS_NR - 1) * ARM_CPU_STACK_SIZE)
  568. #endif
  569. .space (ARM_CPU_STACK_SIZE)
  570. und_stack_top:
  571. abt_stack_n:
  572. #if defined(RT_USING_SMP) && (RT_CPUS_NR > 1)
  573. .space ((RT_CPUS_NR - 1) * ARM_CPU_STACK_SIZE)
  574. #endif
  575. .space (ARM_CPU_STACK_SIZE)
  576. abt_stack_top: