start_gcc.S 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727
  1. /*
  2. * Copyright (c) 2006-2018, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Change Logs:
  7. * Date Author Notes
  8. * 2013-07-05 Bernard the first version
  9. * 2018-11-22 Jesven in the interrupt context, use rt_scheduler_do_irq_switch checks
  10. * and switches to a new thread
  11. */
  12. #include "rtconfig.h"
  13. .equ Mode_USR, 0x10
  14. .equ Mode_FIQ, 0x11
  15. .equ Mode_IRQ, 0x12
  16. .equ Mode_SVC, 0x13
  17. .equ Mode_ABT, 0x17
  18. .equ Mode_UND, 0x1B
  19. .equ Mode_SYS, 0x1F
  20. .equ I_Bit, 0x80 /* when I bit is set, IRQ is disabled */
  21. .equ F_Bit, 0x40 /* when F bit is set, FIQ is disabled */
  22. #ifdef RT_USING_SMART
  23. .data
  24. .align 14
  25. init_mtbl:
  26. .space 16*1024
  27. #endif
  28. .text
  29. /* reset entry */
  30. .globl _reset
  31. _reset:
  32. #ifdef ARCH_ARMV8
  33. /* Check for HYP mode */
  34. mrs r0, cpsr_all
  35. and r0, r0, #0x1F
  36. mov r8, #0x1A
  37. cmp r0, r8
  38. beq overHyped
  39. b continue
  40. overHyped: /* Get out of HYP mode */
  41. adr r1, continue
  42. msr ELR_hyp, r1
  43. mrs r1, cpsr_all
  44. and r1, r1, #0x1f /* CPSR_MODE_MASK */
  45. orr r1, r1, #0x13 /* CPSR_MODE_SUPERVISOR */
  46. msr SPSR_hyp, r1
  47. eret
  48. continue:
  49. #endif
  50. #ifdef SOC_BCM283x
  51. /* Suspend the other cpu cores */
  52. mrc p15, 0, r0, c0, c0, 5
  53. ands r0, #3
  54. bne _halt
  55. /* Disable IRQ & FIQ */
  56. cpsid if
  57. /* Check for HYP mode */
  58. mrs r0, cpsr_all
  59. and r0, r0, #0x1F
  60. mov r8, #0x1A
  61. cmp r0, r8
  62. beq overHyped
  63. b continue
  64. overHyped: /* Get out of HYP mode */
  65. adr r1, continue
  66. msr ELR_hyp, r1
  67. mrs r1, cpsr_all
  68. and r1, r1, #0x1f /* CPSR_MODE_MASK */
  69. orr r1, r1, #0x13 /* CPSR_MODE_SUPERVISOR */
  70. msr SPSR_hyp, r1
  71. eret
  72. continue:
  73. /* set the cpu to SVC32 mode and disable interrupt */
  74. mrs r0, cpsr
  75. bic r0, r0, #0x1f
  76. orr r0, r0, #0x13
  77. msr cpsr_c, r0
  78. #endif
  79. /* invalid tlb before enable mmu */
  80. mrc p15, 0, r0, c1, c0, 0
  81. bic r0, #1
  82. mcr p15, 0, r0, c1, c0, 0
  83. dsb
  84. isb
  85. mov r0, #0
  86. mcr p15, 0, r0, c8, c7, 0
  87. mcr p15, 0, r0, c7, c5, 0 /* iciallu */
  88. mcr p15, 0, r0, c7, c5, 6 /* bpiall */
  89. dsb
  90. isb
  91. #ifdef RT_USING_SMART
  92. /* load r5 with PV_OFFSET */
  93. ldr r7, =_reset
  94. adr r5, _reset
  95. sub r5, r5, r7
  96. mov r7, #0x100000
  97. sub r7, #1
  98. mvn r8, r7
  99. ldr r9, =KERNEL_VADDR_START
  100. ldr r6, =__bss_end
  101. add r6, r7
  102. and r6, r8 /* r6 end vaddr align up to 1M */
  103. sub r6, r9 /* r6 is size */
  104. ldr sp, =svc_stack_n_limit
  105. add sp, r5 /* use paddr */
  106. ldr r0, =init_mtbl
  107. add r0, r5
  108. mov r1, r6
  109. mov r2, r5
  110. bl init_mm_setup
  111. ldr lr, =after_enable_mmu
  112. ldr r0, =init_mtbl
  113. add r0, r5
  114. b enable_mmu
  115. after_enable_mmu:
  116. #endif
  117. #ifndef SOC_BCM283x
  118. /* set the cpu to SVC32 mode and disable interrupt */
  119. cps #Mode_SVC
  120. #endif
  121. #ifdef RT_USING_FPU
  122. mov r4, #0xfffffff
  123. mcr p15, 0, r4, c1, c0, 2
  124. #endif
  125. /* disable the data alignment check */
  126. mrc p15, 0, r1, c1, c0, 0
  127. bic r1, #(1<<1) /* Disable Alignment fault checking */
  128. #ifndef RT_USING_SMART
  129. bic r1, #(1<<0) /* Disable MMU */
  130. bic r1, #(1<<2) /* Disable data cache */
  131. bic r1, #(1<<11) /* Disable program flow prediction */
  132. bic r1, #(1<<12) /* Disable instruction cache */
  133. bic r1, #(3<<19) /* bit[20:19] must be zero */
  134. #endif /* RT_USING_SMART */
  135. mcr p15, 0, r1, c1, c0, 0
  136. #ifndef RT_USING_SMART
  137. #ifdef RT_USING_SMP
  138. /* Use spin-table to start secondary cores */
  139. @ get cpu id, and subtract the offset from the stacks base address
  140. bl rt_hw_cpu_id
  141. mov r5, r0
  142. cmp r5, #0 @ cpu id == 0
  143. beq normal_setup
  144. @ cpu id > 0, stop or wait
  145. #ifdef RT_SMP_AUTO_BOOT
  146. ldr r0, =secondary_cpu_entry
  147. mov r1, #0
  148. str r1, [r0] /* clean secondary_cpu_entry */
  149. #endif /* RT_SMP_AUTO_BOOT */
  150. secondary_loop:
  151. @ cpu core 1 goes into sleep until core 0 wakeup it
  152. wfe
  153. #ifdef RT_SMP_AUTO_BOOT
  154. ldr r1, =secondary_cpu_entry
  155. ldr r0, [r1]
  156. cmp r0, #0
  157. blxne r0 /* if(secondary_cpu_entry) secondary_cpu_entry(); */
  158. #endif /* RT_SMP_AUTO_BOOT */
  159. b secondary_loop
  160. normal_setup:
  161. #endif /* RT_USING_SMP */
  162. #endif /* RT_USING_SMART */
  163. /* enable I cache + branch prediction */
  164. mrc p15, 0, r0, c1, c0, 0
  165. orr r0, r0, #(1<<12)
  166. orr r0, r0, #(1<<11)
  167. mcr p15, 0, r0, c1, c0, 0
  168. /* setup stack */
  169. bl stack_setup
  170. /* clear .bss */
  171. mov r0,#0 /* get a zero */
  172. ldr r1,=__bss_start /* bss start */
  173. ldr r2,=__bss_end /* bss end */
  174. bss_loop:
  175. cmp r1,r2 /* check if data to clear */
  176. strlo r0,[r1],#4 /* clear 4 bytes */
  177. blo bss_loop /* loop until done */
  178. mov r0, r5
  179. bl rt_kmem_pvoff_set
  180. #ifdef RT_USING_SMP
  181. mrc p15, 0, r1, c1, c0, 1
  182. mov r0, #(1<<6)
  183. orr r1, r0
  184. mcr p15, 0, r1, c1, c0, 1 /* enable smp */
  185. #endif
  186. /**
  187. * void rt_hw_init_mmu_table(struct mem_desc *mdesc, rt_uint32_t size)
  188. * initialize the mmu table and enable mmu
  189. */
  190. ldr r0, =platform_mem_desc
  191. ldr r1, =platform_mem_desc_size
  192. ldr r1, [r1]
  193. bl rt_hw_init_mmu_table
  194. #ifdef RT_USING_SMART
  195. ldr r0, =MMUTable /* vaddr */
  196. add r0, r5 /* to paddr */
  197. bl rt_hw_mmu_switch
  198. #else
  199. bl rt_hw_mmu_init
  200. #endif
  201. /* start RT-Thread Kernel */
  202. ldr pc, _rtthread_startup
  203. _rtthread_startup:
  204. .word rtthread_startup
  205. .weak rt_asm_cpu_id
  206. rt_asm_cpu_id:
  207. mrc p15, 0, r0, c0, c0, 5
  208. and r0, r0, #0xf
  209. mov pc, lr
  210. stack_setup:
  211. #ifdef RT_USING_SMP
  212. /* cpu id */
  213. mov r10, lr
  214. bl rt_asm_cpu_id
  215. mov lr, r10
  216. add r0, r0, #1
  217. #else
  218. mov r0, #1
  219. #endif
  220. cps #Mode_UND
  221. ldr r1, =und_stack_n
  222. add sp, r1, r0, asl #12
  223. cps #Mode_IRQ
  224. ldr r1, =irq_stack_n
  225. add sp, r1, r0, asl #12
  226. cps #Mode_FIQ
  227. ldr r1, =irq_stack_n
  228. add sp, r1, r0, asl #12
  229. cps #Mode_ABT
  230. ldr r1, =abt_stack_n
  231. add sp, r1, r0, asl #12
  232. cps #Mode_SVC
  233. ldr r1, =svc_stack_n
  234. add sp, r1, r0, asl #12
  235. bx lr
  236. #ifdef RT_USING_SMART
  237. .align 2
  238. .global enable_mmu
  239. enable_mmu:
  240. orr r0, #0x18
  241. mcr p15, 0, r0, c2, c0, 0 /* ttbr0 */
  242. mov r0, #(1 << 5) /* PD1=1 */
  243. mcr p15, 0, r0, c2, c0, 2 /* ttbcr */
  244. mov r0, #1
  245. mcr p15, 0, r0, c3, c0, 0 /* dacr */
  246. /* invalid tlb before enable mmu */
  247. mov r0, #0
  248. mcr p15, 0, r0, c8, c7, 0
  249. mcr p15, 0, r0, c7, c5, 0 /* iciallu */
  250. mcr p15, 0, r0, c7, c5, 6 /* bpiall */
  251. mrc p15, 0, r0, c1, c0, 0
  252. orr r0, #((1 << 12) | (1 << 11)) /* instruction cache, branch prediction */
  253. orr r0, #((1 << 2) | (1 << 0)) /* data cache, mmu enable */
  254. mcr p15, 0, r0, c1, c0, 0
  255. dsb
  256. isb
  257. mov pc, lr
  258. .global rt_hw_set_process_id
  259. rt_hw_set_process_id:
  260. LSL r0, r0, #8
  261. MCR p15, 0, r0, c13, c0, 1
  262. mov pc, lr
  263. #endif
  264. .global rt_hw_mmu_switch
  265. rt_hw_mmu_switch:
  266. orr r0, #0x18
  267. mcr p15, 0, r0, c2, c0, 0 // ttbr0
  268. //invalid tlb
  269. mov r0, #0
  270. mcr p15, 0, r0, c8, c7, 0
  271. mcr p15, 0, r0, c7, c5, 0 //iciallu
  272. mcr p15, 0, r0, c7, c5, 6 //bpiall
  273. dsb
  274. isb
  275. mov pc, lr
  276. .global rt_hw_mmu_tbl_get
  277. rt_hw_mmu_tbl_get:
  278. mrc p15, 0, r0, c2, c0, 0 /* ttbr0 */
  279. bic r0, #0x18
  280. mov pc, lr
  281. _halt:
  282. wfe
  283. b _halt
  284. #ifdef RT_USING_SMP
  285. .global rt_secondary_cpu_entry
  286. rt_secondary_cpu_entry:
  287. #ifdef RT_USING_SMART
  288. ldr r0, =_reset
  289. adr r5, _reset
  290. sub r5, r5, r0
  291. ldr lr, =after_enable_mmu_n
  292. ldr r0, =init_mtbl
  293. add r0, r5
  294. b enable_mmu
  295. after_enable_mmu_n:
  296. ldr r0, =MMUTable
  297. add r0, r5
  298. bl rt_hw_mmu_switch
  299. #endif
  300. #ifdef RT_USING_FPU
  301. mov r4, #0xfffffff
  302. mcr p15, 0, r4, c1, c0, 2
  303. #endif
  304. mrc p15, 0, r1, c1, c0, 1
  305. mov r0, #(1<<6)
  306. orr r1, r0
  307. mcr p15, 0, r1, c1, c0, 1 /* enable smp */
  308. mrc p15, 0, r0, c1, c0, 0
  309. bic r0, #(1<<13)
  310. mcr p15, 0, r0, c1, c0, 0
  311. bl stack_setup
  312. /* initialize the mmu table and enable mmu */
  313. #ifndef RT_USING_SMART
  314. bl rt_hw_mmu_init
  315. #endif
  316. b rt_hw_secondary_cpu_bsp_start
  317. #endif
  318. /* exception handlers: undef, swi, padt, dabt, resv, irq, fiq */
  319. .section .text.isr, "ax"
  320. .align 5
  321. .globl vector_fiq
  322. vector_fiq:
  323. stmfd sp!,{r0-r7,lr}
  324. bl rt_hw_trap_fiq
  325. ldmfd sp!,{r0-r7,lr}
  326. subs pc, lr, #4
  327. .globl rt_interrupt_enter
  328. .globl rt_interrupt_leave
  329. .globl rt_thread_switch_interrupt_flag
  330. .globl rt_interrupt_from_thread
  331. .globl rt_interrupt_to_thread
  332. .globl rt_current_thread
  333. .globl vmm_thread
  334. .globl vmm_virq_check
  335. .align 5
  336. .globl vector_irq
  337. vector_irq:
  338. #ifdef RT_USING_SMP
  339. stmfd sp!, {r0, r1}
  340. cps #Mode_SVC
  341. mov r0, sp /* svc_sp */
  342. mov r1, lr /* svc_lr */
  343. cps #Mode_IRQ
  344. sub lr, #4
  345. stmfd r0!, {r1, lr} /* svc_lr, svc_pc */
  346. stmfd r0!, {r2 - r12}
  347. ldmfd sp!, {r1, r2} /* original r0, r1 */
  348. stmfd r0!, {r1 - r2}
  349. mrs r1, spsr /* original mode */
  350. stmfd r0!, {r1}
  351. #ifdef RT_USING_SMART
  352. stmfd r0, {r13, r14}^ /* usr_sp, usr_lr */
  353. sub r0, #8
  354. #endif
  355. #ifdef RT_USING_FPU
  356. /* fpu context */
  357. vmrs r6, fpexc
  358. tst r6, #(1<<30)
  359. beq 1f
  360. vstmdb r0!, {d0-d15}
  361. vstmdb r0!, {d16-d31}
  362. vmrs r5, fpscr
  363. stmfd r0!, {r5}
  364. 1:
  365. stmfd r0!, {r6}
  366. #endif
  367. /* now irq stack is clean */
  368. /* r0 is task svc_sp */
  369. /* backup r0 -> r8 */
  370. mov r8, r0
  371. cps #Mode_SVC
  372. mov sp, r8
  373. bl rt_interrupt_enter
  374. bl rt_hw_trap_irq
  375. bl rt_interrupt_leave
  376. mov r0, r8
  377. bl rt_scheduler_do_irq_switch
  378. b rt_hw_context_switch_exit
  379. #else
  380. stmfd sp!, {r0-r12,lr}
  381. bl rt_interrupt_enter
  382. bl rt_hw_trap_irq
  383. bl rt_interrupt_leave
  384. /* if rt_thread_switch_interrupt_flag set, jump to
  385. * rt_hw_context_switch_interrupt_do and don't return */
  386. ldr r0, =rt_thread_switch_interrupt_flag
  387. ldr r1, [r0]
  388. cmp r1, #1
  389. beq rt_hw_context_switch_interrupt_do
  390. #ifdef RT_USING_SMART
  391. ldmfd sp!, {r0-r12,lr}
  392. cps #Mode_SVC
  393. push {r0-r12}
  394. mov r7, lr
  395. cps #Mode_IRQ
  396. mrs r4, spsr
  397. sub r5, lr, #4
  398. cps #Mode_SVC
  399. and r6, r4, #0x1f
  400. cmp r6, #0x10
  401. bne 1f
  402. msr spsr_csxf, r4
  403. mov lr, r5
  404. pop {r0-r12}
  405. b arch_ret_to_user
  406. 1:
  407. mov lr, r7
  408. cps #Mode_IRQ
  409. msr spsr_csxf, r4
  410. mov lr, r5
  411. cps #Mode_SVC
  412. pop {r0-r12}
  413. cps #Mode_IRQ
  414. movs pc, lr
  415. #else
  416. ldmfd sp!, {r0-r12,lr}
  417. subs pc, lr, #4
  418. #endif
  419. rt_hw_context_switch_interrupt_do:
  420. mov r1, #0 /* clear flag */
  421. str r1, [r0]
  422. mov r1, sp /* r1 point to {r0-r3} in stack */
  423. add sp, sp, #4*4
  424. ldmfd sp!, {r4-r12,lr} /* reload saved registers */
  425. mrs r0, spsr /* get cpsr of interrupt thread */
  426. sub r2, lr, #4 /* save old task's pc to r2 */
  427. /* Switch to SVC mode with no interrupt. If the usr mode guest is
  428. * interrupted, this will just switch to the stack of kernel space.
  429. * save the registers in kernel space won't trigger data abort. */
  430. msr cpsr_c, #I_Bit|F_Bit|Mode_SVC
  431. stmfd sp!, {r2} /* push old task's pc */
  432. stmfd sp!, {r4-r12,lr} /* push old task's lr,r12-r4 */
  433. ldmfd r1, {r1-r4} /* restore r0-r3 of the interrupt thread */
  434. stmfd sp!, {r1-r4} /* push old task's r0-r3 */
  435. stmfd sp!, {r0} /* push old task's cpsr */
  436. #ifdef RT_USING_SMART
  437. stmfd sp, {r13, r14}^ /*push usr_sp, usr_lr */
  438. sub sp, #8
  439. #endif
  440. #ifdef RT_USING_FPU
  441. /* fpu context */
  442. vmrs r6, fpexc
  443. tst r6, #(1<<30)
  444. beq 1f
  445. vstmdb sp!, {d0-d15}
  446. vstmdb sp!, {d16-d31}
  447. vmrs r5, fpscr
  448. stmfd sp!, {r5}
  449. 1:
  450. stmfd sp!, {r6}
  451. #endif
  452. ldr r4, =rt_interrupt_from_thread
  453. ldr r5, [r4]
  454. str sp, [r5] /* store sp in preempted tasks's TCB */
  455. ldr r6, =rt_interrupt_to_thread
  456. ldr r6, [r6]
  457. ldr sp, [r6] /* get new task's stack pointer */
  458. #ifdef RT_USING_SMART
  459. bl rt_thread_self
  460. mov r4, r0
  461. bl lwp_aspace_switch
  462. mov r0, r4
  463. bl lwp_user_setting_restore
  464. #endif
  465. #ifdef RT_USING_FPU
  466. /* fpu context */
  467. ldmfd sp!, {r6}
  468. vmsr fpexc, r6
  469. tst r6, #(1<<30)
  470. beq 1f
  471. ldmfd sp!, {r5}
  472. vmsr fpscr, r5
  473. vldmia sp!, {d16-d31}
  474. vldmia sp!, {d0-d15}
  475. 1:
  476. #endif
  477. #ifdef RT_USING_SMART
  478. ldmfd sp, {r13, r14}^ /*pop usr_sp, usr_lr */
  479. add sp, #8
  480. #endif
  481. ldmfd sp!, {r4} /* pop new task's cpsr to spsr */
  482. msr spsr_cxsf, r4
  483. #ifdef RT_USING_SMART
  484. and r4, #0x1f
  485. cmp r4, #0x10
  486. bne 1f
  487. ldmfd sp!, {r0-r12,lr}
  488. ldmfd sp!, {lr}
  489. b arch_ret_to_user
  490. 1:
  491. #endif
  492. /* pop new task's r0-r12,lr & pc, copy spsr to cpsr */
  493. ldmfd sp!, {r0-r12,lr,pc}^
  494. #endif
  495. .macro push_svc_reg
  496. sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
  497. stmia sp, {r0 - r12} /* Calling r0-r12 */
  498. mov r0, sp
  499. add sp, sp, #17 * 4
  500. mrs r6, spsr /* Save CPSR */
  501. str lr, [r0, #15*4] /* Push PC */
  502. str r6, [r0, #16*4] /* Push CPSR */
  503. and r1, r6, #0x1f
  504. cmp r1, #0x10
  505. cps #Mode_SYS
  506. streq sp, [r0, #13*4] /* Save calling SP */
  507. streq lr, [r0, #14*4] /* Save calling PC */
  508. cps #Mode_SVC
  509. strne sp, [r0, #13*4] /* Save calling SP */
  510. strne lr, [r0, #14*4] /* Save calling PC */
  511. .endm
  512. .align 5
  513. .weak vector_swi
  514. vector_swi:
  515. push_svc_reg
  516. bl rt_hw_trap_swi
  517. b .
  518. .align 5
  519. .globl vector_undef
  520. vector_undef:
  521. push_svc_reg
  522. bl rt_hw_trap_undef
  523. #ifdef RT_USING_FPU
  524. cps #Mode_UND
  525. sub sp, sp, #17 * 4
  526. ldr lr, [sp, #15*4]
  527. ldmia sp, {r0 - r12}
  528. add sp, sp, #17 * 4
  529. movs pc, lr
  530. #endif
  531. b .
  532. .align 5
  533. .globl vector_pabt
  534. vector_pabt:
  535. push_svc_reg
  536. #ifdef RT_USING_SMART
  537. /* cp Mode_ABT stack to SVC */
  538. sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
  539. mov lr, r0
  540. ldmia lr, {r0 - r12}
  541. stmia sp, {r0 - r12}
  542. add r1, lr, #13 * 4
  543. add r2, sp, #13 * 4
  544. ldmia r1, {r4 - r7}
  545. stmia r2, {r4 - r7}
  546. mov r0, sp
  547. bl rt_hw_trap_pabt
  548. /* return to user */
  549. ldr lr, [sp, #16*4] /* orign spsr */
  550. msr spsr_cxsf, lr
  551. ldr lr, [sp, #15*4] /* orign pc */
  552. ldmia sp, {r0 - r12}
  553. add sp, #17 * 4
  554. b arch_ret_to_user
  555. #else
  556. bl rt_hw_trap_pabt
  557. b .
  558. #endif
  559. .align 5
  560. .globl vector_dabt
  561. vector_dabt:
  562. push_svc_reg
  563. #ifdef RT_USING_SMART
  564. /* cp Mode_ABT stack to SVC */
  565. sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
  566. mov lr, r0
  567. ldmia lr, {r0 - r12}
  568. stmia sp, {r0 - r12}
  569. add r1, lr, #13 * 4
  570. add r2, sp, #13 * 4
  571. ldmia r1, {r4 - r7}
  572. stmia r2, {r4 - r7}
  573. mov r0, sp
  574. bl rt_hw_trap_dabt
  575. /* return to user */
  576. ldr lr, [sp, #16*4] /* orign spsr */
  577. msr spsr_cxsf, lr
  578. ldr lr, [sp, #15*4] /* orign pc */
  579. ldmia sp, {r0 - r12}
  580. add sp, #17 * 4
  581. b arch_ret_to_user
  582. #else
  583. bl rt_hw_trap_dabt
  584. b .
  585. #endif
  586. .align 5
  587. .globl vector_resv
  588. vector_resv:
  589. push_svc_reg
  590. bl rt_hw_trap_resv
  591. b .
  592. .global rt_hw_clz
  593. rt_hw_clz:
  594. clz r0, r0
  595. bx lr
  596. #ifndef RT_CPUS_NR
  597. #define RT_CPUS_NR 1
  598. #endif
  599. #include "asm-generic.h"
  600. START_POINT(_thread_start)
  601. mov r10, lr
  602. blx r1
  603. blx r10
  604. b . /* never here */
  605. START_POINT_END(_thread_start)
  606. .bss
  607. .align 3 /* align to 2~3=8 */
  608. svc_stack_n:
  609. .space (RT_CPUS_NR << 12)
  610. svc_stack_n_limit:
  611. irq_stack_n:
  612. .space (RT_CPUS_NR << 12)
  613. und_stack_n:
  614. .space (RT_CPUS_NR << 12)
  615. abt_stack_n:
  616. .space (RT_CPUS_NR << 12)