entry_point.S 8.5 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280
  1. /*
  2. * Copyright (c) 2006-2020, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Date Author Notes
  7. * 2020-01-15 bigmagic the first version
  8. * 2020-08-10 SummerGift support clang compiler
  9. */
  10. #include "rtconfig.h"
  11. .section ".text.entrypoint","ax"
  12. .global __start
  13. __start:
  14. bl rt_hw_cpu_id_set
  15. /* read cpu id, stop slave cores */
  16. mrs x0, tpidr_el1
  17. cbz x0, .L__cpu_0 /* .L prefix is the local label in ELF */
  18. /* cpu id > 0, stop */
  19. /* cpu id == 0 will also goto here after returned from entry() if possible */
  20. .L__current_cpu_idle:
  21. wfe
  22. b .L__current_cpu_idle
  23. .L__cpu_0:
  24. /* set stack before our code, Define stack pointer for current exception level */
  25. adr x1, __start
  26. /* set up EL1 */
  27. mrs x0, CurrentEL /* CurrentEL Register. bit 2, 3. Others reserved */
  28. and x0, x0, #12 /* clear reserved bits */
  29. /* running at EL3? */
  30. cmp x0, #12 /* 1100b. So, EL3 */
  31. bne .L__not_in_el3 /* 11? !EL3 -> 5: */
  32. /* should never be executed, just for completeness. (EL3) */
  33. mov x2, #0x5b1
  34. msr scr_el3, x2 /* SCR_ELn Secure Configuration Register */
  35. mov x2, #0x3c9
  36. msr spsr_el3, x2 /* SPSR_ELn. Saved Program Status Register. 1111001001 */
  37. adr x2, .L__not_in_el3
  38. msr elr_el3, x2
  39. eret /* Exception Return: from EL3, continue from .L__not_in_el3 */
  40. .L__not_in_el3: /* running at EL2 or EL1 */
  41. cmp x0, #4 /* 0x04 0100 EL1 */
  42. beq .L__in_el1 /* EL1 -> 5: */
  43. mrs x0, hcr_el2
  44. bic x0, x0, #0xff
  45. msr hcr_el2, x0
  46. msr sp_el1, x1 /* in EL2, set sp of EL1 to _start */
  47. /* enable CNTP for EL1 */
  48. mrs x0, cnthctl_el2 /* Counter-timer Hypervisor Control register */
  49. orr x0, x0, #3
  50. msr cnthctl_el2, x0
  51. msr cntvoff_el2, xzr
  52. /* enable AArch64 in EL1 */
  53. mov x0, #(1 << 31) /* AArch64 */
  54. orr x0, x0, #(1 << 1) /* SWIO hardwired on Pi3 */
  55. msr hcr_el2, x0
  56. mrs x0, hcr_el2
  57. /* change execution level to EL1 */
  58. mov x2, #0x3c4
  59. msr spsr_el2, x2 /* 1111000100 */
  60. adr x2, .L__in_el1
  61. msr elr_el2, x2
  62. eret /* exception return. from EL2. continue from .L__in_el1 */
  63. .L__in_el1:
  64. ldr x9, =PV_OFFSET
  65. mov sp, x1 /* in EL1. Set sp to _start */
  66. /* Set CPACR_EL1 (Architecture Feature Access Control Register) to avoid trap from SIMD or float point instruction */
  67. mov x1, #0x00300000 /* Don't trap any SIMD/FP instructions in both EL0 and EL1 */
  68. msr cpacr_el1, x1
  69. /* clear bss */
  70. ldr x1, =__bss_start /* get bss start address */
  71. ldr x2, =__bss_end
  72. sub x2, x2, x1 /* get bss size */
  73. add x1, x1, x9
  74. and x3, x2, #7 /* x3 is < 7 */
  75. ldr x4, =~0x7
  76. and x2, x2, x4 /* mask ~7 */
  77. .L__clean_bss_loop:
  78. cbz x2, .L__clean_bss_loop_1
  79. str xzr, [x1], #8
  80. sub x2, x2, #8
  81. b .L__clean_bss_loop
  82. .L__clean_bss_loop_1:
  83. cbz x3, .L__jump_to_entry
  84. strb wzr, [x1], #1
  85. sub x3, x3, #1
  86. b .L__clean_bss_loop_1
  87. .L__jump_to_entry: /* jump to C code, should not return */
  88. bl mmu_tcr_init
  89. adr x1, __start
  90. ldr x0, =~0x1fffff
  91. and x0, x1, x0
  92. add x1, x0, #0x1000
  93. msr ttbr0_el1, x0
  94. msr ttbr1_el1, x1
  95. dsb sy
  96. ldr x2, =0x40000000 /* map 1G memory for kernel space */
  97. ldr x3, =PV_OFFSET
  98. bl rt_hw_mmu_setup_early
  99. ldr x30, =after_mmu_enable /* set LR to after_mmu_enable function, it's a v_addr */
  100. mrs x1, sctlr_el1
  101. bic x1, x1, #(3 << 3) /* dis SA, SA0 */
  102. bic x1, x1, #(1 << 1) /* dis A */
  103. orr x1, x1, #(1 << 12) /* I */
  104. orr x1, x1, #(1 << 2) /* C */
  105. orr x1, x1, #(1 << 0) /* M */
  106. msr sctlr_el1, x1 /* enable MMU */
  107. dsb sy
  108. isb sy
  109. ic ialluis /* Invalidate all instruction caches in Inner Shareable domain to Point of Unification */
  110. dsb sy
  111. isb sy
  112. tlbi vmalle1 /* Invalidate all stage 1 translations used at EL1 with the current VMID */
  113. dsb sy
  114. isb sy
  115. ret
  116. after_mmu_enable:
  117. mrs x0, tcr_el1 /* disable ttbr0, only using kernel space */
  118. orr x0, x0, #(1 << 7)
  119. msr tcr_el1, x0
  120. msr ttbr0_el1, xzr
  121. dsb sy
  122. mov x0, #1
  123. msr spsel, x0
  124. adr x1, __start
  125. mov sp, x1 /* sp_el1 set to _start */
  126. b rtthread_startup
  127. #ifdef RT_USING_SMP
  128. /**
  129. * secondary cpu
  130. */
  131. .globl _secondary_cpu_entry
  132. _secondary_cpu_entry:
  133. bl rt_hw_cpu_id_set
  134. adr x1, __start
  135. /* set up EL1 */
  136. mrs x0, CurrentEL /* CurrentEL Register. bit 2, 3. Others reserved */
  137. and x0, x0, #12 /* clear reserved bits */
  138. /* running at EL3? */
  139. cmp x0, #12 /* 1100b. So, EL3 */
  140. bne .L__not_in_el3_cpux /* 11? !EL3 -> 5: */
  141. /* should never be executed, just for completeness. (EL3) */
  142. mov x2, #0x5b1
  143. msr scr_el3, x2 /* SCR_ELn Secure Configuration Register */
  144. mov x2, #0x3c9
  145. msr spsr_el3, x2 /* SPSR_ELn. Saved Program Status Register. 1111001001 */
  146. adr x2, .L__not_in_el3_cpux
  147. msr elr_el3, x2
  148. eret /* Exception Return: from EL3, continue from .L__not_in_el3 */
  149. .L__not_in_el3_cpux: /* running at EL2 or EL1 */
  150. cmp x0, #4 /* 0x04 0100 EL1 */
  151. beq .L__in_el1_cpux /* EL1 -> 5: */
  152. mrs x0, hcr_el2
  153. bic x0, x0, #0xff
  154. msr hcr_el2, x0
  155. msr sp_el1, x1 /* in EL2, set sp of EL1 to _start */
  156. /* enable CNTP for EL1 */
  157. mrs x0, cnthctl_el2 /* Counter-timer Hypervisor Control register */
  158. orr x0, x0, #3
  159. msr cnthctl_el2, x0
  160. msr cntvoff_el2, xzr
  161. /* enable AArch64 in EL1 */
  162. mov x0, #(1 << 31) /* AArch64 */
  163. orr x0, x0, #(1 << 1) /* SWIO hardwired on Pi3 */
  164. msr hcr_el2, x0
  165. mrs x0, hcr_el2
  166. /* change execution level to EL1 */
  167. mov x2, #0x3c4
  168. msr spsr_el2, x2 /* 1111000100 */
  169. adr x2, .L__in_el1_cpux
  170. msr elr_el2, x2
  171. eret /* exception return. from EL2. continue from .L__in_el1 */
  172. .L__in_el1_cpux:
  173. adr x19, .L__in_el1_cpux
  174. ldr x8, =.L__in_el1_cpux
  175. sub x19, x19, x8 /* get PV_OFFSET */
  176. mrs x0, tpidr_el1
  177. /* each cpu init stack is 8k */
  178. sub x1, x1, x0, lsl #13
  179. mov sp, x1 /* in EL1. Set sp to _start */
  180. /* Set CPACR_EL1 (Architecture Feature Access Control Register) to avoid trap from SIMD or float point instruction */
  181. mov x1, #0x00300000 /* Don't trap any SIMD/FP instructions in both EL0 and EL1 */
  182. msr cpacr_el1, x1
  183. .L__jump_to_entry_cpux: /* jump to C code, should not return */
  184. /* init mmu early */
  185. bl mmu_tcr_init
  186. adr x1, __start
  187. ldr x0, =~0x1fffff
  188. and x0, x1, x0
  189. add x1, x0, #0x1000
  190. msr ttbr0_el1, x0
  191. msr ttbr1_el1, x1
  192. dsb sy
  193. ldr x30, =after_mmu_enable_cpux /* set LR to after_mmu_enable function, it's a v_addr */
  194. mrs x1, sctlr_el1
  195. bic x1, x1, #(3 << 3) /* dis SA, SA0 */
  196. bic x1, x1, #(1 << 1) /* dis A */
  197. orr x1, x1, #(1 << 12) /* I */
  198. orr x1, x1, #(1 << 2) /* C */
  199. orr x1, x1, #(1 << 0) /* M */
  200. msr sctlr_el1, x1 /* enable MMU */
  201. dsb sy
  202. isb sy
  203. ic ialluis /* Invalidate all instruction caches in Inner Shareable domain to Point of Unification */
  204. dsb sy
  205. isb sy
  206. tlbi vmalle1 /* Invalidate all stage 1 translations used at EL1 with the current VMID */
  207. dsb sy
  208. isb sy
  209. ret
  210. after_mmu_enable_cpux:
  211. mrs x0, tcr_el1 /* disable ttbr0, only using kernel space */
  212. orr x0, x0, #(1 << 7)
  213. msr tcr_el1, x0
  214. msr ttbr0_el1, xzr
  215. dsb sy
  216. mov x0, #1
  217. msr spsel, x0
  218. mrs x0, tpidr_el1
  219. /* each cpu init stack is 8k */
  220. adr x1, __start
  221. sub x1, x1, x0, lsl #13
  222. mov sp, x1 /* in EL1. Set sp to _start */
  223. b rt_hw_secondary_cpu_bsp_start
  224. #endif