entry_point.S 4.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163
  1. /*
  2. * Copyright (c) 2006-2020, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Date Author Notes
  7. * 2020-01-15 bigmagic the first version
  8. * 2020-08-10 SummerGift support clang compiler
  9. */
  10. #include "rtconfig.h"
  11. .section ".text.entrypoint","ax"
  12. .set EL1_stack, __el1_stack
  13. .global __start
  14. /* This symbol is set to 0x80000 in ld script. That is the address that raspi3's firmware
  15. loads 'kernel8.img' file in. */
  16. __start:
  17. /* read cpu id, stop slave cores */
  18. mrs x1, mpidr_el1 /* MPIDR_EL1: Multi-Processor Affinity Register */
  19. and x1, x1, #3
  20. cbz x1, .L__cpu_0 /* .L prefix is the local label in ELF */
  21. /* cpu id > 0, stop */
  22. /* cpu id == 0 will also goto here after returned from entry() if possible */
  23. .L__current_cpu_idle:
  24. wfe
  25. b .L__current_cpu_idle
  26. .L__cpu_0: /* cpu id == 0 */
  27. /* set stack before our code */
  28. /* Define stack pointer for current exception level */
  29. /* ldr x2, =EL1_stack */
  30. /* mov sp, x2 */
  31. adr x1, __start
  32. /* set up EL1 */
  33. mrs x0, CurrentEL /* CurrentEL Register. bit 2, 3. Others reserved */
  34. and x0, x0, #12 /* clear reserved bits */
  35. /* running at EL3? */
  36. cmp x0, #12 /* 1100b. So, EL3 */
  37. bne .L__not_in_el3 /* 11? !EL3 -> 5: */
  38. /* should never be executed, just for completeness. (EL3) */
  39. mov x2, #0x5b1
  40. msr scr_el3, x2 /* SCR_ELn Secure Configuration Register */
  41. mov x2, #0x3c9
  42. msr spsr_el3, x2 /* SPSR_ELn. Saved Program Status Register. 1111001001 */
  43. adr x2, .L__not_in_el3
  44. msr elr_el3, x2
  45. eret /* Exception Return: from EL3, continue from .L__not_in_el3 */
  46. /* running at EL2 or EL1 */
  47. .L__not_in_el3:
  48. cmp x0, #4 /* 0x04 0100 EL1 */
  49. beq .L__in_el1 /* EL1 -> 5: */
  50. mrs x0, hcr_el2
  51. bic x0, x0, #0xff
  52. msr hcr_el2, x0
  53. /* in EL2 */
  54. msr sp_el1, x1 /* Set sp of EL1 to _start */
  55. /* enable CNTP for EL1 */
  56. mrs x0, cnthctl_el2 /* Counter-timer Hypervisor Control register */
  57. orr x0, x0, #3
  58. msr cnthctl_el2, x0
  59. msr cntvoff_el2, xzr
  60. /* enable AArch64 in EL1 */
  61. mov x0, #(1 << 31) /* AArch64 */
  62. orr x0, x0, #(1 << 1) /* SWIO hardwired on Pi3 */
  63. msr hcr_el2, x0
  64. mrs x0, hcr_el2
  65. /* change execution level to EL1 */
  66. mov x2, #0x3c4
  67. msr spsr_el2, x2 /* 1111000100 */
  68. adr x2, .L__in_el1
  69. msr elr_el2, x2
  70. eret /* exception return. from EL2. continue from .L__in_el1 */
  71. .L__in_el1:
  72. ldr x9, =PV_OFFSET
  73. mov sp, x1 /* in EL1. Set sp to _start */
  74. /* Set CPACR_EL1 (Architecture Feature Access Control Register) to avoid trap from SIMD or float point instruction */
  75. mov x1, #0x00300000 /* Don't trap any SIMD/FP instructions in both EL0 and EL1 */
  76. msr cpacr_el1, x1
  77. /* clear bss */
  78. ldr x1, =__bss_start
  79. add x1, x1, x9
  80. ldr w2, =__bss_size
  81. .L__clean_bss_loop:
  82. cbz w2, .L__jump_to_entry
  83. str xzr, [x1], #8
  84. sub w2, w2, #1
  85. cbnz w2, .L__clean_bss_loop
  86. /* jump to C code, should not return */
  87. .L__jump_to_entry:
  88. bl get_free_page
  89. mov x21, x0
  90. bl get_free_page
  91. mov x20, x0
  92. mov x1, x21
  93. bl mmu_tcr_init
  94. mov x0, x20
  95. mov x1, x21
  96. msr ttbr0_el1, x0
  97. msr ttbr1_el1, x1
  98. dsb sy
  99. ldr x2, =0x40000000 /* map 1G memory for kernel space */
  100. ldr x3, =PV_OFFSET
  101. bl rt_hw_mmu_setup_early
  102. ldr x30, =after_mmu_enable
  103. mrs x1, sctlr_el1
  104. bic x1, x1, #(3 << 3) /* dis SA, SA0 */
  105. bic x1, x1, #(1 << 1) /* dis A */
  106. orr x1, x1, #(1 << 12) /* I */
  107. orr x1, x1, #(1 << 2) /* C */
  108. orr x1, x1, #(1 << 0) /* M */
  109. msr sctlr_el1, x1 /* enable MMU */
  110. dsb sy
  111. isb sy
  112. ic ialluis /* Invalidate all instruction caches in Inner Shareable domain to Point of Unification */
  113. dsb sy
  114. isb sy
  115. tlbi vmalle1 /* Invalidate all stage 1 translations used at EL1 with the current VMID */
  116. dsb sy
  117. isb sy
  118. ret
  119. after_mmu_enable:
  120. mrs x0, tcr_el1 /* disable ttbr0, only using kernel space */
  121. orr x0, x0, #(1 << 7)
  122. msr tcr_el1, x0
  123. msr ttbr0_el1, xzr
  124. dsb sy
  125. mov x0, #1
  126. msr spsel, x0
  127. adr x1, __start
  128. mov sp, x1 /* sp_el1 set to _start */
  129. b rtthread_startup