123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695 |
- /*
- * Copyright (c) 2006-2018, RT-Thread Development Team
- *
- * SPDX-License-Identifier: Apache-2.0
- *
- * Change Logs:
- * Date Author Notes
- * 2013-07-05 Bernard the first version
- * 2018-11-22 Jesven in the interrupt context, use rt_scheduler_do_irq_switch checks
- * and switches to a new thread
- * 2024-01-16 huanghe restructure this code section following the aarch64 architectural style
- */
- #include "rtconfig.h"
- #define ARM_CPU_STACK_SIZE_OFFSET 12
- #define ARM_CPU_STACK_SIZE (1<<ARM_CPU_STACK_SIZE_OFFSET)
- .equ Mode_USR, 0x10
- .equ Mode_FIQ, 0x11
- .equ Mode_IRQ, 0x12
- .equ Mode_SVC, 0x13
- .equ Mode_ABT, 0x17
- .equ Mode_UND, 0x1B
- .equ Mode_SYS, 0x1F
- .equ I_Bit, 0x80 /* when I bit is set, IRQ is disabled */
- .equ F_Bit, 0x40 /* when F bit is set, FIQ is disabled */
- /*Load the physical address of a symbol into a register.
- Through pv_off calculates the offset of the physical address */
- .macro get_phy, reg, symbol, _pvoff
- ldr \reg, =\symbol
- add \reg, \_pvoff
- .endm
- /*Calculate the offset between the physical address and the virtual address of the "_reset".*/
- .macro get_pvoff, tmp, out
- ldr \tmp, =_reset
- adr \out, _reset
- sub \out, \out, \tmp
- .endm
- pv_off .req r11 /* Used to store the offset between physical address and the virtual address */
- cpu_id .req r10 /* Used to store the cpu id */
- /* reset entry */
- .globl _reset
- _reset:
- /* Calculate the offset between the physical address and the virtual address */
- get_pvoff r0, pv_off
- /* exit hyp mode */
- bl init_cpu_mode
- /* clear bss section */
- bl init_kernel_bss
- /* Initializes the assembly environment stack */
- bl init_cpu_stack_early
- /* init mmu */
- b init_mmu_page_table_early
- init_cpu_stack_early:
- cps #Mode_SVC
-
- get_phy r0, svc_stack_top, pv_off
- mov sp, r0
- #ifdef RT_USING_FPU
- mov r4, #0xfffffff
- mcr p15, 0, r4, c1, c0, 2
- #endif
- mov pc, lr
- init_kernel_bss:
- /* enable I cache + branch prediction */
- mrc p15, 0, r0, c1, c0, 0
- orr r0, r0, #(1<<12)
- orr r0, r0, #(1<<11)
- mcr p15, 0, r0, c1, c0, 0
- mov r0,#0 /* get a zero */
- get_phy r1, __bss_start, pv_off
- get_phy r2, __bss_end, pv_off
- bss_loop:
- cmp r1,r2 /* check if data to clear */
- strlo r0,[r1],#4 /* clear 4 bytes */
- blo bss_loop /* loop until done */
- mov pc, lr
- init_cpu_mode:
- #ifdef ARCH_ARMV8
- /* Check for HYP mode */
- mrs r0, cpsr_all
- and r0, r0, #0x1F
- mov r8, #0x1A
- cmp r0, r8
- beq overHyped
- b continue_exit
- overHyped: /* Get out of HYP mode */
- mov r9, lr
- /* HYP mode has a dedicated register, called ELR_hyp,
- to store the exception return address.
- The lr register needs to be temporarily saved,
- otherwise "mov pc lr" cannot be used after switching modes. */
- adr r1, continue_exit
- msr ELR_hyp, r1
- mrs r1, cpsr_all
- and r1, r1, #0xFFFFFFE0 /* CPSR_MODE_MASK */
- orr r1, r1, #0x13 /* CPSR_MODE_SUPERVISOR */
- msr SPSR_hyp, r1
- eret
- continue_exit:
- mov lr ,r9
-
- #endif
- #ifdef SOC_BCM283x
- /* Suspend the other cpu cores */
- mrc p15, 0, r0, c0, c0, 5
- ands r0, #3
- bne _halt
- /* Disable IRQ & FIQ */
- cpsid if
- /* Check for HYP mode */
- mrs r0, cpsr_all
- and r0, r0, #0x1F
- mov r8, #0x1A
- cmp r0, r8
- beq overHyped
- b continue_exit
- overHyped: /* Get out of HYP mode */
- mov r9, lr
- /* HYP mode has a dedicated register, called ELR_hyp,
- to store the exception return address.
- The lr register needs to be temporarily saved,
- otherwise "mov pc lr" cannot be used after switching modes. */
- adr r1, continue_exit
- msr ELR_hyp, r1
- mrs r1, cpsr_all
- and r1, r1, #0xFFFFFFE0 /* CPSR_MODE_MASK */
- orr r1, r1, #0x13 /* CPSR_MODE_SUPERVISOR */
- msr SPSR_hyp, r1
- eret
- continue_exit:
- mov lr ,r9
- /* set the cpu to SVC32 mode and disable interrupt */
- mrs r0, cpsr
- bic r0, r0, #0x1f
- orr r0, r0, #0x13
- msr cpsr_c, r0
-
- #endif
- /* invalid tlb before enable mmu */
- mrc p15, 0, r0, c1, c0, 0
- bic r0, #1
- mcr p15, 0, r0, c1, c0, 0
- dsb
- isb
- mov r0, #0
- mcr p15, 0, r0, c8, c7, 0
- mcr p15, 0, r0, c7, c5, 0 /* iciallu */
- mcr p15, 0, r0, c7, c5, 6 /* bpiall */
- dsb
- isb
- mov pc, lr
- init_mmu_page_table_early:
- get_phy r0, init_mtbl, pv_off
- mov r1, pv_off
- bl rt_hw_mem_setup_early
- /* get cpu id */
- bl rt_hw_cpu_id_early
- mov cpu_id ,r0
- /* enable_mmu_page_table_early is changed to master_core_startup */
- ldr lr, =master_core_startup
-
- cmp cpu_id, #0
- beq enable_mmu_page_table_early
- #ifdef RT_USING_SMP
- #ifdef RT_SMP_AUTO_BOOT
- /* if cpu id > 0, stop or wait */
- ldr r0, =secondary_cpu_entry
- mov r1, #0
- str r1, [r0] /* clean secondary_cpu_entry */
- #endif
- #endif
- secondary_loop:
- @ cpu core 1 goes into sleep until core 0 wakeup it
- wfe
- #ifdef RT_SMP_AUTO_BOOT
- ldr r1, =secondary_cpu_entry
- ldr r0, [r1]
- cmp r0, #0
- blxne r0 /* if(secondary_cpu_entry) secondary_cpu_entry(); */
- #endif /* RT_SMP_AUTO_BOOT */
- b secondary_loop
-
- enable_mmu_page_table_early:
- /* init TTBR0 */
- get_phy r0, init_mtbl, pv_off
- mcr p15, #0, r0, c2, c0, #0
- dmb
-
- ldr r0,=#0x55555555
- mcr p15, #0, r0, c3, c0, #0
- /* disable ttbr1 */
- mov r0, #(1 << 5) /* PD1=1 */
- mcr p15, 0, r0, c2, c0, 2 /* ttbcr */
-
- /* init stack for cpu mod */
- cps #Mode_UND
- ldr r1,=und_stack_top
- sub sp, r1, cpu_id, asl #ARM_CPU_STACK_SIZE_OFFSET
- cps #Mode_IRQ
- ldr r1, =irq_stack_top
- sub sp, r1, cpu_id, asl #ARM_CPU_STACK_SIZE_OFFSET
- cps #Mode_FIQ
- ldr r1, =irq_stack_top
- sub sp, r1, cpu_id, asl #ARM_CPU_STACK_SIZE_OFFSET
- cps #Mode_ABT
- ldr r1, =abt_stack_top
- sub sp, r1, cpu_id, asl #ARM_CPU_STACK_SIZE_OFFSET
- cps #Mode_SVC
- ldr r1, =svc_stack_top
- sub sp, r1, cpu_id, asl #ARM_CPU_STACK_SIZE_OFFSET
- /* invalid tlb before enable mmu */
- mov r0, #0
- mcr p15, 0, r0, c8, c7, 0
- mcr p15, 0, r0, c7, c5, 0 /* iciallu */
- mcr p15, 0, r0, c7, c5, 6 /* bpiall */
- mrc p15, 0, r0, c1, c0, 0
- bic r0, r0, #0x7 /* clear bit1~3 */
- orr r0, #((1 << 12) | (1 << 11)) /* instruction cache, branch prediction */
- orr r0, #((1 << 2) | (1 << 0)) /* data cache, mmu enable */
- mcr p15, 0, r0, c1, c0, 0
- dsb
- isb
- mov pc, lr
- master_core_startup :
- mov r0 ,pv_off
- bl rt_kmem_pvoff_set
- ldr lr, =rtthread_startup
- mov pc, lr
- .global rt_hw_mmu_tbl_get
- rt_hw_mmu_tbl_get:
- mrc p15, 0, r0, c2, c0, 0 /* ttbr0 */
- bic r0, #0x18
- mov pc, lr
- .weak rt_hw_cpu_id_early
- rt_hw_cpu_id_early:
- mrc p15, 0, r0, c0, c0, 5
- and r0, r0, #0xf
- mov pc, lr
- #ifdef RT_USING_SMP
- .global rt_secondary_cpu_entry
- rt_secondary_cpu_entry:
- ldr r0, =_reset
- adr pv_off, _reset
- sub pv_off, pv_off, r0
- bl init_cpu_stack_early
- /* init mmu */
- bl rt_hw_cpu_id_early
- mov cpu_id ,r0
- ldr lr ,= rt_hw_secondary_cpu_bsp_start
- b enable_mmu_page_table_early
- #endif
- /* exception handlers: undef, swi, padt, dabt, resv, irq, fiq */
- .section .text.isr, "ax"
- .align 5
- .globl vector_fiq
- vector_fiq:
- stmfd sp!,{r0-r7,lr}
- bl rt_hw_trap_fiq
- ldmfd sp!,{r0-r7,lr}
- subs pc, lr, #4
- .globl rt_interrupt_enter
- .globl rt_interrupt_leave
- .globl rt_thread_switch_interrupt_flag
- .globl rt_interrupt_from_thread
- .globl rt_interrupt_to_thread
- .globl rt_current_thread
- .globl vmm_thread
- .globl vmm_virq_check
- .align 5
- .globl vector_irq
- vector_irq:
- #ifdef RT_USING_SMP
- stmfd sp!, {r0, r1}
- cps #Mode_SVC
- mov r0, sp /* svc_sp */
- mov r1, lr /* svc_lr */
- cps #Mode_IRQ
- sub lr, #4
- stmfd r0!, {r1, lr} /* svc_lr, svc_pc */
- stmfd r0!, {r2 - r12}
- ldmfd sp!, {r1, r2} /* original r0, r1 */
- stmfd r0!, {r1 - r2}
- mrs r1, spsr /* original mode */
- stmfd r0!, {r1}
- #ifdef RT_USING_SMART
- stmfd r0, {r13, r14}^ /* usr_sp, usr_lr */
- sub r0, #8
- #endif
- #ifdef RT_USING_FPU
- /* fpu context */
- vmrs r6, fpexc
- tst r6, #(1<<30)
- beq 1f
- vstmdb r0!, {d0-d15}
- vstmdb r0!, {d16-d31}
- vmrs r5, fpscr
- stmfd r0!, {r5}
- 1:
- stmfd r0!, {r6}
- #endif
- /* now irq stack is clean */
- /* r0 is task svc_sp */
- /* backup r0 -> r8 */
- mov r8, r0
- cps #Mode_SVC
- mov sp, r8
- bl rt_interrupt_enter
- bl rt_hw_trap_irq
- bl rt_interrupt_leave
- mov r0, r8
- bl rt_scheduler_do_irq_switch
- b rt_hw_context_switch_exit
- #else
- stmfd sp!, {r0-r12,lr}
- bl rt_interrupt_enter
- bl rt_hw_trap_irq
- bl rt_interrupt_leave
- /* if rt_thread_switch_interrupt_flag set, jump to
- * rt_hw_context_switch_interrupt_do and don't return */
- ldr r0, =rt_thread_switch_interrupt_flag
- ldr r1, [r0]
- cmp r1, #1
- beq rt_hw_context_switch_interrupt_do
- #ifdef RT_USING_SMART
- ldmfd sp!, {r0-r12,lr}
- cps #Mode_SVC
- push {r0-r12}
- mov r7, lr
- cps #Mode_IRQ
- mrs r4, spsr
- sub r5, lr, #4
- cps #Mode_SVC
- and r6, r4, #0x1f
- cmp r6, #0x10
- bne 1f
- msr spsr_csxf, r4
- mov lr, r5
- pop {r0-r12}
- b arch_ret_to_user
- 1:
- mov lr, r7
- cps #Mode_IRQ
- msr spsr_csxf, r4
- mov lr, r5
- cps #Mode_SVC
- pop {r0-r12}
- cps #Mode_IRQ
- movs pc, lr
- #else
- ldmfd sp!, {r0-r12,lr}
- subs pc, lr, #4
- #endif
- rt_hw_context_switch_interrupt_do:
- mov r1, #0 /* clear flag */
- str r1, [r0]
- mov r1, sp /* r1 point to {r0-r3} in stack */
- add sp, sp, #4*4
- ldmfd sp!, {r4-r12,lr} /* reload saved registers */
- mrs r0, spsr /* get cpsr of interrupt thread */
- sub r2, lr, #4 /* save old task's pc to r2 */
- /* Switch to SVC mode with no interrupt. If the usr mode guest is
- * interrupted, this will just switch to the stack of kernel space.
- * save the registers in kernel space won't trigger data abort. */
- msr cpsr_c, #I_Bit|F_Bit|Mode_SVC
- stmfd sp!, {r2} /* push old task's pc */
- stmfd sp!, {r4-r12,lr} /* push old task's lr,r12-r4 */
- ldmfd r1, {r1-r4} /* restore r0-r3 of the interrupt thread */
- stmfd sp!, {r1-r4} /* push old task's r0-r3 */
- stmfd sp!, {r0} /* push old task's cpsr */
- #ifdef RT_USING_SMART
- stmfd sp, {r13, r14}^ /*push usr_sp, usr_lr */
- sub sp, #8
- #endif
- #ifdef RT_USING_FPU
- /* fpu context */
- vmrs r6, fpexc
- tst r6, #(1<<30)
- beq 1f
- vstmdb sp!, {d0-d15}
- vstmdb sp!, {d16-d31}
- vmrs r5, fpscr
- stmfd sp!, {r5}
- 1:
- stmfd sp!, {r6}
- #endif
- ldr r4, =rt_interrupt_from_thread
- ldr r5, [r4]
- str sp, [r5] /* store sp in preempted tasks's TCB */
- ldr r6, =rt_interrupt_to_thread
- ldr r6, [r6]
- ldr sp, [r6] /* get new task's stack pointer */
- #ifdef RT_USING_SMART
- bl rt_thread_self
- mov r4, r0
- bl lwp_aspace_switch
- mov r0, r4
- bl lwp_user_setting_restore
- #endif
- #ifdef RT_USING_FPU
- /* fpu context */
- ldmfd sp!, {r6}
- vmsr fpexc, r6
- tst r6, #(1<<30)
- beq 1f
- ldmfd sp!, {r5}
- vmsr fpscr, r5
- vldmia sp!, {d16-d31}
- vldmia sp!, {d0-d15}
- 1:
- #endif
- #ifdef RT_USING_SMART
- ldmfd sp, {r13, r14}^ /*pop usr_sp, usr_lr */
- add sp, #8
- #endif
- ldmfd sp!, {r4} /* pop new task's cpsr to spsr */
- msr spsr_cxsf, r4
- #ifdef RT_USING_SMART
- and r4, #0x1f
- cmp r4, #0x10
- bne 1f
- ldmfd sp!, {r0-r12,lr}
- ldmfd sp!, {lr}
- b arch_ret_to_user
- 1:
- #endif
- /* pop new task's r0-r12,lr & pc, copy spsr to cpsr */
- ldmfd sp!, {r0-r12,lr,pc}^
- #endif
- .macro push_svc_reg
- sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
- stmia sp, {r0 - r12} /* Calling r0-r12 */
- mov r0, sp
- add sp, sp, #17 * 4
- mrs r6, spsr /* Save CPSR */
- str lr, [r0, #15*4] /* Push PC */
- str r6, [r0, #16*4] /* Push CPSR */
- and r1, r6, #0x1f
- cmp r1, #0x10
- cps #Mode_SYS
- streq sp, [r0, #13*4] /* Save calling SP */
- streq lr, [r0, #14*4] /* Save calling PC */
- cps #Mode_SVC
- strne sp, [r0, #13*4] /* Save calling SP */
- strne lr, [r0, #14*4] /* Save calling PC */
- .endm
- .align 5
- .weak vector_swi
- vector_swi:
- push_svc_reg
- bl rt_hw_trap_swi
- b .
- .align 5
- .globl vector_undef
- vector_undef:
- push_svc_reg
- bl rt_hw_trap_undef
- #ifdef RT_USING_FPU
- cps #Mode_UND
- sub sp, sp, #17 * 4
- ldr lr, [sp, #15*4]
- ldmia sp, {r0 - r12}
- add sp, sp, #17 * 4
- movs pc, lr
- #endif
- b .
- .align 5
- .globl vector_pabt
- vector_pabt:
- push_svc_reg
- #ifdef RT_USING_SMART
- /* cp Mode_ABT stack to SVC */
- sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
- mov lr, r0
- ldmia lr, {r0 - r12}
- stmia sp, {r0 - r12}
- add r1, lr, #13 * 4
- add r2, sp, #13 * 4
- ldmia r1, {r4 - r7}
- stmia r2, {r4 - r7}
- mov r0, sp
- bl rt_hw_trap_pabt
- /* return to user */
- ldr lr, [sp, #16*4] /* orign spsr */
- msr spsr_cxsf, lr
- ldr lr, [sp, #15*4] /* orign pc */
- ldmia sp, {r0 - r12}
- add sp, #17 * 4
- b arch_ret_to_user
- #else
- bl rt_hw_trap_pabt
- b .
- #endif
- .align 5
- .globl vector_dabt
- vector_dabt:
- push_svc_reg
- #ifdef RT_USING_SMART
- /* cp Mode_ABT stack to SVC */
- sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
- mov lr, r0
- ldmia lr, {r0 - r12}
- stmia sp, {r0 - r12}
- add r1, lr, #13 * 4
- add r2, sp, #13 * 4
- ldmia r1, {r4 - r7}
- stmia r2, {r4 - r7}
- mov r0, sp
- bl rt_hw_trap_dabt
- /* return to user */
- ldr lr, [sp, #16*4] /* orign spsr */
- msr spsr_cxsf, lr
- ldr lr, [sp, #15*4] /* orign pc */
- ldmia sp, {r0 - r12}
- add sp, #17 * 4
- b arch_ret_to_user
- #else
- bl rt_hw_trap_dabt
- b .
- #endif
- .align 5
- .globl vector_resv
- vector_resv:
- push_svc_reg
- bl rt_hw_trap_resv
- b .
- .global rt_hw_clz
- rt_hw_clz:
- clz r0, r0
- bx lr
- #include "asm-generic.h"
- START_POINT(_thread_start)
- mov r10, lr
- blx r1
- blx r10
- b . /* never here */
- START_POINT_END(_thread_start)
- .data
- .align 14
- init_mtbl:
- .space (4*4096) /* The L1 translation table therefore contains 4096 32-bit (word-sized) entries. */
- .global rt_hw_mmu_switch
- rt_hw_mmu_switch:
- orr r0, #0x18
- mcr p15, 0, r0, c2, c0, 0 // ttbr0
- //invalid tlb
- mov r0, #0
- mcr p15, 0, r0, c8, c7, 0
- mcr p15, 0, r0, c7, c5, 0 //iciallu
- mcr p15, 0, r0, c7, c5, 6 //bpiall
- dsb
- isb
- mov pc, lr
- .global rt_hw_set_process_id
- rt_hw_set_process_id:
- LSL r0, r0, #8
- MCR p15, 0, r0, c13, c0, 1
- mov pc, lr
- .bss
- .align 3 /* align to 2~3=8 */
- .cpus_stack:
- svc_stack_n:
- #if defined(RT_USING_SMP) && (RT_CPUS_NR > 1)
- .space ((RT_CPUS_NR - 1) * ARM_CPU_STACK_SIZE)
- #endif
- .space (ARM_CPU_STACK_SIZE)
- svc_stack_top:
- irq_stack_n:
- #if defined(RT_USING_SMP) && (RT_CPUS_NR > 1)
- .space ((RT_CPUS_NR - 1) * ARM_CPU_STACK_SIZE)
- #endif
- .space (ARM_CPU_STACK_SIZE)
- irq_stack_top:
- und_stack_n:
- #if defined(RT_USING_SMP) && (RT_CPUS_NR > 1)
- .space ((RT_CPUS_NR - 1) * ARM_CPU_STACK_SIZE)
- #endif
- .space (ARM_CPU_STACK_SIZE)
- und_stack_top:
- abt_stack_n:
- #if defined(RT_USING_SMP) && (RT_CPUS_NR > 1)
- .space ((RT_CPUS_NR - 1) * ARM_CPU_STACK_SIZE)
- #endif
- .space (ARM_CPU_STACK_SIZE)
- abt_stack_top:
|