Przeglądaj źródła

!302 [fix] some error when porting rt-smart
Merge pull request !302 from 我夏了夏天/rt-smart-summer

bernard 4 lat temu
rodzic
commit
3c8c1df196

+ 2 - 2
libcpu/aarch64/common/context_gcc.S

@@ -120,9 +120,9 @@ rt_hw_get_gtimer_frq:
     BL      lwp_check_exit
 
     BL      rt_thread_self
-    MOV     X8, X0
+    MOV     X19, X0
     BL      lwp_mmu_switch
-    MOV     X0, X8
+    MOV     X0, X19
     BL      lwp_user_setting_restore
 
     LDP     X2, X3, [SP], #0x10  /* SPSR and ELR. */

+ 3 - 3
libcpu/aarch64/common/interrupt.c

@@ -25,8 +25,8 @@ rt_ubase_t rt_thread_switch_interrupt_flag = 0;
 #endif
 
 const unsigned int VECTOR_BASE = 0x00;
-extern void rt_cpu_vector_set_base(unsigned int addr);
-extern int system_vectors;
+extern void rt_cpu_vector_set_base(void *addr);
+extern void *system_vectors;
 
 #ifdef RT_USING_SMP
 #define rt_interrupt_nest rt_cpu_self()->irq_nest
@@ -47,7 +47,7 @@ static void default_isr_handler(int vector, void *param)
 
 void rt_hw_vector_init(void)
 {
-    rt_cpu_vector_set_base((unsigned long)&system_vectors);
+    rt_cpu_vector_set_base(&system_vectors);
 }
 
 /**

+ 53 - 61
libcpu/aarch64/cortex-a/entry_point.S

@@ -4,97 +4,90 @@
  * SPDX-License-Identifier: Apache-2.0
  *
  * Date           Author       Notes
- * 2020-01-15     bigmagic    the first version
+ * 2020-01-15     bigmagic     the first version
+ * 2020-08-10     SummerGift   support clang compiler
  */
 
 #include "rtconfig.h"
 .section ".text.entrypoint","ax"
-
 .set EL1_stack,         __el1_stack
-
 .global __start
 
-// This symbol is set to 0x80000 in ld script. That is the address that raspi3's firmware
-// loads 'kernel8.img' file in.
 __start:
-    // read cpu id, stop slave cores
-    mrs     x1, mpidr_el1            // MPIDR_EL1: Multi-Processor Affinity Register
+    /* read cpu id, stop slave cores */
+    mrs     x1, mpidr_el1           /* MPIDR_EL1: Multi-Processor Affinity Register */
     and     x1, x1, #3
-    cbz     x1, .L__cpu_0             // .L prefix is the local label in ELF
+    cbz     x1, .L__cpu_0           /* .L prefix is the local label in ELF */
 
-    // cpu id > 0, stop
-    // cpu id == 0 will also goto here after returned from entry() if possible
+    /* cpu id > 0, stop */
+    /* cpu id == 0 will also goto here after returned from entry() if possible */
 .L__current_cpu_idle:
     wfe
     b       .L__current_cpu_idle
 
-.L__cpu_0:  // cpu id == 0
-
-    // set stack before our code
-
-	/* Define stack pointer for current exception level */
-	// ldr	x2, =EL1_stack
-	// mov	sp, x2
+.L__cpu_0:
+    /* set stack before our code, Define stack pointer for current exception level */
+    /* ldr	x2, =EL1_stack */
+    /* mov	sp, x2         */
 
     adr     x1, __start
 
-    // set up EL1
-    mrs     x0, CurrentEL           // CurrentEL Register. bit 2, 3. Others reserved
-    and     x0, x0, #12             // clear reserved bits
+    /* set up EL1 */
+    mrs     x0, CurrentEL           /* CurrentEL Register. bit 2, 3. Others reserved */
+    and     x0, x0, #12             /* clear reserved bits */
 
-    // running at EL3?
-    cmp     x0, #12                // 1100b. So, EL3
-    bne     .L__not_in_el3         // 11?  !EL3 -> 5:
+    /* running at EL3? */
+    cmp     x0, #12                 /* 1100b. So, EL3 */
+    bne     .L__not_in_el3          /* 11?  !EL3 -> 5: */
 
-    // should never be executed, just for completeness. (EL3)
+    /* should never be executed, just for completeness. (EL3) */
     mov     x2, #0x5b1
-    msr     scr_el3, x2            // SCR_ELn  Secure Configuration Register
+    msr     scr_el3, x2             /* SCR_ELn  Secure Configuration Register */
     mov     x2, #0x3c9
-    msr     spsr_el3, x2           // SPSR_ELn. Saved Program Status Register. 1111001001
+    msr     spsr_el3, x2            /* SPSR_ELn. Saved Program Status Register. 1111001001 */
     adr     x2, .L__not_in_el3
     msr     elr_el3, x2
-    eret                           // Exception Return: from EL3, continue from .L__not_in_el3
-
-    // running at EL2 or EL1
-.L__not_in_el3:
-    cmp     x0, #4                 // 0x04  0100 EL1
-    beq     .L__in_el1             // EL1 -> 5:
+    eret                            /* Exception Return: from EL3, continue from .L__not_in_el3 */
 
+.L__not_in_el3:                     /* running at EL2 or EL1 */
+    cmp     x0, #4                  /* 0x04  0100 EL1 */
+    beq     .L__in_el1              /* EL1 -> 5: */
+ 
     mrs     x0, hcr_el2
     bic     x0, x0, #0xff
     msr     hcr_el2, x0
-    // in EL2
-    msr     sp_el1, x1             // Set sp of EL1 to _start
 
-    // enable CNTP for EL1
-    mrs     x0, cnthctl_el2         // Counter-timer Hypervisor Control register
+    msr     sp_el1, x1              /* in EL2, set sp of EL1 to _start */
+
+    /* enable CNTP for EL1 */
+    mrs     x0, cnthctl_el2         /* Counter-timer Hypervisor Control register */
     orr     x0, x0, #3
     msr     cnthctl_el2, x0
     msr     cntvoff_el2, xzr
 
-    // enable AArch64 in EL1
-    mov     x0, #(1 << 31)          // AArch64
-    orr     x0, x0, #(1 << 1)       // SWIO hardwired on Pi3
+    /* enable AArch64 in EL1 */
+    mov     x0, #(1 << 31)          /* AArch64 */
+    orr     x0, x0, #(1 << 1)       /* SWIO hardwired on Pi3 */
     msr     hcr_el2, x0
     mrs     x0, hcr_el2
 
-    // change execution level to EL1
+    /* change execution level to EL1 */
     mov     x2, #0x3c4
-    msr     spsr_el2, x2        // 1111000100
+    msr     spsr_el2, x2            /* 1111000100 */
     adr     x2, .L__in_el1
     msr     elr_el2, x2
 
-    eret                        // exception return. from EL2. continue from .L__in_el1
+    eret                            /* exception return. from EL2. continue from .L__in_el1 */
 
 .L__in_el1:
     ldr     x9, =PV_OFFSET
-    mov     sp, x1             // in EL1. Set sp to _start
+    mov     sp, x1                  /* in EL1. Set sp to _start */
 
-    // Set CPACR_EL1 (Architecture Feature Access Control Register) to avoid trap from SIMD or float point instruction
-    mov     x1, #0x00300000       // Don't trap any SIMD/FP instructions in both EL0 and EL1
+    /* Set CPACR_EL1 (Architecture Feature Access Control Register) to avoid trap from SIMD or float point instruction */
+    mov     x1, #0x00300000         /* Don't trap any SIMD/FP instructions in both EL0 and EL1 */
     msr     cpacr_el1, x1
 
-    // clear bss
+    /* clear bss */
     ldr     x1, =__bss_start
     add     x1, x1, x9
     ldr     w2, =__bss_size
@@ -105,8 +98,7 @@ __start:
     sub     w2, w2, #1
     cbnz    w2, .L__clean_bss_loop
 
-    // jump to C code, should not return
-.L__jump_to_entry:
+.L__jump_to_entry:          /* jump to C code, should not return */        
 
     bl get_free_page
     mov x21, x0
@@ -123,32 +115,32 @@ __start:
     msr ttbr1_el1, x1
     dsb sy
 
-    ldr x2, =0x40000000 //1G
-    mov x3, x9
+    ldr x2, =0x40000000      /* map 1G memory for kernel space */
+    ldr x3, =PV_OFFSET
     bl rt_hw_mmu_setup_early
 
     ldr x30, =after_mmu_enable
 
     mrs x1, sctlr_el1
-    bic x1, x1, #(3 << 3) /* dis SA, SA0 */
-    bic x1, x1, #(1 << 1) /* dis A */
-    orr x1, x1, #(1 << 12) /* I */
-    orr x1, x1, #(1 << 2) /* C */
-    orr x1, x1, #(1 << 0) /* M */
-    msr sctlr_el1, x1
+    bic x1, x1, #(3 << 3)    /* dis SA, SA0 */
+    bic x1, x1, #(1 << 1)    /* dis A */
+    orr x1, x1, #(1 << 12)   /* I */
+    orr x1, x1, #(1 << 2)    /* C */
+    orr x1, x1, #(1 << 0)    /* M */
+    msr sctlr_el1, x1        /* enable MMU */
+
     dsb sy
     isb sy
-    ic ialluis
+    ic ialluis               /* Invalidate all instruction caches in Inner Shareable domain to Point of Unification */
     dsb sy
     isb sy
-    tlbi vmalle1
+    tlbi vmalle1             /* Invalidate all stage 1 translations used at EL1 with the current VMID */
     dsb sy
     isb sy
     ret
 
 after_mmu_enable:
-    // disable ttbr0
-    mrs x0, tcr_el1
+    mrs x0, tcr_el1          /* disable ttbr0, only using kernel space */
     orr x0, x0, #(1 << 7)
     msr tcr_el1, x0
     msr ttbr0_el1, xzr
@@ -157,6 +149,6 @@ after_mmu_enable:
     mov     x0, #1
     msr     spsel, x0
     adr     x1, __start
-    mov     sp, x1             // sp_el1 set to _start
+    mov     sp, x1           /* sp_el1 set to _start */
 
     b  rtthread_startup