Browse Source

更新gicv3中断亲和性设置,重构api: arm_gic_send_affinity_sgi, 增加AMP宏判断, rt_ioremap和宏RT_USING_SMART解耦
修改:中断安装默认绑定至当前核心, 修改rt_hw_interrupt_set_target_cpus函数,

wangqinglin 2 years ago
parent
commit
8335c5add2

+ 150 - 51
libcpu/aarch64/common/gicv3.c

@@ -35,9 +35,8 @@
 #define ARM_SPI_BIND_CPU_ID 0
 #endif
 
-#ifndef RT_USING_SMP
+#if !defined(RT_USING_SMP) && !defined(RT_USING_AMP)
 #define RT_CPUS_NR 1
-extern int rt_hw_cpu_id(void);
 #else
 extern rt_uint64_t rt_cpu_mpidr_early[];
 #endif /* RT_USING_SMP */
@@ -114,6 +113,7 @@ static unsigned int _gic_max_irq;
 /* Macro to access the Generic Interrupt Controller Distributor (GICD) */
 #define GIC_DIST_CTRL(hw_base)              HWREG32((hw_base) + 0x000U)
 #define GIC_DIST_TYPE(hw_base)              HWREG32((hw_base) + 0x004U)
+#define GIC_DIST_IIDR(hw_base)              HWREG32((hw_base) + 0x008U)
 #define GIC_DIST_IGROUP(hw_base, n)         HWREG32((hw_base) + 0x080U + ((n) / 32U) * 4U)
 #define GIC_DIST_ENABLE_SET(hw_base, n)     HWREG32((hw_base) + 0x100U + ((n) / 32U) * 4U)
 #define GIC_DIST_ENABLE_CLEAR(hw_base, n)   HWREG32((hw_base) + 0x180U + ((n) / 32U) * 4U)
@@ -335,6 +335,26 @@ void arm_gic_clear_active(rt_uint64_t index, int irq)
     GIC_DIST_ACTIVE_CLEAR(_gic_table[index].dist_hw_base, irq) = mask;
 }
 
+void arm_gic_set_router_cpu(rt_uint64_t index, int irq, rt_uint64_t aff)
+{
+    RT_ASSERT(index < ARM_GIC_MAX_NR);
+
+    irq = irq - _gic_table[index].offset;
+    RT_ASSERT(irq >= 32);
+
+    GIC_DIST_IROUTER(_gic_table[index].dist_hw_base, irq) = aff & 0xff00ffffffULL;
+}
+
+rt_uint64_t arm_gic_get_router_cpu(rt_uint64_t index, int irq)
+{
+    RT_ASSERT(index < ARM_GIC_MAX_NR);
+
+    irq = irq - _gic_table[index].offset;
+    RT_ASSERT(irq >= 32);
+
+    return GIC_DIST_IROUTER(_gic_table[index].dist_hw_base, irq);
+}
+
 /* Set up the cpu mask for the specific interrupt */
 void arm_gic_set_cpu(rt_uint64_t index, int irq, unsigned int cpumask)
 {
@@ -478,76 +498,118 @@ rt_uint64_t arm_gic_get_irq_status(rt_uint64_t index, int irq)
     return ((active << 1) | pending);
 }
 
-#ifdef RT_USING_SMP
-void arm_gic_send_affinity_sgi(rt_uint64_t index, int irq, rt_uint32_t cpu_masks[], rt_uint64_t routing_mode)
+#if defined(RT_USING_SMP) || defined(RT_USING_AMP)
+struct gicv3_sgi_aff
 {
-    const int cpu_mask_cpu_max_nr = sizeof(cpu_masks[0]) * 8;
-    rt_uint64_t int_id = (irq & 0xf) << 24;
-    rt_uint64_t irm = routing_mode << 40; /* Interrupt Routing Mode */
+    rt_uint64_t aff;
+    rt_uint32_t cpu_mask[(RT_CPUS_NR + 31) >> 5];
+    rt_uint16_t target_list;
+};
 
-    if (routing_mode == GICV3_ROUTED_TO_SPEC)
+static struct gicv3_sgi_aff sgi_aff_table[RT_CPUS_NR];
+static rt_uint64_t sgi_aff_table_num;
+static void sgi_aff_add_table(rt_uint64_t aff, rt_uint64_t cpu_index)
+{
+    rt_uint64_t i;
+
+    for (i = 0; i < sgi_aff_table_num; i++)
     {
-        int cpu_id, cpu_mask_bit, i, cpu_masks_nr = RT_CPUS_NR / cpu_mask_cpu_max_nr;
-        rt_uint16_t target_list;
-        rt_uint64_t rs = 0; /* Range Selector */
-        rt_uint64_t affinity_val, next_affinity_val;
+        if (sgi_aff_table[i].aff == aff)
+        {
+            sgi_aff_table[i].cpu_mask[cpu_index >> 5] |= (1 << (cpu_index & 0x1F));
+            return;
+        }
+    }
+
+    sgi_aff_table[sgi_aff_table_num].aff = aff;
+    sgi_aff_table[sgi_aff_table_num].cpu_mask[cpu_index >> 5] |= (1 << (cpu_index & 0x1F));
+    sgi_aff_table_num++;
+}
+
+static rt_uint64_t gicv3_sgi_init(void)
+{
+    rt_uint64_t i, icc_sgi1r_value;
 
-        if (cpu_masks_nr * cpu_mask_cpu_max_nr != RT_CPUS_NR)
+    for (i = 0; i < RT_CPUS_NR; i++)
+    {
+        icc_sgi1r_value =  (rt_uint64_t)((rt_cpu_mpidr_early[i] >> 8)  & 0xFF) << 16;
+        icc_sgi1r_value |= (rt_uint64_t)((rt_cpu_mpidr_early[i] >> 16) & 0xFF) << 32;
+        icc_sgi1r_value |= (rt_uint64_t)((rt_cpu_mpidr_early[i] >> 32) & 0xFF) << 48;
+        icc_sgi1r_value |= (rt_uint64_t)((rt_cpu_mpidr_early[i] >> 4)  & 0xF)  << 44;
+        sgi_aff_add_table(icc_sgi1r_value, i);
+    }
+
+    return (RT_CPUS_NR + 31) >> 5;
+}
+
+rt_inline void gicv3_sgi_send(rt_uint64_t int_id)
+{
+    rt_uint64_t i;
+    for (i = 0; i < sgi_aff_table_num; i++)
+    {
+        if (sgi_aff_table[i].target_list)
         {
-            ++cpu_masks_nr;
+            __DSB();
+            /* Interrupts routed to the PEs specified by Aff3.Aff2.Aff1.<target list>. */
+            SET_GICV3_REG(ICC_SGI1R_EL1, sgi_aff_table[i].aff | int_id | sgi_aff_table[i].target_list);
+            __ISB();
+            sgi_aff_table[i].target_list = 0;
         }
+    }
+}
+
+rt_inline void gicv3_sgi_target_list_set(rt_uint64_t array, rt_uint32_t cpu_mask)
+{
+    rt_uint64_t i, value;
 
-        for (i = cpu_id = 0; i < cpu_masks_nr;)
+    for (i = 0; i < sgi_aff_table_num; i++)
+    {
+        if (sgi_aff_table[i].cpu_mask[array] & cpu_mask)
         {
-            /* No cpu in this mask */
-            if (cpu_masks[i] == 0)
+            while (cpu_mask)
             {
-                ++i;
-                cpu_id += cpu_mask_cpu_max_nr;
-                continue;
+                value = __builtin_ctzl(cpu_mask);
+                cpu_mask &= ~(1 << value);
+                sgi_aff_table[i].target_list |= 1 << (rt_cpu_mpidr_early[(array << 5) | value] & 0xF);
             }
+        }
+    }
+}
+
+void arm_gic_send_affinity_sgi(rt_uint64_t index, int irq, rt_uint32_t cpu_masks[], rt_uint64_t routing_mode)
+{
+    rt_uint64_t i;
+    rt_uint64_t int_id = (irq & 0xf) << 24;
+    static rt_uint64_t masks_nrs = 0;
 
-            /* Get last cpu affinity value */
-            affinity_val = rt_cpu_mpidr_early[cpu_id] & 0xff00ffff00ULL;
+    if (routing_mode == GICV3_ROUTED_TO_SPEC)
+    {
+        if (!masks_nrs)
+        {
+            masks_nrs = gicv3_sgi_init();
+        }
 
-            /* Read 16 cpus information */
-            for (cpu_mask_bit = 0; cpu_mask_bit < 16; ++cpu_mask_bit, ++cpu_id)
+        for (i = 0; i < masks_nrs; i++)
+        {
+            if (cpu_masks[i] == 0)
             {
-                /* MPIDR_EL1: aff3[39:32], aff2[23:16], aff1[15:8] */
-                next_affinity_val = rt_cpu_mpidr_early[cpu_id] & 0xff00ffff00ULL;
-
-                /* Affinity value is different, read end */
-                if (affinity_val != next_affinity_val)
-                {
-                    break;
-                }
+                continue;
             }
 
-            /* Get all valid cpu mask */
-            target_list = (0xffff >> (16 - cpu_mask_bit)) & cpu_masks[i];
-            /* Clear read mask */
-            cpu_masks[i] >>= cpu_mask_bit;
-            /* ICC_SGI1R_EL1: aff3[55:48], aff2[39:32], aff1[23:16] */
-            affinity_val <<= 8;
-
-            __DSB();
-            /* Interrupts routed to the PEs specified by Aff3.Aff2.Aff1.<target list>. */
-            SET_GICV3_REG(ICC_SGI1R_EL1, affinity_val | (rs << 44) | irm | int_id | target_list);
-            __ISB();
-
-            /* Check if reset the range selector */
-            rs = affinity_val != next_affinity_val ? 0 : rs + 1;
+            gicv3_sgi_target_list_set(i, cpu_masks[i]);
         }
+
+        gicv3_sgi_send(int_id);
     }
     else
     {
         __DSB();
         /* Interrupts routed to all PEs in the system, excluding "self". */
-        SET_GICV3_REG(ICC_SGI1R_EL1, irm | int_id);
+        SET_GICV3_REG(ICC_SGI1R_EL1, (0x10000000000ULL) | int_id);
         __ISB();
     }
 }
-#endif /* RT_USING_SMP */
+#endif /* defined(RT_USING_SMP) || defined(RT_USING_AMP) */
 
 rt_uint64_t arm_gic_get_high_pending_irq(rt_uint64_t index)
 {
@@ -636,6 +698,8 @@ int arm_gic_dist_init(rt_uint64_t index, rt_uint64_t dist_base, int irq_start)
     unsigned int gic_type;
     rt_uint64_t main_cpu_affinity_val;
 
+    RT_UNUSED(i);
+    RT_UNUSED(main_cpu_affinity_val);
     RT_ASSERT(index < ARM_GIC_MAX_NR);
 
     _gic_table[index].dist_hw_base = dist_base;
@@ -660,6 +724,8 @@ int arm_gic_dist_init(rt_uint64_t index, rt_uint64_t dist_base, int irq_start)
         _gic_max_irq = ARM_GIC_NR_IRQS;
     }
 
+#ifndef RT_AMP_SLAVE
+
     GIC_DIST_CTRL(dist_base) = 0;
     /* Wait for register write pending */
     arm_gicv3_wait_rwp(0, 32);
@@ -724,6 +790,7 @@ int arm_gic_dist_init(rt_uint64_t index, rt_uint64_t dist_base, int irq_start)
      */
     GIC_DIST_CTRL(dist_base) = GICD_CTLR_ARE_NS | GICD_CTLR_ENGRP1NS;
 
+#endif /* RT_AMP_SLAVE */
     return 0;
 }
 
@@ -811,13 +878,19 @@ int arm_gic_cpu_init(rt_uint64_t index, rt_uint64_t cpu_base)
 void arm_gic_dump_type(rt_uint64_t index)
 {
     unsigned int gic_type;
+    unsigned int gic_version;
+    unsigned int gic_rp;
 
+    gic_version = (GIC_DIST_IIDR(_gic_table[index].dist_hw_base) >> 24) & 0xfUL;
+    gic_rp = (GIC_DIST_IIDR(_gic_table[index].dist_hw_base) >> 12) & 0xfUL;
     gic_type = GIC_DIST_TYPE(_gic_table[index].dist_hw_base);
-    rt_kprintf("GICv%d on %p, max IRQs: %d, %s security extension(%08x)\n",
-               (GIC_DIST_ICPIDR2(_gic_table[index].dist_hw_base) >> 4) & 0xf,
+    rt_kprintf("GICv3-%d r%dp%d on %p, max IRQs: %d, %s security extension(%08x)\n",
+               (gic_version == 0) ? 500 : (gic_version == 2) ? 600 : 0,
+               (gic_rp >> 4) & 0xF,
+                gic_rp & 0xF,
                _gic_table[index].dist_hw_base,
                _gic_max_irq,
-               gic_type & (1 << 10) ? "has" : "no",
+               gic_type & (1U << 10U) ? "has" : "no",
                gic_type);
 }
 
@@ -850,10 +923,36 @@ void arm_gic_dump(rt_uint64_t index)
     rt_kprintf("\b\b\n");
 }
 
+static void arm_gic_bind_dump(void)
+{
+#ifdef BSP_USING_GICV3
+    int i;
+    for (i = 32; i < _gic_max_irq; i++)
+    {
+        rt_kprintf("irq(%d) -> 0x%X\n", i, arm_gic_get_router_cpu(0, i));
+    }
+#endif  /* BSP_USING_GICV3 */
+}
+
+static void arm_gic_sgi_dump(rt_uint64_t index)
+{
+    rt_int32_t cpu_id = rt_hw_cpu_id();
+
+    rt_kprintf("redist_hw_base = 0x%X\n", _gic_table[index].redist_hw_base[cpu_id]);
+    rt_kprintf("--- sgi mask ---\n");
+    rt_kprintf("0x%08x\n", GIC_RDISTSGI_ISENABLER0(_gic_table[index].redist_hw_base[cpu_id]));
+    rt_kprintf("--- sgi pending ---\n");
+    rt_kprintf("0x%08x\n", GIC_RDISTSGI_ISPENDR0(_gic_table[index].redist_hw_base[cpu_id]));
+    rt_kprintf("--- sgi active ---\n");
+    rt_kprintf("0x%08x\n", GIC_RDISTSGI_ISACTIVER0(_gic_table[index].redist_hw_base[cpu_id]));
+}
+
 long gic_dump(void)
 {
     arm_gic_dump_type(0);
     arm_gic_dump(0);
+    arm_gic_bind_dump();
+    arm_gic_sgi_dump(0);
 
     return 0;
 }

+ 2 - 1
libcpu/aarch64/common/gicv3.h

@@ -41,6 +41,7 @@ rt_uint64_t arm_gic_get_configuration(rt_uint64_t index, int irq);
 
 void arm_gic_clear_active(rt_uint64_t index, int irq);
 
+void arm_gic_set_router_cpu(rt_uint64_t index, int irq, rt_uint64_t aff);
 void arm_gic_set_cpu(rt_uint64_t index, int irq, unsigned int cpumask);
 rt_uint64_t arm_gic_get_target_cpu(rt_uint64_t index, int irq);
 
@@ -55,7 +56,7 @@ rt_uint64_t arm_gic_get_binary_point(rt_uint64_t index);
 
 rt_uint64_t arm_gic_get_irq_status(rt_uint64_t index, int irq);
 
-#ifdef RT_USING_SMP
+#if defined(RT_USING_SMP) || defined(RT_USING_AMP)
 void arm_gic_send_affinity_sgi(rt_uint64_t index, int irq, rt_uint32_t cpu_masks[], rt_uint64_t routing_mode);
 #endif
 

+ 26 - 9
libcpu/aarch64/common/interrupt.c

@@ -14,12 +14,8 @@
 #include "interrupt.h"
 #include "gic.h"
 #include "gicv3.h"
-
-#ifdef RT_USING_SMART
 #include "ioremap.h"
-#else
-#define rt_ioremap(x, ...) (x)
-#endif
+
 
 /* exception and interrupt handler table */
 struct rt_irq_desc isr_table[MAX_HANDLERS];
@@ -215,9 +211,15 @@ void rt_hw_interrupt_ack(int vector)
  * @param vector:   the interrupt number
  *        cpu_mask: target cpus mask, one bit for one core
  */
-void rt_hw_interrupt_set_target_cpus(int vector, unsigned int cpu_mask)
+void rt_hw_interrupt_set_target_cpus(int vector, unsigned long cpu_mask)
 {
-    arm_gic_set_cpu(0, vector, cpu_mask);
+#ifdef BSP_USING_GIC
+#ifdef BSP_USING_GICV3
+    arm_gic_set_router_cpu(0, vector, cpu_mask);
+#else
+    arm_gic_set_cpu(0, vector, (unsigned int) cpu_mask);
+#endif
+#endif
 }
 
 /**
@@ -379,16 +381,31 @@ rt_isr_handler_t rt_hw_interrupt_install(int vector, rt_isr_handler_t handler,
         }
     }
 
+#ifdef BSP_USING_GIC
+    if (vector > 32)
+    {
+#ifdef BSP_USING_GICV3
+        rt_uint64_t cpu_affinity_val;
+        __asm__ volatile ("mrs %0, mpidr_el1":"=r"(cpu_affinity_val));
+        rt_hw_interrupt_set_target_cpus(vector, cpu_affinity_val);
+#else
+        rt_hw_interrupt_set_target_cpus(vector, rt_hw_cpu_id());
+#endif /* BSP_USING_GICV3 */
+    }
+#endif
+
     return old_handler;
 }
 
-#ifdef RT_USING_SMP
+#if defined(RT_USING_SMP) || defined(RT_USING_AMP)
 void rt_hw_ipi_send(int ipi_vector, unsigned int cpu_mask)
 {
 #ifdef BSP_USING_GICV2
     arm_gic_send_sgi(0, ipi_vector, cpu_mask, 0);
 #elif defined(BSP_USING_GICV3)
-    arm_gic_send_affinity_sgi(0, ipi_vector, (unsigned int *)&cpu_mask, GICV3_ROUTED_TO_SPEC);
+    rt_uint32_t gicv3_cpu_mask[(RT_CPUS_NR + 31) >> 5];
+    gicv3_cpu_mask[0] = cpu_mask;
+    arm_gic_send_affinity_sgi(0, ipi_vector, gicv3_cpu_mask, GICV3_ROUTED_TO_SPEC);
 #endif
 }
 

+ 2 - 3
libcpu/aarch64/common/interrupt.h

@@ -30,7 +30,7 @@ void rt_hw_interrupt_umask(int vector);
 int rt_hw_interrupt_get_irq(void);
 void rt_hw_interrupt_ack(int vector);
 
-void rt_hw_interrupt_set_target_cpus(int vector, unsigned int cpu_mask);
+void rt_hw_interrupt_set_target_cpus(int vector, unsigned long cpu_mask);
 unsigned int rt_hw_interrupt_get_target_cpus(int vector);
 
 void rt_hw_interrupt_set_triger_mode(int vector, unsigned int mode);
@@ -52,8 +52,7 @@ unsigned int rt_hw_interrupt_get_prior_group_bits(void);
 rt_isr_handler_t rt_hw_interrupt_install(int vector, rt_isr_handler_t handler,
         void *param, const char *name);
 
-#ifdef RT_USING_SMP
-void rt_hw_ipi_send(int ipi_vector, unsigned int cpu_mask);
+#if defined(RT_USING_SMP) || defined(RT_USING_AMP)
 void rt_hw_ipi_handler_install(int ipi_vector, rt_isr_handler_t ipi_isr_handler);
 #endif