123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505 |
- /*
- * Copyright (c) 2006-2021, RT-Thread Development Team
- *
- * SPDX-License-Identifier: Apache-2.0
- *
- * Change Logs:
- * Date Author Notes
- * 2021-01-30 lizhirui first version
- */
- #include "rtconfig.h"
- #include <rtthread.h>
- #include <stddef.h>
- #include <stdint.h>
- #include <cache.h>
- #include <mm_aspace.h>
- #include <mm_page.h>
- #include <mmu.h>
- #include <riscv_mmu.h>
- #include <tlb.h>
- #ifdef RT_USING_SMART
- #include <ioremap.h>
- #include <lwp_user_mm.h>
- #include <tlb.h>
- #endif
- #define DBG_TAG "MMU"
- #define DBG_LVL DBG_LOG
- #include <rtdbg.h>
- #ifndef RT_USING_SMART
- #define PV_OFFSET 0
- #define USER_VADDR_START 0
- #endif
- static size_t _unmap_area(struct rt_aspace *aspace, void *v_addr, size_t size);
- void rt_hw_aspace_switch(rt_aspace_t aspace)
- {
- uintptr_t page_table = (uintptr_t)_rt_kmem_v2p(aspace->page_table);
- write_csr(satp, (((size_t)SATP_MODE) << SATP_MODE_OFFSET) |
- ((rt_ubase_t)page_table >> PAGE_OFFSET_BIT));
- rt_hw_tlb_invalidate_all_local();
- }
- static void *current_mmu_table = RT_NULL;
- volatile __attribute__((aligned(4 * 1024)))
- rt_ubase_t MMUTable[__SIZE(VPN2_BIT)];
- void *rt_hw_mmu_tbl_get()
- {
- return current_mmu_table;
- }
- static int _map_one_page(struct rt_aspace *aspace, void *va, void *pa,
- size_t attr)
- {
- rt_size_t l1_off, l2_off, l3_off;
- rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
- l1_off = GET_L1((size_t)va);
- l2_off = GET_L2((size_t)va);
- l3_off = GET_L3((size_t)va);
- mmu_l1 = ((rt_size_t *)aspace->page_table) + l1_off;
- if (PTE_USED(*mmu_l1))
- {
- mmu_l2 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l1), PV_OFFSET);
- }
- else
- {
- mmu_l2 = (rt_size_t *)rt_pages_alloc(0);
- if (mmu_l2)
- {
- rt_memset(mmu_l2, 0, PAGE_SIZE);
- rt_hw_cpu_dcache_clean(mmu_l2, PAGE_SIZE);
- *mmu_l1 = COMBINEPTE((rt_size_t)VPN_TO_PPN(mmu_l2, PV_OFFSET),
- PAGE_DEFAULT_ATTR_NEXT);
- rt_hw_cpu_dcache_clean(mmu_l1, sizeof(*mmu_l1));
- }
- else
- {
- return -1;
- }
- }
- if (PTE_USED(*(mmu_l2 + l2_off)))
- {
- RT_ASSERT(!PAGE_IS_LEAF(*(mmu_l2 + l2_off)));
- mmu_l3 =
- (rt_size_t *)PPN_TO_VPN(GET_PADDR(*(mmu_l2 + l2_off)), PV_OFFSET);
- }
- else
- {
- mmu_l3 = (rt_size_t *)rt_pages_alloc(0);
- if (mmu_l3)
- {
- rt_memset(mmu_l3, 0, PAGE_SIZE);
- rt_hw_cpu_dcache_clean(mmu_l3, PAGE_SIZE);
- *(mmu_l2 + l2_off) =
- COMBINEPTE((rt_size_t)VPN_TO_PPN(mmu_l3, PV_OFFSET),
- PAGE_DEFAULT_ATTR_NEXT);
- rt_hw_cpu_dcache_clean(mmu_l2, sizeof(*mmu_l2));
- // declares a reference to parent page table
- rt_page_ref_inc((void *)mmu_l2, 0);
- }
- else
- {
- return -1;
- }
- }
- RT_ASSERT(!PTE_USED(*(mmu_l3 + l3_off)));
- // declares a reference to parent page table
- rt_page_ref_inc((void *)mmu_l3, 0);
- *(mmu_l3 + l3_off) = COMBINEPTE((rt_size_t)pa, attr);
- rt_hw_cpu_dcache_clean(mmu_l3 + l3_off, sizeof(*(mmu_l3 + l3_off)));
- return 0;
- }
- /** rt_hw_mmu_map will never override existed page table entry */
- void *rt_hw_mmu_map(struct rt_aspace *aspace, void *v_addr, void *p_addr,
- size_t size, size_t attr)
- {
- int ret = -1;
- void *unmap_va = v_addr;
- size_t npages = size >> ARCH_PAGE_SHIFT;
- // TODO trying with HUGEPAGE here
- while (npages--)
- {
- ret = _map_one_page(aspace, v_addr, p_addr, attr);
- if (ret != 0)
- {
- /* error, undo map */
- while (unmap_va != v_addr)
- {
- MM_PGTBL_LOCK(aspace);
- _unmap_area(aspace, unmap_va, ARCH_PAGE_SIZE);
- MM_PGTBL_UNLOCK(aspace);
- unmap_va += ARCH_PAGE_SIZE;
- }
- break;
- }
- v_addr += ARCH_PAGE_SIZE;
- p_addr += ARCH_PAGE_SIZE;
- }
- if (ret == 0)
- {
- return unmap_va;
- }
- return NULL;
- }
- static void _unmap_pte(rt_size_t *pentry, rt_size_t *lvl_entry[], int level)
- {
- int loop_flag = 1;
- while (loop_flag)
- {
- loop_flag = 0;
- *pentry = 0;
- rt_hw_cpu_dcache_clean(pentry, sizeof(*pentry));
- // we don't handle level 0, which is maintained by caller
- if (level > 0)
- {
- void *page = (void *)((rt_ubase_t)pentry & ~ARCH_PAGE_MASK);
- // decrease reference from child page to parent
- rt_pages_free(page, 0);
- int free = rt_page_ref_get(page, 0);
- if (free == 1)
- {
- rt_pages_free(page, 0);
- pentry = lvl_entry[--level];
- loop_flag = 1;
- }
- }
- }
- }
- static size_t _unmap_area(struct rt_aspace *aspace, void *v_addr, size_t size)
- {
- rt_size_t loop_va = __UMASKVALUE((rt_size_t)v_addr, PAGE_OFFSET_MASK);
- size_t unmapped = 0;
- int i = 0;
- rt_size_t lvl_off[3];
- rt_size_t *lvl_entry[3];
- lvl_off[0] = (rt_size_t)GET_L1(loop_va);
- lvl_off[1] = (rt_size_t)GET_L2(loop_va);
- lvl_off[2] = (rt_size_t)GET_L3(loop_va);
- unmapped = 1 << (ARCH_PAGE_SHIFT + ARCH_INDEX_WIDTH * 2ul);
- rt_size_t *pentry;
- lvl_entry[i] = ((rt_size_t *)aspace->page_table + lvl_off[i]);
- pentry = lvl_entry[i];
- // find leaf page table entry
- while (PTE_USED(*pentry) && !PAGE_IS_LEAF(*pentry))
- {
- i += 1;
- lvl_entry[i] = ((rt_size_t *)PPN_TO_VPN(GET_PADDR(*pentry), PV_OFFSET) +
- lvl_off[i]);
- pentry = lvl_entry[i];
- unmapped >>= ARCH_INDEX_WIDTH;
- }
- // clear PTE & setup its
- if (PTE_USED(*pentry))
- {
- _unmap_pte(pentry, lvl_entry, i);
- }
- return unmapped;
- }
- /** unmap is different from map that it can handle multiple pages */
- void rt_hw_mmu_unmap(struct rt_aspace *aspace, void *v_addr, size_t size)
- {
- // caller guarantee that v_addr & size are page aligned
- if (!aspace->page_table)
- {
- return;
- }
- size_t unmapped = 0;
- while (size > 0)
- {
- MM_PGTBL_LOCK(aspace);
- unmapped = _unmap_area(aspace, v_addr, size);
- MM_PGTBL_UNLOCK(aspace);
- // when unmapped == 0, region not exist in pgtbl
- if (!unmapped || unmapped > size)
- break;
- size -= unmapped;
- v_addr += unmapped;
- }
- }
- #ifdef RT_USING_SMART
- static inline void _init_region(void *vaddr, size_t size)
- {
- rt_ioremap_start = vaddr;
- rt_ioremap_size = size;
- rt_mpr_start = rt_ioremap_start - rt_mpr_size;
- rt_kprintf("rt_ioremap_start: %p, rt_mpr_start: %p\n", rt_ioremap_start, rt_mpr_start);
- }
- #else
- static inline void _init_region(void *vaddr, size_t size)
- {
- rt_mpr_start = vaddr - rt_mpr_size;
- }
- #endif
- int rt_hw_mmu_map_init(rt_aspace_t aspace, void *v_address, rt_size_t size,
- rt_size_t *vtable, rt_size_t pv_off)
- {
- size_t l1_off, va_s, va_e;
- rt_base_t level;
- if ((!aspace) || (!vtable))
- {
- return -1;
- }
- va_s = (rt_size_t)v_address;
- va_e = ((rt_size_t)v_address) + size - 1;
- if (va_e < va_s)
- {
- return -1;
- }
- // convert address to PPN2 index
- va_s = GET_L1(va_s);
- va_e = GET_L1(va_e);
- if (va_s == 0)
- {
- return -1;
- }
- // vtable initialization check
- for (l1_off = va_s; l1_off <= va_e; l1_off++)
- {
- size_t v = vtable[l1_off];
- if (v)
- {
- return -1;
- }
- }
- rt_aspace_init(&rt_kernel_space, (void *)0x1000, USER_VADDR_START - 0x1000,
- vtable);
- _init_region(v_address, size);
- return 0;
- }
- const static int max_level =
- (ARCH_VADDR_WIDTH - ARCH_PAGE_SHIFT) / ARCH_INDEX_WIDTH;
- static inline uintptr_t _get_level_size(int level)
- {
- return 1ul << (ARCH_PAGE_SHIFT + (max_level - level) * ARCH_INDEX_WIDTH);
- }
- static rt_size_t *_query(struct rt_aspace *aspace, void *vaddr, int *level)
- {
- rt_size_t l1_off, l2_off, l3_off;
- rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
- rt_size_t pa;
- l1_off = GET_L1((rt_size_t)vaddr);
- l2_off = GET_L2((rt_size_t)vaddr);
- l3_off = GET_L3((rt_size_t)vaddr);
- if (!aspace)
- {
- LOG_W("%s: no aspace", __func__);
- return RT_NULL;
- }
- mmu_l1 = ((rt_size_t *)aspace->page_table) + l1_off;
- if (PTE_USED(*mmu_l1))
- {
- if (*mmu_l1 & PTE_XWR_MASK)
- {
- *level = 1;
- return mmu_l1;
- }
- mmu_l2 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l1), PV_OFFSET);
- if (PTE_USED(*(mmu_l2 + l2_off)))
- {
- if (*(mmu_l2 + l2_off) & PTE_XWR_MASK)
- {
- *level = 2;
- return mmu_l2 + l2_off;
- }
- mmu_l3 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*(mmu_l2 + l2_off)),
- PV_OFFSET);
- if (PTE_USED(*(mmu_l3 + l3_off)))
- {
- *level = 3;
- return mmu_l3 + l3_off;
- }
- }
- }
- return RT_NULL;
- }
- void *rt_hw_mmu_v2p(struct rt_aspace *aspace, void *vaddr)
- {
- int level;
- uintptr_t *pte = _query(aspace, vaddr, &level);
- uintptr_t paddr;
- if (pte)
- {
- paddr = GET_PADDR(*pte);
- paddr |= ((intptr_t)vaddr & (_get_level_size(level) - 1));
- }
- else
- {
- paddr = 0;
- }
- return (void *)paddr;
- }
- static int _noncache(uintptr_t *pte)
- {
- return 0;
- }
- static int _cache(uintptr_t *pte)
- {
- return 0;
- }
- static int (*control_handler[MMU_CNTL_DUMMY_END])(uintptr_t *pte) = {
- [MMU_CNTL_CACHE] = _cache,
- [MMU_CNTL_NONCACHE] = _noncache,
- };
- int rt_hw_mmu_control(struct rt_aspace *aspace, void *vaddr, size_t size,
- enum rt_mmu_cntl cmd)
- {
- int level;
- int err = -RT_EINVAL;
- void *vend = vaddr + size;
- int (*handler)(uintptr_t * pte);
- if (cmd >= 0 && cmd < MMU_CNTL_DUMMY_END)
- {
- handler = control_handler[cmd];
- while (vaddr < vend)
- {
- uintptr_t *pte = _query(aspace, vaddr, &level);
- void *range_end = vaddr + _get_level_size(level);
- RT_ASSERT(range_end < vend);
- if (pte)
- {
- err = handler(pte);
- RT_ASSERT(err == RT_EOK);
- }
- vaddr = range_end;
- }
- }
- else
- {
- err = -RT_ENOSYS;
- }
- return err;
- }
- /**
- * @brief setup Page Table for kernel space. It's a fixed map
- * and all mappings cannot be changed after initialization.
- *
- * Memory region in struct mem_desc must be page aligned,
- * otherwise is a failure and no report will be
- * returned.
- *
- * @param mmu_info
- * @param mdesc
- * @param desc_nr
- */
- void rt_hw_mmu_setup(rt_aspace_t aspace, struct mem_desc *mdesc, int desc_nr)
- {
- void *err;
- for (size_t i = 0; i < desc_nr; i++)
- {
- size_t attr;
- switch (mdesc->attr)
- {
- case NORMAL_MEM:
- attr = MMU_MAP_K_RWCB;
- break;
- case NORMAL_NOCACHE_MEM:
- attr = MMU_MAP_K_RWCB;
- break;
- case DEVICE_MEM:
- attr = MMU_MAP_K_DEVICE;
- break;
- default:
- attr = MMU_MAP_K_DEVICE;
- }
- struct rt_mm_va_hint hint = {.flags = MMF_MAP_FIXED,
- .limit_start = aspace->start,
- .limit_range_size = aspace->size,
- .map_size = mdesc->vaddr_end -
- mdesc->vaddr_start + 1,
- .prefer = (void *)mdesc->vaddr_start};
- rt_aspace_map_phy_static(aspace, &mdesc->varea, &hint, attr,
- mdesc->paddr_start >> MM_PAGE_SHIFT, &err);
- mdesc++;
- }
- rt_hw_aspace_switch(&rt_kernel_space);
- rt_page_cleanup();
- }
- void rt_hw_mmu_kernel_map_init(rt_aspace_t aspace, rt_size_t vaddr_start, rt_size_t size)
- {
- rt_size_t paddr_start =
- __UMASKVALUE(VPN_TO_PPN(vaddr_start, PV_OFFSET), PAGE_OFFSET_MASK);
- rt_size_t va_s = GET_L1(vaddr_start);
- rt_size_t va_e = GET_L1(vaddr_start + size - 1);
- rt_size_t i;
- for (i = va_s; i <= va_e; i++)
- {
- MMUTable[i] =
- COMBINEPTE(paddr_start, PAGE_ATTR_RWX | PTE_G | PTE_V | PTE_CACHE |
- PTE_SHARE | PTE_BUF | PTE_A | PTE_D);
- paddr_start += L1_PAGE_SIZE;
- }
- rt_hw_tlb_invalidate_all_local();
- }
|