123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646 |
- /*
- * Copyright (c) 2006-2021, RT-Thread Development Team
- *
- * SPDX-License-Identifier: Apache-2.0
- *
- * Change Logs:
- * Date Author Notes
- * 2021-01-30 lizhirui first version
- */
- #include <rtthread.h>
- #include <rthw.h>
- #ifdef RT_USING_SMART
- #include <board.h>
- #include <page.h>
- #include <stdlib.h>
- #include <string.h>
- #include <lwp_mm.h>
- #include <cache.h>
- #define DBG_TAG "mmu"
- #define DBG_LVL DBG_INFO
- #include <rtdbg.h>
- #include "riscv.h"
- #include "riscv_mmu.h"
- #include "mmu.h"
- void *current_mmu_table = RT_NULL;
- volatile rt_ubase_t MMUTable[__SIZE(VPN2_BIT)] __attribute__((aligned(4 * 1024)));
- static void rt_hw_cpu_tlb_invalidate()
- {
- rt_size_t satpv = read_csr(satp);
- write_csr(satp, satpv);
- mmu_flush_tlb();
- }
- void *rt_hw_mmu_tbl_get()
- {
- return current_mmu_table;
- }
- void rt_hw_mmu_switch(void *mmu_table)
- {
- current_mmu_table = mmu_table;
- RT_ASSERT(__CHECKALIGN(mmu_table, PAGE_OFFSET_BIT));
- mmu_set_pagetable((rt_ubase_t)mmu_table);
- rt_hw_cpu_dcache_clean_all();
- rt_hw_cpu_icache_invalidate_all();
- }
- int rt_hw_mmu_map_init(rt_mmu_info *mmu_info, void *v_address, rt_size_t size, rt_size_t *vtable, rt_size_t pv_off)
- {
- size_t l1_off, va_s, va_e;
- rt_base_t level;
- if ((!mmu_info) || (!vtable))
- {
- return -1;
- }
- va_s = (rt_size_t)v_address;
- va_e = ((rt_size_t)v_address) + size - 1;
- if (va_e < va_s)
- {
- return -1;
- }
- // convert address to PPN2 index
- va_s = GET_L1(va_s);
- va_e = GET_L1(va_e);
- if (va_s == 0)
- {
- return -1;
- }
- rt_mm_lock();
- // vtable initialization check
- for (l1_off = va_s; l1_off <= va_e; l1_off++)
- {
- size_t v = vtable[l1_off];
- if (v)
- {
- rt_mm_unlock();
- return -1;
- }
- }
- rt_mm_unlock();
- mmu_info->vtable = vtable;
- mmu_info->vstart = va_s;
- mmu_info->vend = va_e;
- mmu_info->pv_off = pv_off;
- return 0;
- }
- void rt_hw_mmu_kernel_map_init(rt_mmu_info *mmu_info, rt_size_t vaddr_start, rt_size_t size)
- {
- rt_size_t paddr_start = __UMASKVALUE(VPN_TO_PPN(vaddr_start, mmu_info->pv_off), PAGE_OFFSET_MASK);
- rt_size_t va_s = GET_L1(vaddr_start);
- rt_size_t va_e = GET_L1(vaddr_start + size - 1);
- rt_size_t i;
- for (i = va_s; i <= va_e; i++)
- {
- mmu_info->vtable[i] = COMBINEPTE(paddr_start, PAGE_ATTR_RWX | PTE_G | PTE_V);
- paddr_start += L1_PAGE_SIZE;
- }
- rt_hw_cpu_tlb_invalidate();
- }
- // find a range of free virtual address specified by pages
- static size_t find_vaddr(rt_mmu_info *mmu_info, int pages)
- {
- size_t loop_pages;
- size_t va;
- size_t find_va = 0;
- int n = 0;
- size_t i;
- if (!pages || !mmu_info)
- {
- return 0;
- }
- loop_pages = (mmu_info->vend - mmu_info->vstart) ? (mmu_info->vend - mmu_info->vstart) : 1;
- loop_pages <<= (ARCH_INDEX_WIDTH * 2);
- va = mmu_info->vstart;
- va <<= (ARCH_PAGE_SHIFT + ARCH_INDEX_WIDTH * 2);
- for (i = 0; i < loop_pages; i++, va += ARCH_PAGE_SIZE)
- {
- if (_rt_hw_mmu_v2p(mmu_info, (void *)va))
- {
- n = 0;
- find_va = 0;
- continue;
- }
- if (!find_va)
- {
- find_va = va;
- }
- n++;
- if (n >= pages)
- {
- return find_va;
- }
- }
- return 0;
- }
- // check whether the range of virtual address are free
- static int check_vaddr(rt_mmu_info *mmu_info, void *va, rt_size_t pages)
- {
- rt_size_t loop_va = __UMASKVALUE((rt_size_t)va, PAGE_OFFSET_MASK);
- rt_size_t l1_off, l2_off, l3_off;
- rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
- if (!pages)
- {
- return -1;
- }
- if (!mmu_info)
- {
- return -1;
- }
- while (pages--)
- {
- l1_off = GET_L1(loop_va);
- l2_off = GET_L2(loop_va);
- l3_off = GET_L3(loop_va);
- mmu_l1 = ((rt_size_t *)mmu_info->vtable) + l1_off;
- if (PTE_USED(*mmu_l1))
- {
- RT_ASSERT(!PAGE_IS_LEAF(*mmu_l1));
- mmu_l2 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l1), mmu_info->pv_off) + l2_off;
- if (PTE_USED(*mmu_l2))
- {
- RT_ASSERT(!PAGE_IS_LEAF(*mmu_l2));
- mmu_l3 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l2), mmu_info->pv_off) + l3_off;
- if (PTE_USED(*mmu_l3))
- {
- RT_ASSERT(PAGE_IS_LEAF(*mmu_l3));
- return -1;
- }
- }
- }
- loop_va += PAGE_SIZE;
- }
- return 0;
- }
- static void __rt_hw_mmu_unmap(rt_mmu_info *mmu_info, void *v_addr, rt_size_t npages)
- {
- rt_size_t loop_va = __UMASKVALUE((rt_size_t)v_addr, PAGE_OFFSET_MASK);
- rt_size_t l1_off, l2_off, l3_off;
- rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
- RT_ASSERT(mmu_info);
- while (npages--)
- {
- l1_off = (rt_size_t)GET_L1(loop_va);
- RT_ASSERT((l1_off >= mmu_info->vstart) && (l1_off <= mmu_info->vend));
- l2_off = (rt_size_t)GET_L2(loop_va);
- l3_off = (rt_size_t)GET_L3(loop_va);
- mmu_l1 = ((rt_size_t *)mmu_info->vtable) + l1_off;
- RT_ASSERT(PTE_USED(*mmu_l1))
- RT_ASSERT(!PAGE_IS_LEAF(*mmu_l1));
- mmu_l2 = ((rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l1), mmu_info->pv_off)) + l2_off;
- RT_ASSERT(PTE_USED(*mmu_l2));
- RT_ASSERT(!PAGE_IS_LEAF(*mmu_l2));
- mmu_l3 = ((rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l2), mmu_info->pv_off)) + l3_off;
- RT_ASSERT(PTE_USED(*mmu_l3));
- RT_ASSERT(PAGE_IS_LEAF(*(mmu_l3)));
- *mmu_l3 = 0;
- rt_hw_cpu_dcache_clean(mmu_l3, sizeof(*mmu_l3));
- // decrease reference from leaf page to l3 page
- mmu_l3 -= l3_off;
- rt_pages_free(mmu_l3, 0);
- int free = rt_page_ref_get(mmu_l3, 0);
- if (free == 1)
- {
- // free l3 page
- rt_pages_free(mmu_l3, 0);
- *mmu_l2 = 0;
- rt_hw_cpu_dcache_clean(mmu_l2, sizeof(*mmu_l2));
- // decrease reference from l3 page to l2 page
- mmu_l2 -= l2_off;
- rt_pages_free(mmu_l2, 0);
- free = rt_page_ref_get(mmu_l2, 0);
- if (free == 1)
- {
- // free l3 page
- rt_pages_free(mmu_l2, 0);
- // reset PTE in l1
- *mmu_l1 = 0;
- rt_hw_cpu_dcache_clean(mmu_l1, sizeof(*mmu_l1));
- }
- }
- loop_va += PAGE_SIZE;
- }
- }
- static int _mmu_map_one_page(rt_mmu_info *mmu_info, size_t va, size_t pa, size_t attr)
- {
- rt_size_t l1_off, l2_off, l3_off;
- rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
- l1_off = GET_L1(va);
- l2_off = GET_L2(va);
- l3_off = GET_L3(va);
- mmu_l1 = ((rt_size_t *)mmu_info->vtable) + l1_off;
- if (PTE_USED(*mmu_l1))
- {
- RT_ASSERT(!PAGE_IS_LEAF(*mmu_l1));
- mmu_l2 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l1), mmu_info->pv_off);
- }
- else
- {
- mmu_l2 = (rt_size_t *)rt_pages_alloc(0);
- if (mmu_l2)
- {
- rt_memset(mmu_l2, 0, PAGE_SIZE);
- rt_hw_cpu_dcache_clean(mmu_l2, PAGE_SIZE);
- *mmu_l1 = COMBINEPTE((rt_size_t)VPN_TO_PPN(mmu_l2, mmu_info->pv_off), PAGE_DEFAULT_ATTR_NEXT);
- rt_hw_cpu_dcache_clean(mmu_l1, sizeof(*mmu_l1));
- }
- else
- {
- return -1;
- }
- }
- if (PTE_USED(*(mmu_l2 + l2_off)))
- {
- RT_ASSERT(!PAGE_IS_LEAF(*(mmu_l2 + l2_off)));
- mmu_l3 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*(mmu_l2 + l2_off)), mmu_info->pv_off);
- }
- else
- {
- mmu_l3 = (rt_size_t *)rt_pages_alloc(0);
- if (mmu_l3)
- {
- rt_memset(mmu_l3, 0, PAGE_SIZE);
- rt_hw_cpu_dcache_clean(mmu_l3, PAGE_SIZE);
- *(mmu_l2 + l2_off) = COMBINEPTE((rt_size_t)VPN_TO_PPN(mmu_l3, mmu_info->pv_off), PAGE_DEFAULT_ATTR_NEXT);
- rt_hw_cpu_dcache_clean(mmu_l2, sizeof(*mmu_l2));
- // declares a reference to parent page table
- rt_page_ref_inc((void *)mmu_l2, 0);
- }
- else
- {
- return -1;
- }
- }
- RT_ASSERT(!PTE_USED(*(mmu_l3 + l3_off)));
- // declares a reference to parent page table
- rt_page_ref_inc((void *)mmu_l3, 0);
- *(mmu_l3 + l3_off) = COMBINEPTE((rt_size_t)pa, attr);
- rt_hw_cpu_dcache_clean(mmu_l3 + l3_off, sizeof(*(mmu_l3 + l3_off)));
- return 0;
- }
- static int __rt_hw_mmu_map(rt_mmu_info *mmu_info, void *v_addr, void *p_addr, rt_size_t npages, rt_size_t attr)
- {
- rt_size_t loop_va = __UMASKVALUE((rt_size_t)v_addr, PAGE_OFFSET_MASK);
- rt_size_t loop_pa = __UMASKVALUE((rt_size_t)p_addr, PAGE_OFFSET_MASK);
- if (!mmu_info)
- {
- return -1;
- }
- while (npages--)
- {
- if (_mmu_map_one_page(mmu_info, loop_va, loop_pa, attr) != 0)
- {
- __rt_hw_mmu_unmap(mmu_info, v_addr, npages);
- return -1;
- }
- loop_va += PAGE_SIZE;
- loop_pa += PAGE_SIZE;
- }
- return 0;
- }
- void *_rt_hw_mmu_map(rt_mmu_info *mmu_info, void *v_addr, void *p_addr, rt_size_t size, rt_size_t attr)
- {
- rt_size_t pa_s, pa_e;
- rt_size_t vaddr;
- rt_size_t pages;
- int ret;
- if (!size)
- {
- return 0;
- }
- pa_s = (rt_size_t)p_addr;
- pa_e = ((rt_size_t)p_addr) + size - 1;
- pa_s = GET_PF_ID(pa_s);
- pa_e = GET_PF_ID(pa_e);
- pages = pa_e - pa_s + 1;
- if (v_addr)
- {
- vaddr = (rt_size_t)v_addr;
- pa_s = (rt_size_t)p_addr;
- if (GET_PF_OFFSET(vaddr) != GET_PF_OFFSET(pa_s))
- {
- return 0;
- }
- vaddr = __UMASKVALUE(vaddr, PAGE_OFFSET_MASK);
- if (check_vaddr(mmu_info, (void *)vaddr, pages) != 0)
- {
- return 0;
- }
- }
- else
- {
- vaddr = find_vaddr(mmu_info, pages);
- }
- if (vaddr)
- {
- ret = __rt_hw_mmu_map(mmu_info, (void *)vaddr, p_addr, pages, attr);
- if (ret == 0)
- {
- rt_hw_cpu_tlb_invalidate();
- return (void *)(vaddr | GET_PF_OFFSET((rt_size_t)p_addr));
- }
- }
- return 0;
- }
- static int __rt_hw_mmu_map_auto(rt_mmu_info *mmu_info, void *v_addr, rt_size_t npages, rt_size_t attr)
- {
- rt_size_t loop_va = __UMASKVALUE((rt_size_t)v_addr, PAGE_OFFSET_MASK);
- rt_size_t loop_pa;
- rt_size_t l1_off, l2_off, l3_off;
- rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
- rt_size_t *ref_cnt;
- rt_size_t i;
- void *va, *pa;
- if (!mmu_info)
- {
- return -1;
- }
- while (npages--)
- {
- loop_pa = (rt_size_t)rt_pages_alloc(0);
- if (!loop_pa)
- {
- goto err;
- }
- if (__rt_hw_mmu_map(mmu_info, (void *)loop_va, (void *)loop_pa, 1, attr) < 0)
- {
- goto err;
- }
- loop_va += PAGE_SIZE;
- }
- return 0;
- err:
- va = (void *)__UMASKVALUE((rt_size_t)v_addr, PAGE_OFFSET_MASK);
- for (i = 0; i < npages; i++)
- {
- pa = rt_hw_mmu_v2p(mmu_info, va);
- if (pa)
- {
- rt_pages_free((void *)PPN_TO_VPN(pa, mmu_info->pv_off), 0);
- }
- va = (void *)((rt_uint8_t *)va + PAGE_SIZE);
- }
- __rt_hw_mmu_unmap(mmu_info, v_addr, npages);
- return -1;
- }
- void *_rt_hw_mmu_map_auto(rt_mmu_info *mmu_info, void *v_addr, rt_size_t size, rt_size_t attr)
- {
- rt_size_t vaddr;
- rt_size_t offset;
- rt_size_t pages;
- int ret;
- if (!size)
- {
- return 0;
- }
- offset = GET_PF_OFFSET((rt_size_t)v_addr);
- size += (offset + PAGE_SIZE - 1);
- pages = size >> PAGE_OFFSET_BIT;
- if (v_addr)
- {
- vaddr = __UMASKVALUE((rt_size_t)v_addr, PAGE_OFFSET_MASK);
- if (check_vaddr(mmu_info, (void *)vaddr, pages) != 0)
- {
- return 0;
- }
- }
- else
- {
- vaddr = find_vaddr(mmu_info, pages);
- }
- if (vaddr)
- {
- ret = __rt_hw_mmu_map_auto(mmu_info, (void *)vaddr, pages, attr);
- if (ret == 0)
- {
- rt_hw_cpu_tlb_invalidate();
- return (void *)(vaddr | offset);
- }
- }
- return 0;
- }
- void _rt_hw_mmu_unmap(rt_mmu_info *mmu_info, void *v_addr, rt_size_t size)
- {
- rt_size_t va_s, va_e;
- rt_size_t pages;
- va_s = ((rt_size_t)v_addr) >> PAGE_OFFSET_BIT;
- va_e = (((rt_size_t)v_addr) + size - 1) >> PAGE_OFFSET_BIT;
- pages = va_e - va_s + 1;
- __rt_hw_mmu_unmap(mmu_info, v_addr, pages);
- rt_hw_cpu_tlb_invalidate();
- }
- void *rt_hw_mmu_map(rt_mmu_info *mmu_info, void *v_addr, void *p_addr, rt_size_t size, rt_size_t attr)
- {
- void *ret;
- rt_base_t level;
- rt_mm_lock();
- ret = _rt_hw_mmu_map(mmu_info, v_addr, p_addr, size, attr);
- rt_mm_unlock();
- return ret;
- }
- void *rt_hw_mmu_map_auto(rt_mmu_info *mmu_info, void *v_addr, rt_size_t size, rt_size_t attr)
- {
- void *ret;
- rt_base_t level;
- rt_mm_lock();
- ret = _rt_hw_mmu_map_auto(mmu_info, v_addr, size, attr);
- rt_mm_unlock();
- return ret;
- }
- void rt_hw_mmu_unmap(rt_mmu_info *mmu_info, void *v_addr, rt_size_t size)
- {
- rt_base_t level;
- rt_mm_lock();
- _rt_hw_mmu_unmap(mmu_info, v_addr, size);
- rt_mm_unlock();
- }
- void *_rt_hw_mmu_v2p(rt_mmu_info *mmu_info, void *v_addr)
- {
- rt_size_t l1_off, l2_off, l3_off;
- rt_size_t *mmu_l1, *mmu_l2, *mmu_l3;
- rt_size_t pa;
- l1_off = GET_L1((rt_size_t)v_addr);
- l2_off = GET_L2((rt_size_t)v_addr);
- l3_off = GET_L3((rt_size_t)v_addr);
- if (!mmu_info)
- {
- return RT_NULL;
- }
- mmu_l1 = ((rt_size_t *)mmu_info->vtable) + l1_off;
- if (PTE_USED(*mmu_l1))
- {
- if (*mmu_l1 & PTE_XWR_MASK)
- return (void *)(GET_PADDR(*mmu_l1) | ((rt_size_t)v_addr & ((1 << 30) - 1)));
- mmu_l2 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*mmu_l1), mmu_info->pv_off);
- if (PTE_USED(*(mmu_l2 + l2_off)))
- {
- if (*(mmu_l2 + l2_off) & PTE_XWR_MASK)
- return (void *)(GET_PADDR(*(mmu_l2 + l2_off)) | ((rt_size_t)v_addr & ((1 << 21) - 1)));
- mmu_l3 = (rt_size_t *)PPN_TO_VPN(GET_PADDR(*(mmu_l2 + l2_off)), mmu_info->pv_off);
- if (PTE_USED(*(mmu_l3 + l3_off)))
- {
- return (void *)(GET_PADDR(*(mmu_l3 + l3_off)) | GET_PF_OFFSET((rt_size_t)v_addr));
- }
- }
- }
- return RT_NULL;
- }
- void *rt_hw_mmu_v2p(rt_mmu_info *mmu_info, void *v_addr)
- {
- void *ret;
- rt_base_t level;
- rt_mm_lock();
- ret = _rt_hw_mmu_v2p(mmu_info, v_addr);
- rt_mm_unlock();
- return ret;
- }
- /**
- * @brief setup Page Table for kernel space. It's a fixed map
- * and all mappings cannot be changed after initialization.
- *
- * Memory region in struct mem_desc must be page aligned,
- * otherwise is a failure and no report will be
- * returned.
- *
- * @param mmu_info
- * @param mdesc
- * @param desc_nr
- */
- void rt_hw_mmu_setup(rt_mmu_info *mmu_info, struct mem_desc *mdesc, int desc_nr)
- {
- void *err;
- for (size_t i = 0; i < desc_nr; i++)
- {
- size_t attr;
- switch (mdesc->attr)
- {
- case NORMAL_MEM:
- attr = MMU_MAP_K_RWCB;
- break;
- case NORMAL_NOCACHE_MEM:
- attr = MMU_MAP_K_RWCB;
- break;
- case DEVICE_MEM:
- attr = MMU_MAP_K_DEVICE;
- break;
- default:
- attr = MMU_MAP_K_DEVICE;
- }
- err = _rt_hw_mmu_map(mmu_info, (void *)mdesc->vaddr_start, (void *)mdesc->paddr_start,
- mdesc->vaddr_end - mdesc->vaddr_start + 1, attr);
- mdesc++;
- }
- rt_hw_mmu_switch((void *)MMUTable);
- }
- #endif /* RT_USING_SMART */
|