| /* SPDX-License-Identifier: GPL-2.0-or-later */ |
| /* |
| * QEMU LoongArch TLB helpers |
| * |
| * Copyright (c) 2021 Loongson Technology Corporation Limited |
| * |
| */ |
| |
| #include "qemu/osdep.h" |
| #include "qemu/guest-random.h" |
| |
| #include "cpu.h" |
| #include "internals.h" |
| #include "exec/helper-proto.h" |
| #include "exec/exec-all.h" |
| #include "exec/cpu_ldst.h" |
| #include "exec/log.h" |
| #include "cpu-csr.h" |
| |
| enum { |
| TLBRET_MATCH = 0, |
| TLBRET_BADADDR = 1, |
| TLBRET_NOMATCH = 2, |
| TLBRET_INVALID = 3, |
| TLBRET_DIRTY = 4, |
| TLBRET_RI = 5, |
| TLBRET_XI = 6, |
| TLBRET_PE = 7, |
| }; |
| |
| static int loongarch_map_tlb_entry(CPULoongArchState *env, hwaddr *physical, |
| int *prot, target_ulong address, |
| int access_type, int index, int mmu_idx) |
| { |
| LoongArchTLB *tlb = &env->tlb[index]; |
| uint64_t plv = mmu_idx; |
| uint64_t tlb_entry, tlb_ppn; |
| uint8_t tlb_ps, n, tlb_v, tlb_d, tlb_plv, tlb_nx, tlb_nr, tlb_rplv; |
| |
| if (index >= LOONGARCH_STLB) { |
| tlb_ps = FIELD_EX64(tlb->tlb_misc, TLB_MISC, PS); |
| } else { |
| tlb_ps = FIELD_EX64(env->CSR_STLBPS, CSR_STLBPS, PS); |
| } |
| n = (address >> tlb_ps) & 0x1;/* Odd or even */ |
| |
| tlb_entry = n ? tlb->tlb_entry1 : tlb->tlb_entry0; |
| tlb_v = FIELD_EX64(tlb_entry, TLBENTRY, V); |
| tlb_d = FIELD_EX64(tlb_entry, TLBENTRY, D); |
| tlb_plv = FIELD_EX64(tlb_entry, TLBENTRY, PLV); |
| if (is_la64(env)) { |
| tlb_ppn = FIELD_EX64(tlb_entry, TLBENTRY_64, PPN); |
| tlb_nx = FIELD_EX64(tlb_entry, TLBENTRY_64, NX); |
| tlb_nr = FIELD_EX64(tlb_entry, TLBENTRY_64, NR); |
| tlb_rplv = FIELD_EX64(tlb_entry, TLBENTRY_64, RPLV); |
| } else { |
| tlb_ppn = FIELD_EX64(tlb_entry, TLBENTRY_32, PPN); |
| tlb_nx = 0; |
| tlb_nr = 0; |
| tlb_rplv = 0; |
| } |
| |
| /* Remove sw bit between bit12 -- bit PS*/ |
| tlb_ppn = tlb_ppn & ~(((0x1UL << (tlb_ps - 12)) -1)); |
| |
| /* Check access rights */ |
| if (!tlb_v) { |
| return TLBRET_INVALID; |
| } |
| |
| if (access_type == MMU_INST_FETCH && tlb_nx) { |
| return TLBRET_XI; |
| } |
| |
| if (access_type == MMU_DATA_LOAD && tlb_nr) { |
| return TLBRET_RI; |
| } |
| |
| if (((tlb_rplv == 0) && (plv > tlb_plv)) || |
| ((tlb_rplv == 1) && (plv != tlb_plv))) { |
| return TLBRET_PE; |
| } |
| |
| if ((access_type == MMU_DATA_STORE) && !tlb_d) { |
| return TLBRET_DIRTY; |
| } |
| |
| *physical = (tlb_ppn << R_TLBENTRY_64_PPN_SHIFT) | |
| (address & MAKE_64BIT_MASK(0, tlb_ps)); |
| *prot = PAGE_READ; |
| if (tlb_d) { |
| *prot |= PAGE_WRITE; |
| } |
| if (!tlb_nx) { |
| *prot |= PAGE_EXEC; |
| } |
| return TLBRET_MATCH; |
| } |
| |
| /* |
| * One tlb entry holds an adjacent odd/even pair, the vpn is the |
| * content of the virtual page number divided by 2. So the |
| * compare vpn is bit[47:15] for 16KiB page. while the vppn |
| * field in tlb entry contains bit[47:13], so need adjust. |
| * virt_vpn = vaddr[47:13] |
| */ |
| static bool loongarch_tlb_search(CPULoongArchState *env, target_ulong vaddr, |
| int *index) |
| { |
| LoongArchTLB *tlb; |
| uint16_t csr_asid, tlb_asid, stlb_idx; |
| uint8_t tlb_e, tlb_ps, tlb_g, stlb_ps; |
| int i, compare_shift; |
| uint64_t vpn, tlb_vppn; |
| |
| csr_asid = FIELD_EX64(env->CSR_ASID, CSR_ASID, ASID); |
| stlb_ps = FIELD_EX64(env->CSR_STLBPS, CSR_STLBPS, PS); |
| vpn = (vaddr & TARGET_VIRT_MASK) >> (stlb_ps + 1); |
| stlb_idx = vpn & 0xff; /* VA[25:15] <==> TLBIDX.index for 16KiB Page */ |
| compare_shift = stlb_ps + 1 - R_TLB_MISC_VPPN_SHIFT; |
| |
| /* Search STLB */ |
| for (i = 0; i < 8; ++i) { |
| tlb = &env->tlb[i * 256 + stlb_idx]; |
| tlb_e = FIELD_EX64(tlb->tlb_misc, TLB_MISC, E); |
| if (tlb_e) { |
| tlb_vppn = FIELD_EX64(tlb->tlb_misc, TLB_MISC, VPPN); |
| tlb_asid = FIELD_EX64(tlb->tlb_misc, TLB_MISC, ASID); |
| tlb_g = FIELD_EX64(tlb->tlb_entry0, TLBENTRY, G); |
| |
| if ((tlb_g == 1 || tlb_asid == csr_asid) && |
| (vpn == (tlb_vppn >> compare_shift))) { |
| *index = i * 256 + stlb_idx; |
| return true; |
| } |
| } |
| } |
| |
| /* Search MTLB */ |
| for (i = LOONGARCH_STLB; i < LOONGARCH_TLB_MAX; ++i) { |
| tlb = &env->tlb[i]; |
| tlb_e = FIELD_EX64(tlb->tlb_misc, TLB_MISC, E); |
| if (tlb_e) { |
| tlb_vppn = FIELD_EX64(tlb->tlb_misc, TLB_MISC, VPPN); |
| tlb_ps = FIELD_EX64(tlb->tlb_misc, TLB_MISC, PS); |
| tlb_asid = FIELD_EX64(tlb->tlb_misc, TLB_MISC, ASID); |
| tlb_g = FIELD_EX64(tlb->tlb_entry0, TLBENTRY, G); |
| compare_shift = tlb_ps + 1 - R_TLB_MISC_VPPN_SHIFT; |
| vpn = (vaddr & TARGET_VIRT_MASK) >> (tlb_ps + 1); |
| if ((tlb_g == 1 || tlb_asid == csr_asid) && |
| (vpn == (tlb_vppn >> compare_shift))) { |
| *index = i; |
| return true; |
| } |
| } |
| } |
| return false; |
| } |
| |
| static int loongarch_map_address(CPULoongArchState *env, hwaddr *physical, |
| int *prot, target_ulong address, |
| MMUAccessType access_type, int mmu_idx) |
| { |
| int index, match; |
| |
| match = loongarch_tlb_search(env, address, &index); |
| if (match) { |
| return loongarch_map_tlb_entry(env, physical, prot, |
| address, access_type, index, mmu_idx); |
| } |
| |
| return TLBRET_NOMATCH; |
| } |
| |
| static hwaddr dmw_va2pa(CPULoongArchState *env, target_ulong va, |
| target_ulong dmw) |
| { |
| if (is_la64(env)) { |
| return va & TARGET_VIRT_MASK; |
| } else { |
| uint32_t pseg = FIELD_EX32(dmw, CSR_DMW_32, PSEG); |
| return (va & MAKE_64BIT_MASK(0, R_CSR_DMW_32_VSEG_SHIFT)) | \ |
| (pseg << R_CSR_DMW_32_VSEG_SHIFT); |
| } |
| } |
| |
| static int get_physical_address(CPULoongArchState *env, hwaddr *physical, |
| int *prot, target_ulong address, |
| MMUAccessType access_type, int mmu_idx) |
| { |
| int user_mode = mmu_idx == MMU_IDX_USER; |
| int kernel_mode = mmu_idx == MMU_IDX_KERNEL; |
| uint32_t plv, base_c, base_v; |
| int64_t addr_high; |
| uint8_t da = FIELD_EX64(env->CSR_CRMD, CSR_CRMD, DA); |
| uint8_t pg = FIELD_EX64(env->CSR_CRMD, CSR_CRMD, PG); |
| |
| /* Check PG and DA */ |
| if (da & !pg) { |
| *physical = address & TARGET_PHYS_MASK; |
| *prot = PAGE_READ | PAGE_WRITE | PAGE_EXEC; |
| return TLBRET_MATCH; |
| } |
| |
| plv = kernel_mode | (user_mode << R_CSR_DMW_PLV3_SHIFT); |
| if (is_la64(env)) { |
| base_v = address >> R_CSR_DMW_64_VSEG_SHIFT; |
| } else { |
| base_v = address >> R_CSR_DMW_32_VSEG_SHIFT; |
| } |
| /* Check direct map window */ |
| for (int i = 0; i < 4; i++) { |
| if (is_la64(env)) { |
| base_c = FIELD_EX64(env->CSR_DMW[i], CSR_DMW_64, VSEG); |
| } else { |
| base_c = FIELD_EX64(env->CSR_DMW[i], CSR_DMW_32, VSEG); |
| } |
| if ((plv & env->CSR_DMW[i]) && (base_c == base_v)) { |
| *physical = dmw_va2pa(env, address, env->CSR_DMW[i]); |
| *prot = PAGE_READ | PAGE_WRITE | PAGE_EXEC; |
| return TLBRET_MATCH; |
| } |
| } |
| |
| /* Check valid extension */ |
| addr_high = sextract64(address, TARGET_VIRT_ADDR_SPACE_BITS, 16); |
| if (!(addr_high == 0 || addr_high == -1)) { |
| return TLBRET_BADADDR; |
| } |
| |
| /* Mapped address */ |
| return loongarch_map_address(env, physical, prot, address, |
| access_type, mmu_idx); |
| } |
| |
| hwaddr loongarch_cpu_get_phys_page_debug(CPUState *cs, vaddr addr) |
| { |
| LoongArchCPU *cpu = LOONGARCH_CPU(cs); |
| CPULoongArchState *env = &cpu->env; |
| hwaddr phys_addr; |
| int prot; |
| |
| if (get_physical_address(env, &phys_addr, &prot, addr, MMU_DATA_LOAD, |
| cpu_mmu_index(env, false)) != 0) { |
| return -1; |
| } |
| return phys_addr; |
| } |
| |
| static void raise_mmu_exception(CPULoongArchState *env, target_ulong address, |
| MMUAccessType access_type, int tlb_error) |
| { |
| CPUState *cs = env_cpu(env); |
| |
| switch (tlb_error) { |
| default: |
| case TLBRET_BADADDR: |
| cs->exception_index = access_type == MMU_INST_FETCH |
| ? EXCCODE_ADEF : EXCCODE_ADEM; |
| break; |
| case TLBRET_NOMATCH: |
| /* No TLB match for a mapped address */ |
| if (access_type == MMU_DATA_LOAD) { |
| cs->exception_index = EXCCODE_PIL; |
| } else if (access_type == MMU_DATA_STORE) { |
| cs->exception_index = EXCCODE_PIS; |
| } else if (access_type == MMU_INST_FETCH) { |
| cs->exception_index = EXCCODE_PIF; |
| } |
| env->CSR_TLBRERA = FIELD_DP64(env->CSR_TLBRERA, CSR_TLBRERA, ISTLBR, 1); |
| break; |
| case TLBRET_INVALID: |
| /* TLB match with no valid bit */ |
| if (access_type == MMU_DATA_LOAD) { |
| cs->exception_index = EXCCODE_PIL; |
| } else if (access_type == MMU_DATA_STORE) { |
| cs->exception_index = EXCCODE_PIS; |
| } else if (access_type == MMU_INST_FETCH) { |
| cs->exception_index = EXCCODE_PIF; |
| } |
| break; |
| case TLBRET_DIRTY: |
| /* TLB match but 'D' bit is cleared */ |
| cs->exception_index = EXCCODE_PME; |
| break; |
| case TLBRET_XI: |
| /* Execute-Inhibit Exception */ |
| cs->exception_index = EXCCODE_PNX; |
| break; |
| case TLBRET_RI: |
| /* Read-Inhibit Exception */ |
| cs->exception_index = EXCCODE_PNR; |
| break; |
| case TLBRET_PE: |
| /* Privileged Exception */ |
| cs->exception_index = EXCCODE_PPI; |
| break; |
| } |
| |
| if (tlb_error == TLBRET_NOMATCH) { |
| env->CSR_TLBRBADV = address; |
| if (is_la64(env)) { |
| env->CSR_TLBREHI = FIELD_DP64(env->CSR_TLBREHI, CSR_TLBREHI_64, |
| VPPN, extract64(address, 13, 35)); |
| } else { |
| env->CSR_TLBREHI = FIELD_DP64(env->CSR_TLBREHI, CSR_TLBREHI_32, |
| VPPN, extract64(address, 13, 19)); |
| } |
| } else { |
| if (!FIELD_EX64(env->CSR_DBG, CSR_DBG, DST)) { |
| env->CSR_BADV = address; |
| } |
| env->CSR_TLBEHI = address & (TARGET_PAGE_MASK << 1); |
| } |
| } |
| |
| static void invalidate_tlb_entry(CPULoongArchState *env, int index) |
| { |
| target_ulong addr, mask, pagesize; |
| uint8_t tlb_ps; |
| LoongArchTLB *tlb = &env->tlb[index]; |
| |
| int mmu_idx = cpu_mmu_index(env, false); |
| uint8_t tlb_v0 = FIELD_EX64(tlb->tlb_entry0, TLBENTRY, V); |
| uint8_t tlb_v1 = FIELD_EX64(tlb->tlb_entry1, TLBENTRY, V); |
| uint64_t tlb_vppn = FIELD_EX64(tlb->tlb_misc, TLB_MISC, VPPN); |
| |
| if (index >= LOONGARCH_STLB) { |
| tlb_ps = FIELD_EX64(tlb->tlb_misc, TLB_MISC, PS); |
| } else { |
| tlb_ps = FIELD_EX64(env->CSR_STLBPS, CSR_STLBPS, PS); |
| } |
| pagesize = MAKE_64BIT_MASK(tlb_ps, 1); |
| mask = MAKE_64BIT_MASK(0, tlb_ps + 1); |
| |
| if (tlb_v0) { |
| addr = (tlb_vppn << R_TLB_MISC_VPPN_SHIFT) & ~mask; /* even */ |
| tlb_flush_range_by_mmuidx(env_cpu(env), addr, pagesize, |
| mmu_idx, TARGET_LONG_BITS); |
| } |
| |
| if (tlb_v1) { |
| addr = (tlb_vppn << R_TLB_MISC_VPPN_SHIFT) & pagesize; /* odd */ |
| tlb_flush_range_by_mmuidx(env_cpu(env), addr, pagesize, |
| mmu_idx, TARGET_LONG_BITS); |
| } |
| } |
| |
| static void invalidate_tlb(CPULoongArchState *env, int index) |
| { |
| LoongArchTLB *tlb; |
| uint16_t csr_asid, tlb_asid, tlb_g; |
| |
| csr_asid = FIELD_EX64(env->CSR_ASID, CSR_ASID, ASID); |
| tlb = &env->tlb[index]; |
| tlb_asid = FIELD_EX64(tlb->tlb_misc, TLB_MISC, ASID); |
| tlb_g = FIELD_EX64(tlb->tlb_entry0, TLBENTRY, G); |
| if (tlb_g == 0 && tlb_asid != csr_asid) { |
| return; |
| } |
| invalidate_tlb_entry(env, index); |
| } |
| |
| static void fill_tlb_entry(CPULoongArchState *env, int index) |
| { |
| LoongArchTLB *tlb = &env->tlb[index]; |
| uint64_t lo0, lo1, csr_vppn; |
| uint16_t csr_asid; |
| uint8_t csr_ps; |
| |
| if (FIELD_EX64(env->CSR_TLBRERA, CSR_TLBRERA, ISTLBR)) { |
| csr_ps = FIELD_EX64(env->CSR_TLBREHI, CSR_TLBREHI, PS); |
| if (is_la64(env)) { |
| csr_vppn = FIELD_EX64(env->CSR_TLBREHI, CSR_TLBREHI_64, VPPN); |
| } else { |
| csr_vppn = FIELD_EX64(env->CSR_TLBREHI, CSR_TLBREHI_32, VPPN); |
| } |
| lo0 = env->CSR_TLBRELO0; |
| lo1 = env->CSR_TLBRELO1; |
| } else { |
| csr_ps = FIELD_EX64(env->CSR_TLBIDX, CSR_TLBIDX, PS); |
| if (is_la64(env)) { |
| csr_vppn = FIELD_EX64(env->CSR_TLBEHI, CSR_TLBEHI_64, VPPN); |
| } else { |
| csr_vppn = FIELD_EX64(env->CSR_TLBEHI, CSR_TLBEHI_32, VPPN); |
| } |
| lo0 = env->CSR_TLBELO0; |
| lo1 = env->CSR_TLBELO1; |
| } |
| |
| if (csr_ps == 0) { |
| qemu_log_mask(CPU_LOG_MMU, "page size is 0\n"); |
| } |
| |
| /* Only MTLB has the ps fields */ |
| if (index >= LOONGARCH_STLB) { |
| tlb->tlb_misc = FIELD_DP64(tlb->tlb_misc, TLB_MISC, PS, csr_ps); |
| } |
| |
| tlb->tlb_misc = FIELD_DP64(tlb->tlb_misc, TLB_MISC, VPPN, csr_vppn); |
| tlb->tlb_misc = FIELD_DP64(tlb->tlb_misc, TLB_MISC, E, 1); |
| csr_asid = FIELD_EX64(env->CSR_ASID, CSR_ASID, ASID); |
| tlb->tlb_misc = FIELD_DP64(tlb->tlb_misc, TLB_MISC, ASID, csr_asid); |
| |
| tlb->tlb_entry0 = lo0; |
| tlb->tlb_entry1 = lo1; |
| } |
| |
| /* Return an random value between low and high */ |
| static uint32_t get_random_tlb(uint32_t low, uint32_t high) |
| { |
| uint32_t val; |
| |
| qemu_guest_getrandom_nofail(&val, sizeof(val)); |
| return val % (high - low + 1) + low; |
| } |
| |
| void helper_tlbsrch(CPULoongArchState *env) |
| { |
| int index, match; |
| |
| if (FIELD_EX64(env->CSR_TLBRERA, CSR_TLBRERA, ISTLBR)) { |
| match = loongarch_tlb_search(env, env->CSR_TLBREHI, &index); |
| } else { |
| match = loongarch_tlb_search(env, env->CSR_TLBEHI, &index); |
| } |
| |
| if (match) { |
| env->CSR_TLBIDX = FIELD_DP64(env->CSR_TLBIDX, CSR_TLBIDX, INDEX, index); |
| env->CSR_TLBIDX = FIELD_DP64(env->CSR_TLBIDX, CSR_TLBIDX, NE, 0); |
| return; |
| } |
| |
| env->CSR_TLBIDX = FIELD_DP64(env->CSR_TLBIDX, CSR_TLBIDX, NE, 1); |
| } |
| |
| void helper_tlbrd(CPULoongArchState *env) |
| { |
| LoongArchTLB *tlb; |
| int index; |
| uint8_t tlb_ps, tlb_e; |
| |
| index = FIELD_EX64(env->CSR_TLBIDX, CSR_TLBIDX, INDEX); |
| tlb = &env->tlb[index]; |
| |
| if (index >= LOONGARCH_STLB) { |
| tlb_ps = FIELD_EX64(tlb->tlb_misc, TLB_MISC, PS); |
| } else { |
| tlb_ps = FIELD_EX64(env->CSR_STLBPS, CSR_STLBPS, PS); |
| } |
| tlb_e = FIELD_EX64(tlb->tlb_misc, TLB_MISC, E); |
| |
| if (!tlb_e) { |
| /* Invalid TLB entry */ |
| env->CSR_TLBIDX = FIELD_DP64(env->CSR_TLBIDX, CSR_TLBIDX, NE, 1); |
| env->CSR_ASID = FIELD_DP64(env->CSR_ASID, CSR_ASID, ASID, 0); |
| env->CSR_TLBEHI = 0; |
| env->CSR_TLBELO0 = 0; |
| env->CSR_TLBELO1 = 0; |
| env->CSR_TLBIDX = FIELD_DP64(env->CSR_TLBIDX, CSR_TLBIDX, PS, 0); |
| } else { |
| /* Valid TLB entry */ |
| env->CSR_TLBIDX = FIELD_DP64(env->CSR_TLBIDX, CSR_TLBIDX, NE, 0); |
| env->CSR_TLBIDX = FIELD_DP64(env->CSR_TLBIDX, CSR_TLBIDX, |
| PS, (tlb_ps & 0x3f)); |
| env->CSR_TLBEHI = FIELD_EX64(tlb->tlb_misc, TLB_MISC, VPPN) << |
| R_TLB_MISC_VPPN_SHIFT; |
| env->CSR_TLBELO0 = tlb->tlb_entry0; |
| env->CSR_TLBELO1 = tlb->tlb_entry1; |
| } |
| } |
| |
| void helper_tlbwr(CPULoongArchState *env) |
| { |
| int index = FIELD_EX64(env->CSR_TLBIDX, CSR_TLBIDX, INDEX); |
| |
| invalidate_tlb(env, index); |
| |
| if (FIELD_EX64(env->CSR_TLBIDX, CSR_TLBIDX, NE)) { |
| env->tlb[index].tlb_misc = FIELD_DP64(env->tlb[index].tlb_misc, |
| TLB_MISC, E, 0); |
| return; |
| } |
| |
| fill_tlb_entry(env, index); |
| } |
| |
| void helper_tlbfill(CPULoongArchState *env) |
| { |
| uint64_t address, entryhi; |
| int index, set, stlb_idx; |
| uint16_t pagesize, stlb_ps; |
| |
| if (FIELD_EX64(env->CSR_TLBRERA, CSR_TLBRERA, ISTLBR)) { |
| entryhi = env->CSR_TLBREHI; |
| pagesize = FIELD_EX64(env->CSR_TLBREHI, CSR_TLBREHI, PS); |
| } else { |
| entryhi = env->CSR_TLBEHI; |
| pagesize = FIELD_EX64(env->CSR_TLBIDX, CSR_TLBIDX, PS); |
| } |
| |
| stlb_ps = FIELD_EX64(env->CSR_STLBPS, CSR_STLBPS, PS); |
| |
| if (pagesize == stlb_ps) { |
| /* Only write into STLB bits [47:13] */ |
| address = entryhi & ~MAKE_64BIT_MASK(0, R_CSR_TLBEHI_64_VPPN_SHIFT); |
| |
| /* Choose one set ramdomly */ |
| set = get_random_tlb(0, 7); |
| |
| /* Index in one set */ |
| stlb_idx = (address >> (stlb_ps + 1)) & 0xff; /* [0,255] */ |
| |
| index = set * 256 + stlb_idx; |
| } else { |
| /* Only write into MTLB */ |
| index = get_random_tlb(LOONGARCH_STLB, LOONGARCH_TLB_MAX - 1); |
| } |
| |
| invalidate_tlb(env, index); |
| fill_tlb_entry(env, index); |
| } |
| |
| void helper_tlbclr(CPULoongArchState *env) |
| { |
| LoongArchTLB *tlb; |
| int i, index; |
| uint16_t csr_asid, tlb_asid, tlb_g; |
| |
| csr_asid = FIELD_EX64(env->CSR_ASID, CSR_ASID, ASID); |
| index = FIELD_EX64(env->CSR_TLBIDX, CSR_TLBIDX, INDEX); |
| |
| if (index < LOONGARCH_STLB) { |
| /* STLB. One line per operation */ |
| for (i = 0; i < 8; i++) { |
| tlb = &env->tlb[i * 256 + (index % 256)]; |
| tlb_asid = FIELD_EX64(tlb->tlb_misc, TLB_MISC, ASID); |
| tlb_g = FIELD_EX64(tlb->tlb_entry0, TLBENTRY, G); |
| if (!tlb_g && tlb_asid == csr_asid) { |
| tlb->tlb_misc = FIELD_DP64(tlb->tlb_misc, TLB_MISC, E, 0); |
| } |
| } |
| } else if (index < LOONGARCH_TLB_MAX) { |
| /* All MTLB entries */ |
| for (i = LOONGARCH_STLB; i < LOONGARCH_TLB_MAX; i++) { |
| tlb = &env->tlb[i]; |
| tlb_asid = FIELD_EX64(tlb->tlb_misc, TLB_MISC, ASID); |
| tlb_g = FIELD_EX64(tlb->tlb_entry0, TLBENTRY, G); |
| if (!tlb_g && tlb_asid == csr_asid) { |
| tlb->tlb_misc = FIELD_DP64(tlb->tlb_misc, TLB_MISC, E, 0); |
| } |
| } |
| } |
| |
| tlb_flush(env_cpu(env)); |
| } |
| |
| void helper_tlbflush(CPULoongArchState *env) |
| { |
| int i, index; |
| |
| index = FIELD_EX64(env->CSR_TLBIDX, CSR_TLBIDX, INDEX); |
| |
| if (index < LOONGARCH_STLB) { |
| /* STLB. One line per operation */ |
| for (i = 0; i < 8; i++) { |
| int s_idx = i * 256 + (index % 256); |
| env->tlb[s_idx].tlb_misc = FIELD_DP64(env->tlb[s_idx].tlb_misc, |
| TLB_MISC, E, 0); |
| } |
| } else if (index < LOONGARCH_TLB_MAX) { |
| /* All MTLB entries */ |
| for (i = LOONGARCH_STLB; i < LOONGARCH_TLB_MAX; i++) { |
| env->tlb[i].tlb_misc = FIELD_DP64(env->tlb[i].tlb_misc, |
| TLB_MISC, E, 0); |
| } |
| } |
| |
| tlb_flush(env_cpu(env)); |
| } |
| |
| void helper_invtlb_all(CPULoongArchState *env) |
| { |
| for (int i = 0; i < LOONGARCH_TLB_MAX; i++) { |
| env->tlb[i].tlb_misc = FIELD_DP64(env->tlb[i].tlb_misc, |
| TLB_MISC, E, 0); |
| } |
| tlb_flush(env_cpu(env)); |
| } |
| |
| void helper_invtlb_all_g(CPULoongArchState *env, uint32_t g) |
| { |
| for (int i = 0; i < LOONGARCH_TLB_MAX; i++) { |
| LoongArchTLB *tlb = &env->tlb[i]; |
| uint8_t tlb_g = FIELD_EX64(tlb->tlb_entry0, TLBENTRY, G); |
| |
| if (tlb_g == g) { |
| tlb->tlb_misc = FIELD_DP64(tlb->tlb_misc, TLB_MISC, E, 0); |
| } |
| } |
| tlb_flush(env_cpu(env)); |
| } |
| |
| void helper_invtlb_all_asid(CPULoongArchState *env, target_ulong info) |
| { |
| uint16_t asid = info & R_CSR_ASID_ASID_MASK; |
| |
| for (int i = 0; i < LOONGARCH_TLB_MAX; i++) { |
| LoongArchTLB *tlb = &env->tlb[i]; |
| uint8_t tlb_g = FIELD_EX64(tlb->tlb_entry0, TLBENTRY, G); |
| uint16_t tlb_asid = FIELD_EX64(tlb->tlb_misc, TLB_MISC, ASID); |
| |
| if (!tlb_g && (tlb_asid == asid)) { |
| tlb->tlb_misc = FIELD_DP64(tlb->tlb_misc, TLB_MISC, E, 0); |
| } |
| } |
| tlb_flush(env_cpu(env)); |
| } |
| |
| void helper_invtlb_page_asid(CPULoongArchState *env, target_ulong info, |
| target_ulong addr) |
| { |
| uint16_t asid = info & 0x3ff; |
| |
| for (int i = 0; i < LOONGARCH_TLB_MAX; i++) { |
| LoongArchTLB *tlb = &env->tlb[i]; |
| uint8_t tlb_g = FIELD_EX64(tlb->tlb_entry0, TLBENTRY, G); |
| uint16_t tlb_asid = FIELD_EX64(tlb->tlb_misc, TLB_MISC, ASID); |
| uint64_t vpn, tlb_vppn; |
| uint8_t tlb_ps, compare_shift; |
| |
| if (i >= LOONGARCH_STLB) { |
| tlb_ps = FIELD_EX64(tlb->tlb_misc, TLB_MISC, PS); |
| } else { |
| tlb_ps = FIELD_EX64(env->CSR_STLBPS, CSR_STLBPS, PS); |
| } |
| tlb_vppn = FIELD_EX64(tlb->tlb_misc, TLB_MISC, VPPN); |
| vpn = (addr & TARGET_VIRT_MASK) >> (tlb_ps + 1); |
| compare_shift = tlb_ps + 1 - R_TLB_MISC_VPPN_SHIFT; |
| |
| if (!tlb_g && (tlb_asid == asid) && |
| (vpn == (tlb_vppn >> compare_shift))) { |
| tlb->tlb_misc = FIELD_DP64(tlb->tlb_misc, TLB_MISC, E, 0); |
| } |
| } |
| tlb_flush(env_cpu(env)); |
| } |
| |
| void helper_invtlb_page_asid_or_g(CPULoongArchState *env, |
| target_ulong info, target_ulong addr) |
| { |
| uint16_t asid = info & 0x3ff; |
| |
| for (int i = 0; i < LOONGARCH_TLB_MAX; i++) { |
| LoongArchTLB *tlb = &env->tlb[i]; |
| uint8_t tlb_g = FIELD_EX64(tlb->tlb_entry0, TLBENTRY, G); |
| uint16_t tlb_asid = FIELD_EX64(tlb->tlb_misc, TLB_MISC, ASID); |
| uint64_t vpn, tlb_vppn; |
| uint8_t tlb_ps, compare_shift; |
| |
| if (i >= LOONGARCH_STLB) { |
| tlb_ps = FIELD_EX64(tlb->tlb_misc, TLB_MISC, PS); |
| } else { |
| tlb_ps = FIELD_EX64(env->CSR_STLBPS, CSR_STLBPS, PS); |
| } |
| tlb_vppn = FIELD_EX64(tlb->tlb_misc, TLB_MISC, VPPN); |
| vpn = (addr & TARGET_VIRT_MASK) >> (tlb_ps + 1); |
| compare_shift = tlb_ps + 1 - R_TLB_MISC_VPPN_SHIFT; |
| |
| if ((tlb_g || (tlb_asid == asid)) && |
| (vpn == (tlb_vppn >> compare_shift))) { |
| tlb->tlb_misc = FIELD_DP64(tlb->tlb_misc, TLB_MISC, E, 0); |
| } |
| } |
| tlb_flush(env_cpu(env)); |
| } |
| |
| bool loongarch_cpu_tlb_fill(CPUState *cs, vaddr address, int size, |
| MMUAccessType access_type, int mmu_idx, |
| bool probe, uintptr_t retaddr) |
| { |
| LoongArchCPU *cpu = LOONGARCH_CPU(cs); |
| CPULoongArchState *env = &cpu->env; |
| hwaddr physical; |
| int prot; |
| int ret; |
| |
| /* Data access */ |
| ret = get_physical_address(env, &physical, &prot, address, |
| access_type, mmu_idx); |
| |
| if (ret == TLBRET_MATCH) { |
| tlb_set_page(cs, address & TARGET_PAGE_MASK, |
| physical & TARGET_PAGE_MASK, prot, |
| mmu_idx, TARGET_PAGE_SIZE); |
| qemu_log_mask(CPU_LOG_MMU, |
| "%s address=%" VADDR_PRIx " physical " HWADDR_FMT_plx |
| " prot %d\n", __func__, address, physical, prot); |
| return true; |
| } else { |
| qemu_log_mask(CPU_LOG_MMU, |
| "%s address=%" VADDR_PRIx " ret %d\n", __func__, address, |
| ret); |
| } |
| if (probe) { |
| return false; |
| } |
| raise_mmu_exception(env, address, access_type, ret); |
| cpu_loop_exit_restore(cs, retaddr); |
| } |
| |
| target_ulong helper_lddir(CPULoongArchState *env, target_ulong base, |
| target_ulong level, uint32_t mem_idx) |
| { |
| CPUState *cs = env_cpu(env); |
| target_ulong badvaddr, index, phys, ret; |
| int shift; |
| uint64_t dir_base, dir_width; |
| bool huge = (base >> LOONGARCH_PAGE_HUGE_SHIFT) & 0x1; |
| |
| badvaddr = env->CSR_TLBRBADV; |
| base = base & TARGET_PHYS_MASK; |
| |
| /* 0:64bit, 1:128bit, 2:192bit, 3:256bit */ |
| shift = FIELD_EX64(env->CSR_PWCL, CSR_PWCL, PTEWIDTH); |
| shift = (shift + 1) * 3; |
| |
| if (huge) { |
| return base; |
| } |
| switch (level) { |
| case 1: |
| dir_base = FIELD_EX64(env->CSR_PWCL, CSR_PWCL, DIR1_BASE); |
| dir_width = FIELD_EX64(env->CSR_PWCL, CSR_PWCL, DIR1_WIDTH); |
| break; |
| case 2: |
| dir_base = FIELD_EX64(env->CSR_PWCL, CSR_PWCL, DIR2_BASE); |
| dir_width = FIELD_EX64(env->CSR_PWCL, CSR_PWCL, DIR2_WIDTH); |
| break; |
| case 3: |
| dir_base = FIELD_EX64(env->CSR_PWCH, CSR_PWCH, DIR3_BASE); |
| dir_width = FIELD_EX64(env->CSR_PWCH, CSR_PWCH, DIR3_WIDTH); |
| break; |
| case 4: |
| dir_base = FIELD_EX64(env->CSR_PWCH, CSR_PWCH, DIR4_BASE); |
| dir_width = FIELD_EX64(env->CSR_PWCH, CSR_PWCH, DIR4_WIDTH); |
| break; |
| default: |
| do_raise_exception(env, EXCCODE_INE, GETPC()); |
| return 0; |
| } |
| index = (badvaddr >> dir_base) & ((1 << dir_width) - 1); |
| phys = base | index << shift; |
| ret = ldq_phys(cs->as, phys) & TARGET_PHYS_MASK; |
| return ret; |
| } |
| |
| void helper_ldpte(CPULoongArchState *env, target_ulong base, target_ulong odd, |
| uint32_t mem_idx) |
| { |
| CPUState *cs = env_cpu(env); |
| target_ulong phys, tmp0, ptindex, ptoffset0, ptoffset1, ps, badv; |
| int shift; |
| bool huge = (base >> LOONGARCH_PAGE_HUGE_SHIFT) & 0x1; |
| uint64_t ptbase = FIELD_EX64(env->CSR_PWCL, CSR_PWCL, PTBASE); |
| uint64_t ptwidth = FIELD_EX64(env->CSR_PWCL, CSR_PWCL, PTWIDTH); |
| |
| base = base & TARGET_PHYS_MASK; |
| |
| if (huge) { |
| /* Huge Page. base is paddr */ |
| tmp0 = base ^ (1 << LOONGARCH_PAGE_HUGE_SHIFT); |
| /* Move Global bit */ |
| tmp0 = ((tmp0 & (1 << LOONGARCH_HGLOBAL_SHIFT)) >> |
| LOONGARCH_HGLOBAL_SHIFT) << R_TLBENTRY_G_SHIFT | |
| (tmp0 & (~(1 << LOONGARCH_HGLOBAL_SHIFT))); |
| ps = ptbase + ptwidth - 1; |
| if (odd) { |
| tmp0 += MAKE_64BIT_MASK(ps, 1); |
| } |
| } else { |
| /* 0:64bit, 1:128bit, 2:192bit, 3:256bit */ |
| shift = FIELD_EX64(env->CSR_PWCL, CSR_PWCL, PTEWIDTH); |
| shift = (shift + 1) * 3; |
| badv = env->CSR_TLBRBADV; |
| |
| ptindex = (badv >> ptbase) & ((1 << ptwidth) - 1); |
| ptindex = ptindex & ~0x1; /* clear bit 0 */ |
| ptoffset0 = ptindex << shift; |
| ptoffset1 = (ptindex + 1) << shift; |
| |
| phys = base | (odd ? ptoffset1 : ptoffset0); |
| tmp0 = ldq_phys(cs->as, phys) & TARGET_PHYS_MASK; |
| ps = ptbase; |
| } |
| |
| if (odd) { |
| env->CSR_TLBRELO1 = tmp0; |
| } else { |
| env->CSR_TLBRELO0 = tmp0; |
| } |
| env->CSR_TLBREHI = FIELD_DP64(env->CSR_TLBREHI, CSR_TLBREHI, PS, ps); |
| } |