| /* |
| * QEMU RISC-V CPU |
| * |
| * Copyright (c) 2016-2017 Sagar Karandikar, sagark@eecs.berkeley.edu |
| * Copyright (c) 2017-2018 SiFive, Inc. |
| * |
| * This program is free software; you can redistribute it and/or modify it |
| * under the terms and conditions of the GNU General Public License, |
| * version 2 or later, as published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope it will be useful, but WITHOUT |
| * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for |
| * more details. |
| * |
| * You should have received a copy of the GNU General Public License along with |
| * this program. If not, see <http://www.gnu.org/licenses/>. |
| */ |
| |
| #ifndef RISCV_CPU_H |
| #define RISCV_CPU_H |
| |
| #include "hw/core/cpu.h" |
| #include "hw/registerfields.h" |
| #include "exec/cpu-defs.h" |
| #include "qemu/cpu-float.h" |
| #include "qom/object.h" |
| #include "qemu/int128.h" |
| #include "cpu_bits.h" |
| |
| #define TCG_GUEST_DEFAULT_MO 0 |
| |
| #define TYPE_RISCV_CPU "riscv-cpu" |
| |
| #define RISCV_CPU_TYPE_SUFFIX "-" TYPE_RISCV_CPU |
| #define RISCV_CPU_TYPE_NAME(name) (name RISCV_CPU_TYPE_SUFFIX) |
| #define CPU_RESOLVING_TYPE TYPE_RISCV_CPU |
| |
| #define TYPE_RISCV_CPU_ANY RISCV_CPU_TYPE_NAME("any") |
| #define TYPE_RISCV_CPU_BASE32 RISCV_CPU_TYPE_NAME("rv32") |
| #define TYPE_RISCV_CPU_BASE64 RISCV_CPU_TYPE_NAME("rv64") |
| #define TYPE_RISCV_CPU_BASE128 RISCV_CPU_TYPE_NAME("x-rv128") |
| #define TYPE_RISCV_CPU_IBEX RISCV_CPU_TYPE_NAME("lowrisc-ibex") |
| #define TYPE_RISCV_CPU_SHAKTI_C RISCV_CPU_TYPE_NAME("shakti-c") |
| #define TYPE_RISCV_CPU_SIFIVE_E31 RISCV_CPU_TYPE_NAME("sifive-e31") |
| #define TYPE_RISCV_CPU_SIFIVE_E34 RISCV_CPU_TYPE_NAME("sifive-e34") |
| #define TYPE_RISCV_CPU_SIFIVE_E51 RISCV_CPU_TYPE_NAME("sifive-e51") |
| #define TYPE_RISCV_CPU_SIFIVE_U34 RISCV_CPU_TYPE_NAME("sifive-u34") |
| #define TYPE_RISCV_CPU_SIFIVE_U54 RISCV_CPU_TYPE_NAME("sifive-u54") |
| #define TYPE_RISCV_CPU_HOST RISCV_CPU_TYPE_NAME("host") |
| |
| #if defined(TARGET_RISCV32) |
| # define TYPE_RISCV_CPU_BASE TYPE_RISCV_CPU_BASE32 |
| #elif defined(TARGET_RISCV64) |
| # define TYPE_RISCV_CPU_BASE TYPE_RISCV_CPU_BASE64 |
| #endif |
| |
| #define RV(x) ((target_ulong)1 << (x - 'A')) |
| |
| #define RVI RV('I') |
| #define RVE RV('E') /* E and I are mutually exclusive */ |
| #define RVM RV('M') |
| #define RVA RV('A') |
| #define RVF RV('F') |
| #define RVD RV('D') |
| #define RVV RV('V') |
| #define RVC RV('C') |
| #define RVS RV('S') |
| #define RVU RV('U') |
| #define RVH RV('H') |
| #define RVJ RV('J') |
| |
| /* S extension denotes that Supervisor mode exists, however it is possible |
| to have a core that support S mode but does not have an MMU and there |
| is currently no bit in misa to indicate whether an MMU exists or not |
| so a cpu features bitfield is required, likewise for optional PMP support */ |
| enum { |
| RISCV_FEATURE_MMU, |
| RISCV_FEATURE_PMP, |
| RISCV_FEATURE_EPMP, |
| RISCV_FEATURE_MISA, |
| RISCV_FEATURE_AIA |
| }; |
| |
| /* Privileged specification version */ |
| enum { |
| PRIV_VERSION_1_10_0 = 0, |
| PRIV_VERSION_1_11_0, |
| }; |
| |
| #define VEXT_VERSION_1_00_0 0x00010000 |
| |
| enum { |
| TRANSLATE_SUCCESS, |
| TRANSLATE_FAIL, |
| TRANSLATE_PMP_FAIL, |
| TRANSLATE_G_STAGE_FAIL |
| }; |
| |
| #define MMU_USER_IDX 3 |
| |
| #define MAX_RISCV_PMPS (16) |
| |
| typedef struct CPUArchState CPURISCVState; |
| |
| #if !defined(CONFIG_USER_ONLY) |
| #include "pmp.h" |
| #endif |
| |
| #define RV_VLEN_MAX 1024 |
| |
| FIELD(VTYPE, VLMUL, 0, 3) |
| FIELD(VTYPE, VSEW, 3, 3) |
| FIELD(VTYPE, VTA, 6, 1) |
| FIELD(VTYPE, VMA, 7, 1) |
| FIELD(VTYPE, VEDIV, 8, 2) |
| FIELD(VTYPE, RESERVED, 10, sizeof(target_ulong) * 8 - 11) |
| |
| struct CPUArchState { |
| target_ulong gpr[32]; |
| target_ulong gprh[32]; /* 64 top bits of the 128-bit registers */ |
| uint64_t fpr[32]; /* assume both F and D extensions */ |
| |
| /* vector coprocessor state. */ |
| uint64_t vreg[32 * RV_VLEN_MAX / 64] QEMU_ALIGNED(16); |
| target_ulong vxrm; |
| target_ulong vxsat; |
| target_ulong vl; |
| target_ulong vstart; |
| target_ulong vtype; |
| bool vill; |
| |
| target_ulong pc; |
| target_ulong load_res; |
| target_ulong load_val; |
| |
| target_ulong frm; |
| |
| target_ulong badaddr; |
| uint32_t bins; |
| |
| target_ulong guest_phys_fault_addr; |
| |
| target_ulong priv_ver; |
| target_ulong bext_ver; |
| target_ulong vext_ver; |
| |
| /* RISCVMXL, but uint32_t for vmstate migration */ |
| uint32_t misa_mxl; /* current mxl */ |
| uint32_t misa_mxl_max; /* max mxl for this cpu */ |
| uint32_t misa_ext; /* current extensions */ |
| uint32_t misa_ext_mask; /* max ext for this cpu */ |
| uint32_t xl; /* current xlen */ |
| |
| /* 128-bit helpers upper part return value */ |
| target_ulong retxh; |
| |
| uint32_t features; |
| |
| #ifdef CONFIG_USER_ONLY |
| uint32_t elf_flags; |
| #endif |
| |
| #ifndef CONFIG_USER_ONLY |
| target_ulong priv; |
| /* This contains QEMU specific information about the virt state. */ |
| target_ulong virt; |
| target_ulong geilen; |
| target_ulong resetvec; |
| |
| target_ulong mhartid; |
| /* |
| * For RV32 this is 32-bit mstatus and 32-bit mstatush. |
| * For RV64 this is a 64-bit mstatus. |
| */ |
| uint64_t mstatus; |
| |
| uint64_t mip; |
| |
| uint64_t miclaim; |
| |
| uint64_t mie; |
| uint64_t mideleg; |
| |
| target_ulong satp; /* since: priv-1.10.0 */ |
| target_ulong stval; |
| target_ulong medeleg; |
| |
| target_ulong stvec; |
| target_ulong sepc; |
| target_ulong scause; |
| |
| target_ulong mtvec; |
| target_ulong mepc; |
| target_ulong mcause; |
| target_ulong mtval; /* since: priv-1.10.0 */ |
| |
| /* Machine and Supervisor interrupt priorities */ |
| uint8_t miprio[64]; |
| uint8_t siprio[64]; |
| |
| /* AIA CSRs */ |
| target_ulong miselect; |
| target_ulong siselect; |
| |
| /* Hypervisor CSRs */ |
| target_ulong hstatus; |
| target_ulong hedeleg; |
| uint64_t hideleg; |
| target_ulong hcounteren; |
| target_ulong htval; |
| target_ulong htinst; |
| target_ulong hgatp; |
| target_ulong hgeie; |
| target_ulong hgeip; |
| uint64_t htimedelta; |
| |
| /* Hypervisor controlled virtual interrupt priorities */ |
| target_ulong hvictl; |
| uint8_t hviprio[64]; |
| |
| /* Upper 64-bits of 128-bit CSRs */ |
| uint64_t mscratchh; |
| uint64_t sscratchh; |
| |
| /* Virtual CSRs */ |
| /* |
| * For RV32 this is 32-bit vsstatus and 32-bit vsstatush. |
| * For RV64 this is a 64-bit vsstatus. |
| */ |
| uint64_t vsstatus; |
| target_ulong vstvec; |
| target_ulong vsscratch; |
| target_ulong vsepc; |
| target_ulong vscause; |
| target_ulong vstval; |
| target_ulong vsatp; |
| |
| /* AIA VS-mode CSRs */ |
| target_ulong vsiselect; |
| |
| target_ulong mtval2; |
| target_ulong mtinst; |
| |
| /* HS Backup CSRs */ |
| target_ulong stvec_hs; |
| target_ulong sscratch_hs; |
| target_ulong sepc_hs; |
| target_ulong scause_hs; |
| target_ulong stval_hs; |
| target_ulong satp_hs; |
| uint64_t mstatus_hs; |
| |
| /* Signals whether the current exception occurred with two-stage address |
| translation active. */ |
| bool two_stage_lookup; |
| |
| target_ulong scounteren; |
| target_ulong mcounteren; |
| |
| target_ulong sscratch; |
| target_ulong mscratch; |
| |
| /* temporary htif regs */ |
| uint64_t mfromhost; |
| uint64_t mtohost; |
| uint64_t timecmp; |
| |
| /* physical memory protection */ |
| pmp_table_t pmp_state; |
| target_ulong mseccfg; |
| |
| /* machine specific rdtime callback */ |
| uint64_t (*rdtime_fn)(uint32_t); |
| uint32_t rdtime_fn_arg; |
| |
| /* machine specific AIA ireg read-modify-write callback */ |
| #define AIA_MAKE_IREG(__isel, __priv, __virt, __vgein, __xlen) \ |
| ((((__xlen) & 0xff) << 24) | \ |
| (((__vgein) & 0x3f) << 20) | \ |
| (((__virt) & 0x1) << 18) | \ |
| (((__priv) & 0x3) << 16) | \ |
| (__isel & 0xffff)) |
| #define AIA_IREG_ISEL(__ireg) ((__ireg) & 0xffff) |
| #define AIA_IREG_PRIV(__ireg) (((__ireg) >> 16) & 0x3) |
| #define AIA_IREG_VIRT(__ireg) (((__ireg) >> 18) & 0x1) |
| #define AIA_IREG_VGEIN(__ireg) (((__ireg) >> 20) & 0x3f) |
| #define AIA_IREG_XLEN(__ireg) (((__ireg) >> 24) & 0xff) |
| int (*aia_ireg_rmw_fn[4])(void *arg, target_ulong reg, |
| target_ulong *val, target_ulong new_val, target_ulong write_mask); |
| void *aia_ireg_rmw_fn_arg[4]; |
| |
| /* True if in debugger mode. */ |
| bool debugger; |
| |
| /* |
| * CSRs for PointerMasking extension |
| */ |
| target_ulong mmte; |
| target_ulong mpmmask; |
| target_ulong mpmbase; |
| target_ulong spmmask; |
| target_ulong spmbase; |
| target_ulong upmmask; |
| target_ulong upmbase; |
| #endif |
| target_ulong cur_pmmask; |
| target_ulong cur_pmbase; |
| |
| float_status fp_status; |
| |
| /* Fields from here on are preserved across CPU reset. */ |
| QEMUTimer *timer; /* Internal timer */ |
| |
| hwaddr kernel_addr; |
| hwaddr fdt_addr; |
| |
| /* kvm timer */ |
| bool kvm_timer_dirty; |
| uint64_t kvm_timer_time; |
| uint64_t kvm_timer_compare; |
| uint64_t kvm_timer_state; |
| uint64_t kvm_timer_frequency; |
| }; |
| |
| OBJECT_DECLARE_CPU_TYPE(RISCVCPU, RISCVCPUClass, RISCV_CPU) |
| |
| /** |
| * RISCVCPUClass: |
| * @parent_realize: The parent class' realize handler. |
| * @parent_reset: The parent class' reset handler. |
| * |
| * A RISCV CPU model. |
| */ |
| struct RISCVCPUClass { |
| /*< private >*/ |
| CPUClass parent_class; |
| /*< public >*/ |
| DeviceRealize parent_realize; |
| DeviceReset parent_reset; |
| }; |
| |
| struct RISCVCPUConfig { |
| bool ext_i; |
| bool ext_e; |
| bool ext_g; |
| bool ext_m; |
| bool ext_a; |
| bool ext_f; |
| bool ext_d; |
| bool ext_c; |
| bool ext_s; |
| bool ext_u; |
| bool ext_h; |
| bool ext_j; |
| bool ext_v; |
| bool ext_zba; |
| bool ext_zbb; |
| bool ext_zbc; |
| bool ext_zbs; |
| bool ext_counters; |
| bool ext_ifencei; |
| bool ext_icsr; |
| bool ext_svinval; |
| bool ext_svnapot; |
| bool ext_svpbmt; |
| bool ext_zdinx; |
| bool ext_zfh; |
| bool ext_zfhmin; |
| bool ext_zfinx; |
| bool ext_zhinx; |
| bool ext_zhinxmin; |
| bool ext_zve32f; |
| bool ext_zve64f; |
| |
| /* Vendor-specific custom extensions */ |
| bool ext_XVentanaCondOps; |
| |
| char *priv_spec; |
| char *user_spec; |
| char *bext_spec; |
| char *vext_spec; |
| uint16_t vlen; |
| uint16_t elen; |
| bool mmu; |
| bool pmp; |
| bool epmp; |
| bool aia; |
| uint64_t resetvec; |
| }; |
| |
| typedef struct RISCVCPUConfig RISCVCPUConfig; |
| |
| /** |
| * RISCVCPU: |
| * @env: #CPURISCVState |
| * |
| * A RISCV CPU. |
| */ |
| struct ArchCPU { |
| /*< private >*/ |
| CPUState parent_obj; |
| /*< public >*/ |
| CPUNegativeOffsetState neg; |
| CPURISCVState env; |
| |
| char *dyn_csr_xml; |
| char *dyn_vreg_xml; |
| |
| /* Configuration Settings */ |
| RISCVCPUConfig cfg; |
| }; |
| |
| static inline int riscv_has_ext(CPURISCVState *env, target_ulong ext) |
| { |
| return (env->misa_ext & ext) != 0; |
| } |
| |
| static inline bool riscv_feature(CPURISCVState *env, int feature) |
| { |
| return env->features & (1ULL << feature); |
| } |
| |
| static inline void riscv_set_feature(CPURISCVState *env, int feature) |
| { |
| env->features |= (1ULL << feature); |
| } |
| |
| #include "cpu_user.h" |
| |
| extern const char * const riscv_int_regnames[]; |
| extern const char * const riscv_int_regnamesh[]; |
| extern const char * const riscv_fpr_regnames[]; |
| |
| const char *riscv_cpu_get_trap_name(target_ulong cause, bool async); |
| void riscv_cpu_do_interrupt(CPUState *cpu); |
| int riscv_cpu_write_elf64_note(WriteCoreDumpFunction f, CPUState *cs, |
| int cpuid, void *opaque); |
| int riscv_cpu_write_elf32_note(WriteCoreDumpFunction f, CPUState *cs, |
| int cpuid, void *opaque); |
| int riscv_cpu_gdb_read_register(CPUState *cpu, GByteArray *buf, int reg); |
| int riscv_cpu_gdb_write_register(CPUState *cpu, uint8_t *buf, int reg); |
| int riscv_cpu_hviprio_index2irq(int index, int *out_irq, int *out_rdzero); |
| uint8_t riscv_cpu_default_priority(int irq); |
| int riscv_cpu_mirq_pending(CPURISCVState *env); |
| int riscv_cpu_sirq_pending(CPURISCVState *env); |
| int riscv_cpu_vsirq_pending(CPURISCVState *env); |
| bool riscv_cpu_fp_enabled(CPURISCVState *env); |
| target_ulong riscv_cpu_get_geilen(CPURISCVState *env); |
| void riscv_cpu_set_geilen(CPURISCVState *env, target_ulong geilen); |
| bool riscv_cpu_vector_enabled(CPURISCVState *env); |
| bool riscv_cpu_virt_enabled(CPURISCVState *env); |
| void riscv_cpu_set_virt_enabled(CPURISCVState *env, bool enable); |
| bool riscv_cpu_two_stage_lookup(int mmu_idx); |
| int riscv_cpu_mmu_index(CPURISCVState *env, bool ifetch); |
| hwaddr riscv_cpu_get_phys_page_debug(CPUState *cpu, vaddr addr); |
| G_NORETURN void riscv_cpu_do_unaligned_access(CPUState *cs, vaddr addr, |
| MMUAccessType access_type, int mmu_idx, |
| uintptr_t retaddr); |
| bool riscv_cpu_tlb_fill(CPUState *cs, vaddr address, int size, |
| MMUAccessType access_type, int mmu_idx, |
| bool probe, uintptr_t retaddr); |
| void riscv_cpu_do_transaction_failed(CPUState *cs, hwaddr physaddr, |
| vaddr addr, unsigned size, |
| MMUAccessType access_type, |
| int mmu_idx, MemTxAttrs attrs, |
| MemTxResult response, uintptr_t retaddr); |
| char *riscv_isa_string(RISCVCPU *cpu); |
| void riscv_cpu_list(void); |
| |
| #define cpu_list riscv_cpu_list |
| #define cpu_mmu_index riscv_cpu_mmu_index |
| |
| #ifndef CONFIG_USER_ONLY |
| bool riscv_cpu_exec_interrupt(CPUState *cs, int interrupt_request); |
| void riscv_cpu_swap_hypervisor_regs(CPURISCVState *env); |
| int riscv_cpu_claim_interrupts(RISCVCPU *cpu, uint64_t interrupts); |
| uint64_t riscv_cpu_update_mip(RISCVCPU *cpu, uint64_t mask, uint64_t value); |
| #define BOOL_TO_MASK(x) (-!!(x)) /* helper for riscv_cpu_update_mip value */ |
| void riscv_cpu_set_rdtime_fn(CPURISCVState *env, uint64_t (*fn)(uint32_t), |
| uint32_t arg); |
| void riscv_cpu_set_aia_ireg_rmw_fn(CPURISCVState *env, uint32_t priv, |
| int (*rmw_fn)(void *arg, |
| target_ulong reg, |
| target_ulong *val, |
| target_ulong new_val, |
| target_ulong write_mask), |
| void *rmw_fn_arg); |
| #endif |
| void riscv_cpu_set_mode(CPURISCVState *env, target_ulong newpriv); |
| |
| void riscv_translate_init(void); |
| G_NORETURN void riscv_raise_exception(CPURISCVState *env, |
| uint32_t exception, uintptr_t pc); |
| |
| target_ulong riscv_cpu_get_fflags(CPURISCVState *env); |
| void riscv_cpu_set_fflags(CPURISCVState *env, target_ulong); |
| |
| #define TB_FLAGS_PRIV_MMU_MASK 3 |
| #define TB_FLAGS_PRIV_HYP_ACCESS_MASK (1 << 2) |
| #define TB_FLAGS_MSTATUS_FS MSTATUS_FS |
| #define TB_FLAGS_MSTATUS_VS MSTATUS_VS |
| |
| #include "exec/cpu-all.h" |
| |
| FIELD(TB_FLAGS, MEM_IDX, 0, 3) |
| FIELD(TB_FLAGS, LMUL, 3, 3) |
| FIELD(TB_FLAGS, SEW, 6, 3) |
| /* Skip MSTATUS_VS (0x600) bits */ |
| FIELD(TB_FLAGS, VL_EQ_VLMAX, 11, 1) |
| FIELD(TB_FLAGS, VILL, 12, 1) |
| /* Skip MSTATUS_FS (0x6000) bits */ |
| /* Is a Hypervisor instruction load/store allowed? */ |
| FIELD(TB_FLAGS, HLSX, 15, 1) |
| FIELD(TB_FLAGS, MSTATUS_HS_FS, 16, 2) |
| FIELD(TB_FLAGS, MSTATUS_HS_VS, 18, 2) |
| /* The combination of MXL/SXL/UXL that applies to the current cpu mode. */ |
| FIELD(TB_FLAGS, XL, 20, 2) |
| /* If PointerMasking should be applied */ |
| FIELD(TB_FLAGS, PM_MASK_ENABLED, 22, 1) |
| FIELD(TB_FLAGS, PM_BASE_ENABLED, 23, 1) |
| |
| #ifdef TARGET_RISCV32 |
| #define riscv_cpu_mxl(env) ((void)(env), MXL_RV32) |
| #else |
| static inline RISCVMXL riscv_cpu_mxl(CPURISCVState *env) |
| { |
| return env->misa_mxl; |
| } |
| #endif |
| #define riscv_cpu_mxl_bits(env) (1UL << (4 + riscv_cpu_mxl(env))) |
| |
| #if defined(TARGET_RISCV32) |
| #define cpu_recompute_xl(env) ((void)(env), MXL_RV32) |
| #else |
| static inline RISCVMXL cpu_recompute_xl(CPURISCVState *env) |
| { |
| RISCVMXL xl = env->misa_mxl; |
| #if !defined(CONFIG_USER_ONLY) |
| /* |
| * When emulating a 32-bit-only cpu, use RV32. |
| * When emulating a 64-bit cpu, and MXL has been reduced to RV32, |
| * MSTATUSH doesn't have UXL/SXL, therefore XLEN cannot be widened |
| * back to RV64 for lower privs. |
| */ |
| if (xl != MXL_RV32) { |
| switch (env->priv) { |
| case PRV_M: |
| break; |
| case PRV_U: |
| xl = get_field(env->mstatus, MSTATUS64_UXL); |
| break; |
| default: /* PRV_S | PRV_H */ |
| xl = get_field(env->mstatus, MSTATUS64_SXL); |
| break; |
| } |
| } |
| #endif |
| return xl; |
| } |
| #endif |
| |
| static inline int riscv_cpu_xlen(CPURISCVState *env) |
| { |
| return 16 << env->xl; |
| } |
| |
| #ifdef TARGET_RISCV32 |
| #define riscv_cpu_sxl(env) ((void)(env), MXL_RV32) |
| #else |
| static inline RISCVMXL riscv_cpu_sxl(CPURISCVState *env) |
| { |
| #ifdef CONFIG_USER_ONLY |
| return env->misa_mxl; |
| #else |
| return get_field(env->mstatus, MSTATUS64_SXL); |
| #endif |
| } |
| #endif |
| |
| /* |
| * Encode LMUL to lmul as follows: |
| * LMUL vlmul lmul |
| * 1 000 0 |
| * 2 001 1 |
| * 4 010 2 |
| * 8 011 3 |
| * - 100 - |
| * 1/8 101 -3 |
| * 1/4 110 -2 |
| * 1/2 111 -1 |
| * |
| * then, we can calculate VLMAX = vlen >> (vsew + 3 - lmul) |
| * e.g. vlen = 256 bits, SEW = 16, LMUL = 1/8 |
| * => VLMAX = vlen >> (1 + 3 - (-3)) |
| * = 256 >> 7 |
| * = 2 |
| */ |
| static inline uint32_t vext_get_vlmax(RISCVCPU *cpu, target_ulong vtype) |
| { |
| uint8_t sew = FIELD_EX64(vtype, VTYPE, VSEW); |
| int8_t lmul = sextract32(FIELD_EX64(vtype, VTYPE, VLMUL), 0, 3); |
| return cpu->cfg.vlen >> (sew + 3 - lmul); |
| } |
| |
| void cpu_get_tb_cpu_state(CPURISCVState *env, target_ulong *pc, |
| target_ulong *cs_base, uint32_t *pflags); |
| |
| void riscv_cpu_update_mask(CPURISCVState *env); |
| |
| RISCVException riscv_csrrw(CPURISCVState *env, int csrno, |
| target_ulong *ret_value, |
| target_ulong new_value, target_ulong write_mask); |
| RISCVException riscv_csrrw_debug(CPURISCVState *env, int csrno, |
| target_ulong *ret_value, |
| target_ulong new_value, |
| target_ulong write_mask); |
| |
| static inline void riscv_csr_write(CPURISCVState *env, int csrno, |
| target_ulong val) |
| { |
| riscv_csrrw(env, csrno, NULL, val, MAKE_64BIT_MASK(0, TARGET_LONG_BITS)); |
| } |
| |
| static inline target_ulong riscv_csr_read(CPURISCVState *env, int csrno) |
| { |
| target_ulong val = 0; |
| riscv_csrrw(env, csrno, &val, 0, 0); |
| return val; |
| } |
| |
| typedef RISCVException (*riscv_csr_predicate_fn)(CPURISCVState *env, |
| int csrno); |
| typedef RISCVException (*riscv_csr_read_fn)(CPURISCVState *env, int csrno, |
| target_ulong *ret_value); |
| typedef RISCVException (*riscv_csr_write_fn)(CPURISCVState *env, int csrno, |
| target_ulong new_value); |
| typedef RISCVException (*riscv_csr_op_fn)(CPURISCVState *env, int csrno, |
| target_ulong *ret_value, |
| target_ulong new_value, |
| target_ulong write_mask); |
| |
| RISCVException riscv_csrrw_i128(CPURISCVState *env, int csrno, |
| Int128 *ret_value, |
| Int128 new_value, Int128 write_mask); |
| |
| typedef RISCVException (*riscv_csr_read128_fn)(CPURISCVState *env, int csrno, |
| Int128 *ret_value); |
| typedef RISCVException (*riscv_csr_write128_fn)(CPURISCVState *env, int csrno, |
| Int128 new_value); |
| |
| typedef struct { |
| const char *name; |
| riscv_csr_predicate_fn predicate; |
| riscv_csr_read_fn read; |
| riscv_csr_write_fn write; |
| riscv_csr_op_fn op; |
| riscv_csr_read128_fn read128; |
| riscv_csr_write128_fn write128; |
| } riscv_csr_operations; |
| |
| /* CSR function table constants */ |
| enum { |
| CSR_TABLE_SIZE = 0x1000 |
| }; |
| |
| /* CSR function table */ |
| extern riscv_csr_operations csr_ops[CSR_TABLE_SIZE]; |
| |
| void riscv_get_csr_ops(int csrno, riscv_csr_operations *ops); |
| void riscv_set_csr_ops(int csrno, riscv_csr_operations *ops); |
| |
| void riscv_cpu_register_gdb_regs_for_features(CPUState *cs); |
| |
| #endif /* RISCV_CPU_H */ |