| /* |
| * QEMU RISC-V CPU |
| * |
| * Copyright (c) 2016-2017 Sagar Karandikar, sagark@eecs.berkeley.edu |
| * Copyright (c) 2017-2018 SiFive, Inc. |
| * |
| * This program is free software; you can redistribute it and/or modify it |
| * under the terms and conditions of the GNU General Public License, |
| * version 2 or later, as published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope it will be useful, but WITHOUT |
| * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for |
| * more details. |
| * |
| * You should have received a copy of the GNU General Public License along with |
| * this program. If not, see <http://www.gnu.org/licenses/>. |
| */ |
| |
| #include "qemu/osdep.h" |
| #include "qemu/qemu-print.h" |
| #include "qemu/ctype.h" |
| #include "qemu/log.h" |
| #include "cpu.h" |
| #include "cpu_vendorid.h" |
| #include "internals.h" |
| #include "exec/exec-all.h" |
| #include "qapi/error.h" |
| #include "qapi/visitor.h" |
| #include "qemu/error-report.h" |
| #include "hw/qdev-properties.h" |
| #include "hw/core/qdev-prop-internal.h" |
| #include "migration/vmstate.h" |
| #include "fpu/softfloat-helpers.h" |
| #include "sysemu/device_tree.h" |
| #include "sysemu/kvm.h" |
| #include "sysemu/tcg.h" |
| #include "kvm/kvm_riscv.h" |
| #include "tcg/tcg-cpu.h" |
| #include "tcg/tcg.h" |
| |
| /* RISC-V CPU definitions */ |
| static const char riscv_single_letter_exts[] = "IEMAFDQCBPVH"; |
| const uint32_t misa_bits[] = {RVI, RVE, RVM, RVA, RVF, RVD, RVV, |
| RVC, RVS, RVU, RVH, RVJ, RVG, RVB, 0}; |
| |
| /* |
| * From vector_helper.c |
| * Note that vector data is stored in host-endian 64-bit chunks, |
| * so addressing bytes needs a host-endian fixup. |
| */ |
| #if HOST_BIG_ENDIAN |
| #define BYTE(x) ((x) ^ 7) |
| #else |
| #define BYTE(x) (x) |
| #endif |
| |
| bool riscv_cpu_is_32bit(RISCVCPU *cpu) |
| { |
| return riscv_cpu_mxl(&cpu->env) == MXL_RV32; |
| } |
| |
| /* Hash that stores general user set numeric options */ |
| static GHashTable *general_user_opts; |
| |
| static void cpu_option_add_user_setting(const char *optname, uint32_t value) |
| { |
| g_hash_table_insert(general_user_opts, (gpointer)optname, |
| GUINT_TO_POINTER(value)); |
| } |
| |
| bool riscv_cpu_option_set(const char *optname) |
| { |
| return g_hash_table_contains(general_user_opts, optname); |
| } |
| |
| #define ISA_EXT_DATA_ENTRY(_name, _min_ver, _prop) \ |
| {#_name, _min_ver, CPU_CFG_OFFSET(_prop)} |
| |
| /* |
| * Here are the ordering rules of extension naming defined by RISC-V |
| * specification : |
| * 1. All extensions should be separated from other multi-letter extensions |
| * by an underscore. |
| * 2. The first letter following the 'Z' conventionally indicates the most |
| * closely related alphabetical extension category, IMAFDQLCBKJTPVH. |
| * If multiple 'Z' extensions are named, they should be ordered first |
| * by category, then alphabetically within a category. |
| * 3. Standard supervisor-level extensions (starts with 'S') should be |
| * listed after standard unprivileged extensions. If multiple |
| * supervisor-level extensions are listed, they should be ordered |
| * alphabetically. |
| * 4. Non-standard extensions (starts with 'X') must be listed after all |
| * standard extensions. They must be separated from other multi-letter |
| * extensions by an underscore. |
| * |
| * Single letter extensions are checked in riscv_cpu_validate_misa_priv() |
| * instead. |
| */ |
| const RISCVIsaExtData isa_edata_arr[] = { |
| ISA_EXT_DATA_ENTRY(zic64b, PRIV_VERSION_1_12_0, ext_zic64b), |
| ISA_EXT_DATA_ENTRY(zicbom, PRIV_VERSION_1_12_0, ext_zicbom), |
| ISA_EXT_DATA_ENTRY(zicbop, PRIV_VERSION_1_12_0, ext_zicbop), |
| ISA_EXT_DATA_ENTRY(zicboz, PRIV_VERSION_1_12_0, ext_zicboz), |
| ISA_EXT_DATA_ENTRY(ziccamoa, PRIV_VERSION_1_11_0, has_priv_1_11), |
| ISA_EXT_DATA_ENTRY(ziccif, PRIV_VERSION_1_11_0, has_priv_1_11), |
| ISA_EXT_DATA_ENTRY(zicclsm, PRIV_VERSION_1_11_0, has_priv_1_11), |
| ISA_EXT_DATA_ENTRY(ziccrse, PRIV_VERSION_1_11_0, has_priv_1_11), |
| ISA_EXT_DATA_ENTRY(zicond, PRIV_VERSION_1_12_0, ext_zicond), |
| ISA_EXT_DATA_ENTRY(zicntr, PRIV_VERSION_1_12_0, ext_zicntr), |
| ISA_EXT_DATA_ENTRY(zicsr, PRIV_VERSION_1_10_0, ext_zicsr), |
| ISA_EXT_DATA_ENTRY(zifencei, PRIV_VERSION_1_10_0, ext_zifencei), |
| ISA_EXT_DATA_ENTRY(zihintntl, PRIV_VERSION_1_10_0, ext_zihintntl), |
| ISA_EXT_DATA_ENTRY(zihintpause, PRIV_VERSION_1_10_0, ext_zihintpause), |
| ISA_EXT_DATA_ENTRY(zihpm, PRIV_VERSION_1_12_0, ext_zihpm), |
| ISA_EXT_DATA_ENTRY(zimop, PRIV_VERSION_1_13_0, ext_zimop), |
| ISA_EXT_DATA_ENTRY(zmmul, PRIV_VERSION_1_12_0, ext_zmmul), |
| ISA_EXT_DATA_ENTRY(za64rs, PRIV_VERSION_1_12_0, has_priv_1_11), |
| ISA_EXT_DATA_ENTRY(zaamo, PRIV_VERSION_1_12_0, ext_zaamo), |
| ISA_EXT_DATA_ENTRY(zabha, PRIV_VERSION_1_13_0, ext_zabha), |
| ISA_EXT_DATA_ENTRY(zacas, PRIV_VERSION_1_12_0, ext_zacas), |
| ISA_EXT_DATA_ENTRY(zama16b, PRIV_VERSION_1_13_0, ext_zama16b), |
| ISA_EXT_DATA_ENTRY(zalrsc, PRIV_VERSION_1_12_0, ext_zalrsc), |
| ISA_EXT_DATA_ENTRY(zawrs, PRIV_VERSION_1_12_0, ext_zawrs), |
| ISA_EXT_DATA_ENTRY(zfa, PRIV_VERSION_1_12_0, ext_zfa), |
| ISA_EXT_DATA_ENTRY(zfbfmin, PRIV_VERSION_1_12_0, ext_zfbfmin), |
| ISA_EXT_DATA_ENTRY(zfh, PRIV_VERSION_1_11_0, ext_zfh), |
| ISA_EXT_DATA_ENTRY(zfhmin, PRIV_VERSION_1_11_0, ext_zfhmin), |
| ISA_EXT_DATA_ENTRY(zfinx, PRIV_VERSION_1_12_0, ext_zfinx), |
| ISA_EXT_DATA_ENTRY(zdinx, PRIV_VERSION_1_12_0, ext_zdinx), |
| ISA_EXT_DATA_ENTRY(zca, PRIV_VERSION_1_12_0, ext_zca), |
| ISA_EXT_DATA_ENTRY(zcb, PRIV_VERSION_1_12_0, ext_zcb), |
| ISA_EXT_DATA_ENTRY(zcf, PRIV_VERSION_1_12_0, ext_zcf), |
| ISA_EXT_DATA_ENTRY(zcd, PRIV_VERSION_1_12_0, ext_zcd), |
| ISA_EXT_DATA_ENTRY(zce, PRIV_VERSION_1_12_0, ext_zce), |
| ISA_EXT_DATA_ENTRY(zcmop, PRIV_VERSION_1_13_0, ext_zcmop), |
| ISA_EXT_DATA_ENTRY(zcmp, PRIV_VERSION_1_12_0, ext_zcmp), |
| ISA_EXT_DATA_ENTRY(zcmt, PRIV_VERSION_1_12_0, ext_zcmt), |
| ISA_EXT_DATA_ENTRY(zba, PRIV_VERSION_1_12_0, ext_zba), |
| ISA_EXT_DATA_ENTRY(zbb, PRIV_VERSION_1_12_0, ext_zbb), |
| ISA_EXT_DATA_ENTRY(zbc, PRIV_VERSION_1_12_0, ext_zbc), |
| ISA_EXT_DATA_ENTRY(zbkb, PRIV_VERSION_1_12_0, ext_zbkb), |
| ISA_EXT_DATA_ENTRY(zbkc, PRIV_VERSION_1_12_0, ext_zbkc), |
| ISA_EXT_DATA_ENTRY(zbkx, PRIV_VERSION_1_12_0, ext_zbkx), |
| ISA_EXT_DATA_ENTRY(zbs, PRIV_VERSION_1_12_0, ext_zbs), |
| ISA_EXT_DATA_ENTRY(zk, PRIV_VERSION_1_12_0, ext_zk), |
| ISA_EXT_DATA_ENTRY(zkn, PRIV_VERSION_1_12_0, ext_zkn), |
| ISA_EXT_DATA_ENTRY(zknd, PRIV_VERSION_1_12_0, ext_zknd), |
| ISA_EXT_DATA_ENTRY(zkne, PRIV_VERSION_1_12_0, ext_zkne), |
| ISA_EXT_DATA_ENTRY(zknh, PRIV_VERSION_1_12_0, ext_zknh), |
| ISA_EXT_DATA_ENTRY(zkr, PRIV_VERSION_1_12_0, ext_zkr), |
| ISA_EXT_DATA_ENTRY(zks, PRIV_VERSION_1_12_0, ext_zks), |
| ISA_EXT_DATA_ENTRY(zksed, PRIV_VERSION_1_12_0, ext_zksed), |
| ISA_EXT_DATA_ENTRY(zksh, PRIV_VERSION_1_12_0, ext_zksh), |
| ISA_EXT_DATA_ENTRY(zkt, PRIV_VERSION_1_12_0, ext_zkt), |
| ISA_EXT_DATA_ENTRY(ztso, PRIV_VERSION_1_12_0, ext_ztso), |
| ISA_EXT_DATA_ENTRY(zvbb, PRIV_VERSION_1_12_0, ext_zvbb), |
| ISA_EXT_DATA_ENTRY(zvbc, PRIV_VERSION_1_12_0, ext_zvbc), |
| ISA_EXT_DATA_ENTRY(zve32f, PRIV_VERSION_1_10_0, ext_zve32f), |
| ISA_EXT_DATA_ENTRY(zve32x, PRIV_VERSION_1_10_0, ext_zve32x), |
| ISA_EXT_DATA_ENTRY(zve64f, PRIV_VERSION_1_10_0, ext_zve64f), |
| ISA_EXT_DATA_ENTRY(zve64d, PRIV_VERSION_1_10_0, ext_zve64d), |
| ISA_EXT_DATA_ENTRY(zve64x, PRIV_VERSION_1_10_0, ext_zve64x), |
| ISA_EXT_DATA_ENTRY(zvfbfmin, PRIV_VERSION_1_12_0, ext_zvfbfmin), |
| ISA_EXT_DATA_ENTRY(zvfbfwma, PRIV_VERSION_1_12_0, ext_zvfbfwma), |
| ISA_EXT_DATA_ENTRY(zvfh, PRIV_VERSION_1_12_0, ext_zvfh), |
| ISA_EXT_DATA_ENTRY(zvfhmin, PRIV_VERSION_1_12_0, ext_zvfhmin), |
| ISA_EXT_DATA_ENTRY(zvkb, PRIV_VERSION_1_12_0, ext_zvkb), |
| ISA_EXT_DATA_ENTRY(zvkg, PRIV_VERSION_1_12_0, ext_zvkg), |
| ISA_EXT_DATA_ENTRY(zvkn, PRIV_VERSION_1_12_0, ext_zvkn), |
| ISA_EXT_DATA_ENTRY(zvknc, PRIV_VERSION_1_12_0, ext_zvknc), |
| ISA_EXT_DATA_ENTRY(zvkned, PRIV_VERSION_1_12_0, ext_zvkned), |
| ISA_EXT_DATA_ENTRY(zvkng, PRIV_VERSION_1_12_0, ext_zvkng), |
| ISA_EXT_DATA_ENTRY(zvknha, PRIV_VERSION_1_12_0, ext_zvknha), |
| ISA_EXT_DATA_ENTRY(zvknhb, PRIV_VERSION_1_12_0, ext_zvknhb), |
| ISA_EXT_DATA_ENTRY(zvks, PRIV_VERSION_1_12_0, ext_zvks), |
| ISA_EXT_DATA_ENTRY(zvksc, PRIV_VERSION_1_12_0, ext_zvksc), |
| ISA_EXT_DATA_ENTRY(zvksed, PRIV_VERSION_1_12_0, ext_zvksed), |
| ISA_EXT_DATA_ENTRY(zvksg, PRIV_VERSION_1_12_0, ext_zvksg), |
| ISA_EXT_DATA_ENTRY(zvksh, PRIV_VERSION_1_12_0, ext_zvksh), |
| ISA_EXT_DATA_ENTRY(zvkt, PRIV_VERSION_1_12_0, ext_zvkt), |
| ISA_EXT_DATA_ENTRY(zhinx, PRIV_VERSION_1_12_0, ext_zhinx), |
| ISA_EXT_DATA_ENTRY(zhinxmin, PRIV_VERSION_1_12_0, ext_zhinxmin), |
| ISA_EXT_DATA_ENTRY(smaia, PRIV_VERSION_1_12_0, ext_smaia), |
| ISA_EXT_DATA_ENTRY(smcntrpmf, PRIV_VERSION_1_12_0, ext_smcntrpmf), |
| ISA_EXT_DATA_ENTRY(smepmp, PRIV_VERSION_1_12_0, ext_smepmp), |
| ISA_EXT_DATA_ENTRY(smstateen, PRIV_VERSION_1_12_0, ext_smstateen), |
| ISA_EXT_DATA_ENTRY(ssaia, PRIV_VERSION_1_12_0, ext_ssaia), |
| ISA_EXT_DATA_ENTRY(ssccptr, PRIV_VERSION_1_11_0, has_priv_1_11), |
| ISA_EXT_DATA_ENTRY(sscofpmf, PRIV_VERSION_1_12_0, ext_sscofpmf), |
| ISA_EXT_DATA_ENTRY(sscounterenw, PRIV_VERSION_1_12_0, has_priv_1_12), |
| ISA_EXT_DATA_ENTRY(sstc, PRIV_VERSION_1_12_0, ext_sstc), |
| ISA_EXT_DATA_ENTRY(sstvala, PRIV_VERSION_1_12_0, has_priv_1_12), |
| ISA_EXT_DATA_ENTRY(sstvecd, PRIV_VERSION_1_12_0, has_priv_1_12), |
| ISA_EXT_DATA_ENTRY(svade, PRIV_VERSION_1_11_0, ext_svade), |
| ISA_EXT_DATA_ENTRY(svadu, PRIV_VERSION_1_12_0, ext_svadu), |
| ISA_EXT_DATA_ENTRY(svinval, PRIV_VERSION_1_12_0, ext_svinval), |
| ISA_EXT_DATA_ENTRY(svnapot, PRIV_VERSION_1_12_0, ext_svnapot), |
| ISA_EXT_DATA_ENTRY(svpbmt, PRIV_VERSION_1_12_0, ext_svpbmt), |
| ISA_EXT_DATA_ENTRY(xtheadba, PRIV_VERSION_1_11_0, ext_xtheadba), |
| ISA_EXT_DATA_ENTRY(xtheadbb, PRIV_VERSION_1_11_0, ext_xtheadbb), |
| ISA_EXT_DATA_ENTRY(xtheadbs, PRIV_VERSION_1_11_0, ext_xtheadbs), |
| ISA_EXT_DATA_ENTRY(xtheadcmo, PRIV_VERSION_1_11_0, ext_xtheadcmo), |
| ISA_EXT_DATA_ENTRY(xtheadcondmov, PRIV_VERSION_1_11_0, ext_xtheadcondmov), |
| ISA_EXT_DATA_ENTRY(xtheadfmemidx, PRIV_VERSION_1_11_0, ext_xtheadfmemidx), |
| ISA_EXT_DATA_ENTRY(xtheadfmv, PRIV_VERSION_1_11_0, ext_xtheadfmv), |
| ISA_EXT_DATA_ENTRY(xtheadmac, PRIV_VERSION_1_11_0, ext_xtheadmac), |
| ISA_EXT_DATA_ENTRY(xtheadmemidx, PRIV_VERSION_1_11_0, ext_xtheadmemidx), |
| ISA_EXT_DATA_ENTRY(xtheadmempair, PRIV_VERSION_1_11_0, ext_xtheadmempair), |
| ISA_EXT_DATA_ENTRY(xtheadsync, PRIV_VERSION_1_11_0, ext_xtheadsync), |
| ISA_EXT_DATA_ENTRY(xventanacondops, PRIV_VERSION_1_12_0, ext_XVentanaCondOps), |
| |
| DEFINE_PROP_END_OF_LIST(), |
| }; |
| |
| bool isa_ext_is_enabled(RISCVCPU *cpu, uint32_t ext_offset) |
| { |
| bool *ext_enabled = (void *)&cpu->cfg + ext_offset; |
| |
| return *ext_enabled; |
| } |
| |
| void isa_ext_update_enabled(RISCVCPU *cpu, uint32_t ext_offset, bool en) |
| { |
| bool *ext_enabled = (void *)&cpu->cfg + ext_offset; |
| |
| *ext_enabled = en; |
| } |
| |
| bool riscv_cpu_is_vendor(Object *cpu_obj) |
| { |
| return object_dynamic_cast(cpu_obj, TYPE_RISCV_VENDOR_CPU) != NULL; |
| } |
| |
| const char * const riscv_int_regnames[] = { |
| "x0/zero", "x1/ra", "x2/sp", "x3/gp", "x4/tp", "x5/t0", "x6/t1", |
| "x7/t2", "x8/s0", "x9/s1", "x10/a0", "x11/a1", "x12/a2", "x13/a3", |
| "x14/a4", "x15/a5", "x16/a6", "x17/a7", "x18/s2", "x19/s3", "x20/s4", |
| "x21/s5", "x22/s6", "x23/s7", "x24/s8", "x25/s9", "x26/s10", "x27/s11", |
| "x28/t3", "x29/t4", "x30/t5", "x31/t6" |
| }; |
| |
| const char * const riscv_int_regnamesh[] = { |
| "x0h/zeroh", "x1h/rah", "x2h/sph", "x3h/gph", "x4h/tph", "x5h/t0h", |
| "x6h/t1h", "x7h/t2h", "x8h/s0h", "x9h/s1h", "x10h/a0h", "x11h/a1h", |
| "x12h/a2h", "x13h/a3h", "x14h/a4h", "x15h/a5h", "x16h/a6h", "x17h/a7h", |
| "x18h/s2h", "x19h/s3h", "x20h/s4h", "x21h/s5h", "x22h/s6h", "x23h/s7h", |
| "x24h/s8h", "x25h/s9h", "x26h/s10h", "x27h/s11h", "x28h/t3h", "x29h/t4h", |
| "x30h/t5h", "x31h/t6h" |
| }; |
| |
| const char * const riscv_fpr_regnames[] = { |
| "f0/ft0", "f1/ft1", "f2/ft2", "f3/ft3", "f4/ft4", "f5/ft5", |
| "f6/ft6", "f7/ft7", "f8/fs0", "f9/fs1", "f10/fa0", "f11/fa1", |
| "f12/fa2", "f13/fa3", "f14/fa4", "f15/fa5", "f16/fa6", "f17/fa7", |
| "f18/fs2", "f19/fs3", "f20/fs4", "f21/fs5", "f22/fs6", "f23/fs7", |
| "f24/fs8", "f25/fs9", "f26/fs10", "f27/fs11", "f28/ft8", "f29/ft9", |
| "f30/ft10", "f31/ft11" |
| }; |
| |
| const char * const riscv_rvv_regnames[] = { |
| "v0", "v1", "v2", "v3", "v4", "v5", "v6", |
| "v7", "v8", "v9", "v10", "v11", "v12", "v13", |
| "v14", "v15", "v16", "v17", "v18", "v19", "v20", |
| "v21", "v22", "v23", "v24", "v25", "v26", "v27", |
| "v28", "v29", "v30", "v31" |
| }; |
| |
| static const char * const riscv_excp_names[] = { |
| "misaligned_fetch", |
| "fault_fetch", |
| "illegal_instruction", |
| "breakpoint", |
| "misaligned_load", |
| "fault_load", |
| "misaligned_store", |
| "fault_store", |
| "user_ecall", |
| "supervisor_ecall", |
| "hypervisor_ecall", |
| "machine_ecall", |
| "exec_page_fault", |
| "load_page_fault", |
| "reserved", |
| "store_page_fault", |
| "reserved", |
| "reserved", |
| "reserved", |
| "reserved", |
| "guest_exec_page_fault", |
| "guest_load_page_fault", |
| "reserved", |
| "guest_store_page_fault", |
| }; |
| |
| static const char * const riscv_intr_names[] = { |
| "u_software", |
| "s_software", |
| "vs_software", |
| "m_software", |
| "u_timer", |
| "s_timer", |
| "vs_timer", |
| "m_timer", |
| "u_external", |
| "s_external", |
| "vs_external", |
| "m_external", |
| "reserved", |
| "reserved", |
| "reserved", |
| "reserved" |
| }; |
| |
| const char *riscv_cpu_get_trap_name(target_ulong cause, bool async) |
| { |
| if (async) { |
| return (cause < ARRAY_SIZE(riscv_intr_names)) ? |
| riscv_intr_names[cause] : "(unknown)"; |
| } else { |
| return (cause < ARRAY_SIZE(riscv_excp_names)) ? |
| riscv_excp_names[cause] : "(unknown)"; |
| } |
| } |
| |
| void riscv_cpu_set_misa_ext(CPURISCVState *env, uint32_t ext) |
| { |
| env->misa_ext_mask = env->misa_ext = ext; |
| } |
| |
| int riscv_cpu_max_xlen(RISCVCPUClass *mcc) |
| { |
| return 16 << mcc->misa_mxl_max; |
| } |
| |
| #ifndef CONFIG_USER_ONLY |
| static uint8_t satp_mode_from_str(const char *satp_mode_str) |
| { |
| if (!strncmp(satp_mode_str, "mbare", 5)) { |
| return VM_1_10_MBARE; |
| } |
| |
| if (!strncmp(satp_mode_str, "sv32", 4)) { |
| return VM_1_10_SV32; |
| } |
| |
| if (!strncmp(satp_mode_str, "sv39", 4)) { |
| return VM_1_10_SV39; |
| } |
| |
| if (!strncmp(satp_mode_str, "sv48", 4)) { |
| return VM_1_10_SV48; |
| } |
| |
| if (!strncmp(satp_mode_str, "sv57", 4)) { |
| return VM_1_10_SV57; |
| } |
| |
| if (!strncmp(satp_mode_str, "sv64", 4)) { |
| return VM_1_10_SV64; |
| } |
| |
| g_assert_not_reached(); |
| } |
| |
| uint8_t satp_mode_max_from_map(uint32_t map) |
| { |
| /* |
| * 'map = 0' will make us return (31 - 32), which C will |
| * happily overflow to UINT_MAX. There's no good result to |
| * return if 'map = 0' (e.g. returning 0 will be ambiguous |
| * with the result for 'map = 1'). |
| * |
| * Assert out if map = 0. Callers will have to deal with |
| * it outside of this function. |
| */ |
| g_assert(map > 0); |
| |
| /* map here has at least one bit set, so no problem with clz */ |
| return 31 - __builtin_clz(map); |
| } |
| |
| const char *satp_mode_str(uint8_t satp_mode, bool is_32_bit) |
| { |
| if (is_32_bit) { |
| switch (satp_mode) { |
| case VM_1_10_SV32: |
| return "sv32"; |
| case VM_1_10_MBARE: |
| return "none"; |
| } |
| } else { |
| switch (satp_mode) { |
| case VM_1_10_SV64: |
| return "sv64"; |
| case VM_1_10_SV57: |
| return "sv57"; |
| case VM_1_10_SV48: |
| return "sv48"; |
| case VM_1_10_SV39: |
| return "sv39"; |
| case VM_1_10_MBARE: |
| return "none"; |
| } |
| } |
| |
| g_assert_not_reached(); |
| } |
| |
| static void set_satp_mode_max_supported(RISCVCPU *cpu, |
| uint8_t satp_mode) |
| { |
| bool rv32 = riscv_cpu_mxl(&cpu->env) == MXL_RV32; |
| const bool *valid_vm = rv32 ? valid_vm_1_10_32 : valid_vm_1_10_64; |
| |
| for (int i = 0; i <= satp_mode; ++i) { |
| if (valid_vm[i]) { |
| cpu->cfg.satp_mode.supported |= (1 << i); |
| } |
| } |
| } |
| |
| /* Set the satp mode to the max supported */ |
| static void set_satp_mode_default_map(RISCVCPU *cpu) |
| { |
| /* |
| * Bare CPUs do not default to the max available. |
| * Users must set a valid satp_mode in the command |
| * line. |
| */ |
| if (object_dynamic_cast(OBJECT(cpu), TYPE_RISCV_BARE_CPU) != NULL) { |
| warn_report("No satp mode set. Defaulting to 'bare'"); |
| cpu->cfg.satp_mode.map = (1 << VM_1_10_MBARE); |
| return; |
| } |
| |
| cpu->cfg.satp_mode.map = cpu->cfg.satp_mode.supported; |
| } |
| #endif |
| |
| static void riscv_max_cpu_init(Object *obj) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| CPURISCVState *env = &cpu->env; |
| |
| cpu->cfg.mmu = true; |
| cpu->cfg.pmp = true; |
| |
| env->priv_ver = PRIV_VERSION_LATEST; |
| #ifndef CONFIG_USER_ONLY |
| #ifdef TARGET_RISCV32 |
| set_satp_mode_max_supported(cpu, VM_1_10_SV32); |
| #else |
| set_satp_mode_max_supported(cpu, VM_1_10_SV57); |
| #endif |
| #endif |
| } |
| |
| #if defined(TARGET_RISCV64) |
| static void rv64_base_cpu_init(Object *obj) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| CPURISCVState *env = &cpu->env; |
| |
| cpu->cfg.mmu = true; |
| cpu->cfg.pmp = true; |
| |
| /* Set latest version of privileged specification */ |
| env->priv_ver = PRIV_VERSION_LATEST; |
| #ifndef CONFIG_USER_ONLY |
| set_satp_mode_max_supported(RISCV_CPU(obj), VM_1_10_SV57); |
| #endif |
| } |
| |
| static void rv64_sifive_u_cpu_init(Object *obj) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| CPURISCVState *env = &cpu->env; |
| riscv_cpu_set_misa_ext(env, RVI | RVM | RVA | RVF | RVD | RVC | RVS | RVU); |
| env->priv_ver = PRIV_VERSION_1_10_0; |
| #ifndef CONFIG_USER_ONLY |
| set_satp_mode_max_supported(RISCV_CPU(obj), VM_1_10_SV39); |
| #endif |
| |
| /* inherited from parent obj via riscv_cpu_init() */ |
| cpu->cfg.ext_zifencei = true; |
| cpu->cfg.ext_zicsr = true; |
| cpu->cfg.mmu = true; |
| cpu->cfg.pmp = true; |
| } |
| |
| static void rv64_sifive_e_cpu_init(Object *obj) |
| { |
| CPURISCVState *env = &RISCV_CPU(obj)->env; |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| |
| riscv_cpu_set_misa_ext(env, RVI | RVM | RVA | RVC | RVU); |
| env->priv_ver = PRIV_VERSION_1_10_0; |
| #ifndef CONFIG_USER_ONLY |
| set_satp_mode_max_supported(cpu, VM_1_10_MBARE); |
| #endif |
| |
| /* inherited from parent obj via riscv_cpu_init() */ |
| cpu->cfg.ext_zifencei = true; |
| cpu->cfg.ext_zicsr = true; |
| cpu->cfg.pmp = true; |
| } |
| |
| static void rv64_thead_c906_cpu_init(Object *obj) |
| { |
| CPURISCVState *env = &RISCV_CPU(obj)->env; |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| |
| riscv_cpu_set_misa_ext(env, RVG | RVC | RVS | RVU); |
| env->priv_ver = PRIV_VERSION_1_11_0; |
| |
| cpu->cfg.ext_zfa = true; |
| cpu->cfg.ext_zfh = true; |
| cpu->cfg.mmu = true; |
| cpu->cfg.ext_xtheadba = true; |
| cpu->cfg.ext_xtheadbb = true; |
| cpu->cfg.ext_xtheadbs = true; |
| cpu->cfg.ext_xtheadcmo = true; |
| cpu->cfg.ext_xtheadcondmov = true; |
| cpu->cfg.ext_xtheadfmemidx = true; |
| cpu->cfg.ext_xtheadmac = true; |
| cpu->cfg.ext_xtheadmemidx = true; |
| cpu->cfg.ext_xtheadmempair = true; |
| cpu->cfg.ext_xtheadsync = true; |
| |
| cpu->cfg.mvendorid = THEAD_VENDOR_ID; |
| #ifndef CONFIG_USER_ONLY |
| set_satp_mode_max_supported(cpu, VM_1_10_SV39); |
| th_register_custom_csrs(cpu); |
| #endif |
| |
| /* inherited from parent obj via riscv_cpu_init() */ |
| cpu->cfg.pmp = true; |
| } |
| |
| static void rv64_veyron_v1_cpu_init(Object *obj) |
| { |
| CPURISCVState *env = &RISCV_CPU(obj)->env; |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| |
| riscv_cpu_set_misa_ext(env, RVG | RVC | RVS | RVU | RVH); |
| env->priv_ver = PRIV_VERSION_1_12_0; |
| |
| /* Enable ISA extensions */ |
| cpu->cfg.mmu = true; |
| cpu->cfg.ext_zifencei = true; |
| cpu->cfg.ext_zicsr = true; |
| cpu->cfg.pmp = true; |
| cpu->cfg.ext_zicbom = true; |
| cpu->cfg.cbom_blocksize = 64; |
| cpu->cfg.cboz_blocksize = 64; |
| cpu->cfg.ext_zicboz = true; |
| cpu->cfg.ext_smaia = true; |
| cpu->cfg.ext_ssaia = true; |
| cpu->cfg.ext_sscofpmf = true; |
| cpu->cfg.ext_sstc = true; |
| cpu->cfg.ext_svinval = true; |
| cpu->cfg.ext_svnapot = true; |
| cpu->cfg.ext_svpbmt = true; |
| cpu->cfg.ext_smstateen = true; |
| cpu->cfg.ext_zba = true; |
| cpu->cfg.ext_zbb = true; |
| cpu->cfg.ext_zbc = true; |
| cpu->cfg.ext_zbs = true; |
| cpu->cfg.ext_XVentanaCondOps = true; |
| |
| cpu->cfg.mvendorid = VEYRON_V1_MVENDORID; |
| cpu->cfg.marchid = VEYRON_V1_MARCHID; |
| cpu->cfg.mimpid = VEYRON_V1_MIMPID; |
| |
| #ifndef CONFIG_USER_ONLY |
| set_satp_mode_max_supported(cpu, VM_1_10_SV48); |
| #endif |
| } |
| |
| #ifdef CONFIG_TCG |
| static void rv128_base_cpu_init(Object *obj) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| CPURISCVState *env = &cpu->env; |
| |
| if (qemu_tcg_mttcg_enabled()) { |
| /* Missing 128-bit aligned atomics */ |
| error_report("128-bit RISC-V currently does not work with Multi " |
| "Threaded TCG. Please use: -accel tcg,thread=single"); |
| exit(EXIT_FAILURE); |
| } |
| |
| cpu->cfg.mmu = true; |
| cpu->cfg.pmp = true; |
| |
| /* Set latest version of privileged specification */ |
| env->priv_ver = PRIV_VERSION_LATEST; |
| #ifndef CONFIG_USER_ONLY |
| set_satp_mode_max_supported(RISCV_CPU(obj), VM_1_10_SV57); |
| #endif |
| } |
| #endif /* CONFIG_TCG */ |
| |
| static void rv64i_bare_cpu_init(Object *obj) |
| { |
| CPURISCVState *env = &RISCV_CPU(obj)->env; |
| riscv_cpu_set_misa_ext(env, RVI); |
| } |
| |
| static void rv64e_bare_cpu_init(Object *obj) |
| { |
| CPURISCVState *env = &RISCV_CPU(obj)->env; |
| riscv_cpu_set_misa_ext(env, RVE); |
| } |
| |
| #else /* !TARGET_RISCV64 */ |
| |
| static void rv32_base_cpu_init(Object *obj) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| CPURISCVState *env = &cpu->env; |
| |
| cpu->cfg.mmu = true; |
| cpu->cfg.pmp = true; |
| |
| /* Set latest version of privileged specification */ |
| env->priv_ver = PRIV_VERSION_LATEST; |
| #ifndef CONFIG_USER_ONLY |
| set_satp_mode_max_supported(RISCV_CPU(obj), VM_1_10_SV32); |
| #endif |
| } |
| |
| static void rv32_sifive_u_cpu_init(Object *obj) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| CPURISCVState *env = &cpu->env; |
| riscv_cpu_set_misa_ext(env, RVI | RVM | RVA | RVF | RVD | RVC | RVS | RVU); |
| env->priv_ver = PRIV_VERSION_1_10_0; |
| #ifndef CONFIG_USER_ONLY |
| set_satp_mode_max_supported(RISCV_CPU(obj), VM_1_10_SV32); |
| #endif |
| |
| /* inherited from parent obj via riscv_cpu_init() */ |
| cpu->cfg.ext_zifencei = true; |
| cpu->cfg.ext_zicsr = true; |
| cpu->cfg.mmu = true; |
| cpu->cfg.pmp = true; |
| } |
| |
| static void rv32_sifive_e_cpu_init(Object *obj) |
| { |
| CPURISCVState *env = &RISCV_CPU(obj)->env; |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| |
| riscv_cpu_set_misa_ext(env, RVI | RVM | RVA | RVC | RVU); |
| env->priv_ver = PRIV_VERSION_1_10_0; |
| #ifndef CONFIG_USER_ONLY |
| set_satp_mode_max_supported(cpu, VM_1_10_MBARE); |
| #endif |
| |
| /* inherited from parent obj via riscv_cpu_init() */ |
| cpu->cfg.ext_zifencei = true; |
| cpu->cfg.ext_zicsr = true; |
| cpu->cfg.pmp = true; |
| } |
| |
| static void rv32_ibex_cpu_init(Object *obj) |
| { |
| CPURISCVState *env = &RISCV_CPU(obj)->env; |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| |
| riscv_cpu_set_misa_ext(env, RVI | RVM | RVC | RVU); |
| env->priv_ver = PRIV_VERSION_1_12_0; |
| #ifndef CONFIG_USER_ONLY |
| set_satp_mode_max_supported(cpu, VM_1_10_MBARE); |
| #endif |
| /* inherited from parent obj via riscv_cpu_init() */ |
| cpu->cfg.ext_zifencei = true; |
| cpu->cfg.ext_zicsr = true; |
| cpu->cfg.pmp = true; |
| cpu->cfg.ext_smepmp = true; |
| } |
| |
| static void rv32_imafcu_nommu_cpu_init(Object *obj) |
| { |
| CPURISCVState *env = &RISCV_CPU(obj)->env; |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| |
| riscv_cpu_set_misa_ext(env, RVI | RVM | RVA | RVF | RVC | RVU); |
| env->priv_ver = PRIV_VERSION_1_10_0; |
| #ifndef CONFIG_USER_ONLY |
| set_satp_mode_max_supported(cpu, VM_1_10_MBARE); |
| #endif |
| |
| /* inherited from parent obj via riscv_cpu_init() */ |
| cpu->cfg.ext_zifencei = true; |
| cpu->cfg.ext_zicsr = true; |
| cpu->cfg.pmp = true; |
| } |
| |
| static void rv32i_bare_cpu_init(Object *obj) |
| { |
| CPURISCVState *env = &RISCV_CPU(obj)->env; |
| riscv_cpu_set_misa_ext(env, RVI); |
| } |
| |
| static void rv32e_bare_cpu_init(Object *obj) |
| { |
| CPURISCVState *env = &RISCV_CPU(obj)->env; |
| riscv_cpu_set_misa_ext(env, RVE); |
| } |
| #endif |
| |
| static ObjectClass *riscv_cpu_class_by_name(const char *cpu_model) |
| { |
| ObjectClass *oc; |
| char *typename; |
| char **cpuname; |
| |
| cpuname = g_strsplit(cpu_model, ",", 1); |
| typename = g_strdup_printf(RISCV_CPU_TYPE_NAME("%s"), cpuname[0]); |
| oc = object_class_by_name(typename); |
| g_strfreev(cpuname); |
| g_free(typename); |
| |
| return oc; |
| } |
| |
| char *riscv_cpu_get_name(RISCVCPU *cpu) |
| { |
| RISCVCPUClass *rcc = RISCV_CPU_GET_CLASS(cpu); |
| const char *typename = object_class_get_name(OBJECT_CLASS(rcc)); |
| |
| g_assert(g_str_has_suffix(typename, RISCV_CPU_TYPE_SUFFIX)); |
| |
| return cpu_model_from_type(typename); |
| } |
| |
| static void riscv_cpu_dump_state(CPUState *cs, FILE *f, int flags) |
| { |
| RISCVCPU *cpu = RISCV_CPU(cs); |
| CPURISCVState *env = &cpu->env; |
| int i, j; |
| uint8_t *p; |
| |
| #if !defined(CONFIG_USER_ONLY) |
| if (riscv_has_ext(env, RVH)) { |
| qemu_fprintf(f, " %s %d\n", "V = ", env->virt_enabled); |
| } |
| #endif |
| qemu_fprintf(f, " %s " TARGET_FMT_lx "\n", "pc ", env->pc); |
| #ifndef CONFIG_USER_ONLY |
| { |
| static const int dump_csrs[] = { |
| CSR_MHARTID, |
| CSR_MSTATUS, |
| CSR_MSTATUSH, |
| /* |
| * CSR_SSTATUS is intentionally omitted here as its value |
| * can be figured out by looking at CSR_MSTATUS |
| */ |
| CSR_HSTATUS, |
| CSR_VSSTATUS, |
| CSR_MIP, |
| CSR_MIE, |
| CSR_MIDELEG, |
| CSR_HIDELEG, |
| CSR_MEDELEG, |
| CSR_HEDELEG, |
| CSR_MTVEC, |
| CSR_STVEC, |
| CSR_VSTVEC, |
| CSR_MEPC, |
| CSR_SEPC, |
| CSR_VSEPC, |
| CSR_MCAUSE, |
| CSR_SCAUSE, |
| CSR_VSCAUSE, |
| CSR_MTVAL, |
| CSR_STVAL, |
| CSR_HTVAL, |
| CSR_MTVAL2, |
| CSR_MSCRATCH, |
| CSR_SSCRATCH, |
| CSR_SATP, |
| CSR_MMTE, |
| CSR_UPMBASE, |
| CSR_UPMMASK, |
| CSR_SPMBASE, |
| CSR_SPMMASK, |
| CSR_MPMBASE, |
| CSR_MPMMASK, |
| }; |
| |
| for (i = 0; i < ARRAY_SIZE(dump_csrs); ++i) { |
| int csrno = dump_csrs[i]; |
| target_ulong val = 0; |
| RISCVException res = riscv_csrrw_debug(env, csrno, &val, 0, 0); |
| |
| /* |
| * Rely on the smode, hmode, etc, predicates within csr.c |
| * to do the filtering of the registers that are present. |
| */ |
| if (res == RISCV_EXCP_NONE) { |
| qemu_fprintf(f, " %-8s " TARGET_FMT_lx "\n", |
| csr_ops[csrno].name, val); |
| } |
| } |
| } |
| #endif |
| |
| for (i = 0; i < 32; i++) { |
| qemu_fprintf(f, " %-8s " TARGET_FMT_lx, |
| riscv_int_regnames[i], env->gpr[i]); |
| if ((i & 3) == 3) { |
| qemu_fprintf(f, "\n"); |
| } |
| } |
| if (flags & CPU_DUMP_FPU) { |
| for (i = 0; i < 32; i++) { |
| qemu_fprintf(f, " %-8s %016" PRIx64, |
| riscv_fpr_regnames[i], env->fpr[i]); |
| if ((i & 3) == 3) { |
| qemu_fprintf(f, "\n"); |
| } |
| } |
| } |
| if (riscv_has_ext(env, RVV) && (flags & CPU_DUMP_VPU)) { |
| static const int dump_rvv_csrs[] = { |
| CSR_VSTART, |
| CSR_VXSAT, |
| CSR_VXRM, |
| CSR_VCSR, |
| CSR_VL, |
| CSR_VTYPE, |
| CSR_VLENB, |
| }; |
| for (i = 0; i < ARRAY_SIZE(dump_rvv_csrs); ++i) { |
| int csrno = dump_rvv_csrs[i]; |
| target_ulong val = 0; |
| RISCVException res = riscv_csrrw_debug(env, csrno, &val, 0, 0); |
| |
| /* |
| * Rely on the smode, hmode, etc, predicates within csr.c |
| * to do the filtering of the registers that are present. |
| */ |
| if (res == RISCV_EXCP_NONE) { |
| qemu_fprintf(f, " %-8s " TARGET_FMT_lx "\n", |
| csr_ops[csrno].name, val); |
| } |
| } |
| uint16_t vlenb = cpu->cfg.vlenb; |
| |
| for (i = 0; i < 32; i++) { |
| qemu_fprintf(f, " %-8s ", riscv_rvv_regnames[i]); |
| p = (uint8_t *)env->vreg; |
| for (j = vlenb - 1 ; j >= 0; j--) { |
| qemu_fprintf(f, "%02x", *(p + i * vlenb + BYTE(j))); |
| } |
| qemu_fprintf(f, "\n"); |
| } |
| } |
| } |
| |
| static void riscv_cpu_set_pc(CPUState *cs, vaddr value) |
| { |
| RISCVCPU *cpu = RISCV_CPU(cs); |
| CPURISCVState *env = &cpu->env; |
| |
| if (env->xl == MXL_RV32) { |
| env->pc = (int32_t)value; |
| } else { |
| env->pc = value; |
| } |
| } |
| |
| static vaddr riscv_cpu_get_pc(CPUState *cs) |
| { |
| RISCVCPU *cpu = RISCV_CPU(cs); |
| CPURISCVState *env = &cpu->env; |
| |
| /* Match cpu_get_tb_cpu_state. */ |
| if (env->xl == MXL_RV32) { |
| return env->pc & UINT32_MAX; |
| } |
| return env->pc; |
| } |
| |
| bool riscv_cpu_has_work(CPUState *cs) |
| { |
| #ifndef CONFIG_USER_ONLY |
| RISCVCPU *cpu = RISCV_CPU(cs); |
| CPURISCVState *env = &cpu->env; |
| /* |
| * Definition of the WFI instruction requires it to ignore the privilege |
| * mode and delegation registers, but respect individual enables |
| */ |
| return riscv_cpu_all_pending(env) != 0 || |
| riscv_cpu_sirq_pending(env) != RISCV_EXCP_NONE || |
| riscv_cpu_vsirq_pending(env) != RISCV_EXCP_NONE; |
| #else |
| return true; |
| #endif |
| } |
| |
| static int riscv_cpu_mmu_index(CPUState *cs, bool ifetch) |
| { |
| return riscv_env_mmu_index(cpu_env(cs), ifetch); |
| } |
| |
| static void riscv_cpu_reset_hold(Object *obj, ResetType type) |
| { |
| #ifndef CONFIG_USER_ONLY |
| uint8_t iprio; |
| int i, irq, rdzero; |
| #endif |
| CPUState *cs = CPU(obj); |
| RISCVCPU *cpu = RISCV_CPU(cs); |
| RISCVCPUClass *mcc = RISCV_CPU_GET_CLASS(obj); |
| CPURISCVState *env = &cpu->env; |
| |
| if (mcc->parent_phases.hold) { |
| mcc->parent_phases.hold(obj, type); |
| } |
| #ifndef CONFIG_USER_ONLY |
| env->misa_mxl = mcc->misa_mxl_max; |
| env->priv = PRV_M; |
| env->mstatus &= ~(MSTATUS_MIE | MSTATUS_MPRV); |
| if (env->misa_mxl > MXL_RV32) { |
| /* |
| * The reset status of SXL/UXL is undefined, but mstatus is WARL |
| * and we must ensure that the value after init is valid for read. |
| */ |
| env->mstatus = set_field(env->mstatus, MSTATUS64_SXL, env->misa_mxl); |
| env->mstatus = set_field(env->mstatus, MSTATUS64_UXL, env->misa_mxl); |
| if (riscv_has_ext(env, RVH)) { |
| env->vsstatus = set_field(env->vsstatus, |
| MSTATUS64_SXL, env->misa_mxl); |
| env->vsstatus = set_field(env->vsstatus, |
| MSTATUS64_UXL, env->misa_mxl); |
| env->mstatus_hs = set_field(env->mstatus_hs, |
| MSTATUS64_SXL, env->misa_mxl); |
| env->mstatus_hs = set_field(env->mstatus_hs, |
| MSTATUS64_UXL, env->misa_mxl); |
| } |
| } |
| env->mcause = 0; |
| env->miclaim = MIP_SGEIP; |
| env->pc = env->resetvec; |
| env->bins = 0; |
| env->two_stage_lookup = false; |
| |
| env->menvcfg = (cpu->cfg.ext_svpbmt ? MENVCFG_PBMTE : 0) | |
| (!cpu->cfg.ext_svade && cpu->cfg.ext_svadu ? |
| MENVCFG_ADUE : 0); |
| env->henvcfg = 0; |
| |
| /* Initialized default priorities of local interrupts. */ |
| for (i = 0; i < ARRAY_SIZE(env->miprio); i++) { |
| iprio = riscv_cpu_default_priority(i); |
| env->miprio[i] = (i == IRQ_M_EXT) ? 0 : iprio; |
| env->siprio[i] = (i == IRQ_S_EXT) ? 0 : iprio; |
| env->hviprio[i] = 0; |
| } |
| i = 0; |
| while (!riscv_cpu_hviprio_index2irq(i, &irq, &rdzero)) { |
| if (!rdzero) { |
| env->hviprio[irq] = env->miprio[irq]; |
| } |
| i++; |
| } |
| /* mmte is supposed to have pm.current hardwired to 1 */ |
| env->mmte |= (EXT_STATUS_INITIAL | MMTE_M_PM_CURRENT); |
| |
| /* |
| * Bits 10, 6, 2 and 12 of mideleg are read only 1 when the Hypervisor |
| * extension is enabled. |
| */ |
| if (riscv_has_ext(env, RVH)) { |
| env->mideleg |= HS_MODE_INTERRUPTS; |
| } |
| |
| /* |
| * Clear mseccfg and unlock all the PMP entries upon reset. |
| * This is allowed as per the priv and smepmp specifications |
| * and is needed to clear stale entries across reboots. |
| */ |
| if (riscv_cpu_cfg(env)->ext_smepmp) { |
| env->mseccfg = 0; |
| } |
| |
| pmp_unlock_entries(env); |
| #endif |
| env->xl = riscv_cpu_mxl(env); |
| riscv_cpu_update_mask(env); |
| cs->exception_index = RISCV_EXCP_NONE; |
| env->load_res = -1; |
| set_default_nan_mode(1, &env->fp_status); |
| |
| #ifndef CONFIG_USER_ONLY |
| if (cpu->cfg.debug) { |
| riscv_trigger_reset_hold(env); |
| } |
| |
| if (kvm_enabled()) { |
| kvm_riscv_reset_vcpu(cpu); |
| } |
| #endif |
| } |
| |
| static void riscv_cpu_disas_set_info(CPUState *s, disassemble_info *info) |
| { |
| RISCVCPU *cpu = RISCV_CPU(s); |
| CPURISCVState *env = &cpu->env; |
| info->target_info = &cpu->cfg; |
| |
| switch (env->xl) { |
| case MXL_RV32: |
| info->print_insn = print_insn_riscv32; |
| break; |
| case MXL_RV64: |
| info->print_insn = print_insn_riscv64; |
| break; |
| case MXL_RV128: |
| info->print_insn = print_insn_riscv128; |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } |
| |
| #ifndef CONFIG_USER_ONLY |
| static void riscv_cpu_satp_mode_finalize(RISCVCPU *cpu, Error **errp) |
| { |
| bool rv32 = riscv_cpu_is_32bit(cpu); |
| uint8_t satp_mode_map_max, satp_mode_supported_max; |
| |
| /* The CPU wants the OS to decide which satp mode to use */ |
| if (cpu->cfg.satp_mode.supported == 0) { |
| return; |
| } |
| |
| satp_mode_supported_max = |
| satp_mode_max_from_map(cpu->cfg.satp_mode.supported); |
| |
| if (cpu->cfg.satp_mode.map == 0) { |
| if (cpu->cfg.satp_mode.init == 0) { |
| /* If unset by the user, we fallback to the default satp mode. */ |
| set_satp_mode_default_map(cpu); |
| } else { |
| /* |
| * Find the lowest level that was disabled and then enable the |
| * first valid level below which can be found in |
| * valid_vm_1_10_32/64. |
| */ |
| for (int i = 1; i < 16; ++i) { |
| if ((cpu->cfg.satp_mode.init & (1 << i)) && |
| (cpu->cfg.satp_mode.supported & (1 << i))) { |
| for (int j = i - 1; j >= 0; --j) { |
| if (cpu->cfg.satp_mode.supported & (1 << j)) { |
| cpu->cfg.satp_mode.map |= (1 << j); |
| break; |
| } |
| } |
| break; |
| } |
| } |
| } |
| } |
| |
| satp_mode_map_max = satp_mode_max_from_map(cpu->cfg.satp_mode.map); |
| |
| /* Make sure the user asked for a supported configuration (HW and qemu) */ |
| if (satp_mode_map_max > satp_mode_supported_max) { |
| error_setg(errp, "satp_mode %s is higher than hw max capability %s", |
| satp_mode_str(satp_mode_map_max, rv32), |
| satp_mode_str(satp_mode_supported_max, rv32)); |
| return; |
| } |
| |
| /* |
| * Make sure the user did not ask for an invalid configuration as per |
| * the specification. |
| */ |
| if (!rv32) { |
| for (int i = satp_mode_map_max - 1; i >= 0; --i) { |
| if (!(cpu->cfg.satp_mode.map & (1 << i)) && |
| (cpu->cfg.satp_mode.init & (1 << i)) && |
| (cpu->cfg.satp_mode.supported & (1 << i))) { |
| error_setg(errp, "cannot disable %s satp mode if %s " |
| "is enabled", satp_mode_str(i, false), |
| satp_mode_str(satp_mode_map_max, false)); |
| return; |
| } |
| } |
| } |
| |
| /* Finally expand the map so that all valid modes are set */ |
| for (int i = satp_mode_map_max - 1; i >= 0; --i) { |
| if (cpu->cfg.satp_mode.supported & (1 << i)) { |
| cpu->cfg.satp_mode.map |= (1 << i); |
| } |
| } |
| } |
| #endif |
| |
| void riscv_cpu_finalize_features(RISCVCPU *cpu, Error **errp) |
| { |
| Error *local_err = NULL; |
| |
| #ifndef CONFIG_USER_ONLY |
| riscv_cpu_satp_mode_finalize(cpu, &local_err); |
| if (local_err != NULL) { |
| error_propagate(errp, local_err); |
| return; |
| } |
| #endif |
| |
| if (tcg_enabled()) { |
| riscv_tcg_cpu_finalize_features(cpu, &local_err); |
| if (local_err != NULL) { |
| error_propagate(errp, local_err); |
| return; |
| } |
| riscv_tcg_cpu_finalize_dynamic_decoder(cpu); |
| } else if (kvm_enabled()) { |
| riscv_kvm_cpu_finalize_features(cpu, &local_err); |
| if (local_err != NULL) { |
| error_propagate(errp, local_err); |
| return; |
| } |
| } |
| } |
| |
| static void riscv_cpu_realize(DeviceState *dev, Error **errp) |
| { |
| CPUState *cs = CPU(dev); |
| RISCVCPU *cpu = RISCV_CPU(dev); |
| RISCVCPUClass *mcc = RISCV_CPU_GET_CLASS(dev); |
| Error *local_err = NULL; |
| |
| cpu_exec_realizefn(cs, &local_err); |
| if (local_err != NULL) { |
| error_propagate(errp, local_err); |
| return; |
| } |
| |
| riscv_cpu_finalize_features(cpu, &local_err); |
| if (local_err != NULL) { |
| error_propagate(errp, local_err); |
| return; |
| } |
| |
| riscv_cpu_register_gdb_regs_for_features(cs); |
| |
| #ifndef CONFIG_USER_ONLY |
| if (cpu->cfg.debug) { |
| riscv_trigger_realize(&cpu->env); |
| } |
| #endif |
| |
| qemu_init_vcpu(cs); |
| cpu_reset(cs); |
| |
| mcc->parent_realize(dev, errp); |
| } |
| |
| bool riscv_cpu_accelerator_compatible(RISCVCPU *cpu) |
| { |
| if (tcg_enabled()) { |
| return riscv_cpu_tcg_compatible(cpu); |
| } |
| |
| return true; |
| } |
| |
| #ifndef CONFIG_USER_ONLY |
| static void cpu_riscv_get_satp(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| RISCVSATPMap *satp_map = opaque; |
| uint8_t satp = satp_mode_from_str(name); |
| bool value; |
| |
| value = satp_map->map & (1 << satp); |
| |
| visit_type_bool(v, name, &value, errp); |
| } |
| |
| static void cpu_riscv_set_satp(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| RISCVSATPMap *satp_map = opaque; |
| uint8_t satp = satp_mode_from_str(name); |
| bool value; |
| |
| if (!visit_type_bool(v, name, &value, errp)) { |
| return; |
| } |
| |
| satp_map->map = deposit32(satp_map->map, satp, 1, value); |
| satp_map->init |= 1 << satp; |
| } |
| |
| void riscv_add_satp_mode_properties(Object *obj) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| |
| if (cpu->env.misa_mxl == MXL_RV32) { |
| object_property_add(obj, "sv32", "bool", cpu_riscv_get_satp, |
| cpu_riscv_set_satp, NULL, &cpu->cfg.satp_mode); |
| } else { |
| object_property_add(obj, "sv39", "bool", cpu_riscv_get_satp, |
| cpu_riscv_set_satp, NULL, &cpu->cfg.satp_mode); |
| object_property_add(obj, "sv48", "bool", cpu_riscv_get_satp, |
| cpu_riscv_set_satp, NULL, &cpu->cfg.satp_mode); |
| object_property_add(obj, "sv57", "bool", cpu_riscv_get_satp, |
| cpu_riscv_set_satp, NULL, &cpu->cfg.satp_mode); |
| object_property_add(obj, "sv64", "bool", cpu_riscv_get_satp, |
| cpu_riscv_set_satp, NULL, &cpu->cfg.satp_mode); |
| } |
| } |
| |
| static void riscv_cpu_set_irq(void *opaque, int irq, int level) |
| { |
| RISCVCPU *cpu = RISCV_CPU(opaque); |
| CPURISCVState *env = &cpu->env; |
| |
| if (irq < IRQ_LOCAL_MAX) { |
| switch (irq) { |
| case IRQ_U_SOFT: |
| case IRQ_S_SOFT: |
| case IRQ_VS_SOFT: |
| case IRQ_M_SOFT: |
| case IRQ_U_TIMER: |
| case IRQ_S_TIMER: |
| case IRQ_VS_TIMER: |
| case IRQ_M_TIMER: |
| case IRQ_U_EXT: |
| case IRQ_VS_EXT: |
| case IRQ_M_EXT: |
| if (kvm_enabled()) { |
| kvm_riscv_set_irq(cpu, irq, level); |
| } else { |
| riscv_cpu_update_mip(env, 1 << irq, BOOL_TO_MASK(level)); |
| } |
| break; |
| case IRQ_S_EXT: |
| if (kvm_enabled()) { |
| kvm_riscv_set_irq(cpu, irq, level); |
| } else { |
| env->external_seip = level; |
| riscv_cpu_update_mip(env, 1 << irq, |
| BOOL_TO_MASK(level | env->software_seip)); |
| } |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } else if (irq < (IRQ_LOCAL_MAX + IRQ_LOCAL_GUEST_MAX)) { |
| /* Require H-extension for handling guest local interrupts */ |
| if (!riscv_has_ext(env, RVH)) { |
| g_assert_not_reached(); |
| } |
| |
| /* Compute bit position in HGEIP CSR */ |
| irq = irq - IRQ_LOCAL_MAX + 1; |
| if (env->geilen < irq) { |
| g_assert_not_reached(); |
| } |
| |
| /* Update HGEIP CSR */ |
| env->hgeip &= ~((target_ulong)1 << irq); |
| if (level) { |
| env->hgeip |= (target_ulong)1 << irq; |
| } |
| |
| /* Update mip.SGEIP bit */ |
| riscv_cpu_update_mip(env, MIP_SGEIP, |
| BOOL_TO_MASK(!!(env->hgeie & env->hgeip))); |
| } else { |
| g_assert_not_reached(); |
| } |
| } |
| #endif /* CONFIG_USER_ONLY */ |
| |
| static bool riscv_cpu_is_dynamic(Object *cpu_obj) |
| { |
| return object_dynamic_cast(cpu_obj, TYPE_RISCV_DYNAMIC_CPU) != NULL; |
| } |
| |
| static void riscv_cpu_post_init(Object *obj) |
| { |
| accel_cpu_instance_init(CPU(obj)); |
| } |
| |
| static void riscv_cpu_init(Object *obj) |
| { |
| RISCVCPUClass *mcc = RISCV_CPU_GET_CLASS(obj); |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| CPURISCVState *env = &cpu->env; |
| |
| env->misa_mxl = mcc->misa_mxl_max; |
| |
| #ifndef CONFIG_USER_ONLY |
| qdev_init_gpio_in(DEVICE(obj), riscv_cpu_set_irq, |
| IRQ_LOCAL_MAX + IRQ_LOCAL_GUEST_MAX); |
| #endif /* CONFIG_USER_ONLY */ |
| |
| general_user_opts = g_hash_table_new(g_str_hash, g_str_equal); |
| |
| /* |
| * The timer and performance counters extensions were supported |
| * in QEMU before they were added as discrete extensions in the |
| * ISA. To keep compatibility we'll always default them to 'true' |
| * for all CPUs. Each accelerator will decide what to do when |
| * users disable them. |
| */ |
| RISCV_CPU(obj)->cfg.ext_zicntr = true; |
| RISCV_CPU(obj)->cfg.ext_zihpm = true; |
| |
| /* Default values for non-bool cpu properties */ |
| cpu->cfg.pmu_mask = MAKE_64BIT_MASK(3, 16); |
| cpu->cfg.vlenb = 128 >> 3; |
| cpu->cfg.elen = 64; |
| cpu->cfg.cbom_blocksize = 64; |
| cpu->cfg.cbop_blocksize = 64; |
| cpu->cfg.cboz_blocksize = 64; |
| cpu->env.vext_ver = VEXT_VERSION_1_00_0; |
| } |
| |
| static void riscv_bare_cpu_init(Object *obj) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| |
| /* |
| * Bare CPUs do not inherit the timer and performance |
| * counters from the parent class (see riscv_cpu_init() |
| * for info on why the parent enables them). |
| * |
| * Users have to explicitly enable these counters for |
| * bare CPUs. |
| */ |
| cpu->cfg.ext_zicntr = false; |
| cpu->cfg.ext_zihpm = false; |
| |
| /* Set to QEMU's first supported priv version */ |
| cpu->env.priv_ver = PRIV_VERSION_1_10_0; |
| |
| /* |
| * Support all available satp_mode settings. The default |
| * value will be set to MBARE if the user doesn't set |
| * satp_mode manually (see set_satp_mode_default()). |
| */ |
| #ifndef CONFIG_USER_ONLY |
| set_satp_mode_max_supported(cpu, VM_1_10_SV64); |
| #endif |
| } |
| |
| typedef struct misa_ext_info { |
| const char *name; |
| const char *description; |
| } MISAExtInfo; |
| |
| #define MISA_INFO_IDX(_bit) \ |
| __builtin_ctz(_bit) |
| |
| #define MISA_EXT_INFO(_bit, _propname, _descr) \ |
| [MISA_INFO_IDX(_bit)] = {.name = _propname, .description = _descr} |
| |
| static const MISAExtInfo misa_ext_info_arr[] = { |
| MISA_EXT_INFO(RVA, "a", "Atomic instructions"), |
| MISA_EXT_INFO(RVC, "c", "Compressed instructions"), |
| MISA_EXT_INFO(RVD, "d", "Double-precision float point"), |
| MISA_EXT_INFO(RVF, "f", "Single-precision float point"), |
| MISA_EXT_INFO(RVI, "i", "Base integer instruction set"), |
| MISA_EXT_INFO(RVE, "e", "Base integer instruction set (embedded)"), |
| MISA_EXT_INFO(RVM, "m", "Integer multiplication and division"), |
| MISA_EXT_INFO(RVS, "s", "Supervisor-level instructions"), |
| MISA_EXT_INFO(RVU, "u", "User-level instructions"), |
| MISA_EXT_INFO(RVH, "h", "Hypervisor"), |
| MISA_EXT_INFO(RVJ, "x-j", "Dynamic translated languages"), |
| MISA_EXT_INFO(RVV, "v", "Vector operations"), |
| MISA_EXT_INFO(RVG, "g", "General purpose (IMAFD_Zicsr_Zifencei)"), |
| MISA_EXT_INFO(RVB, "b", "Bit manipulation (Zba_Zbb_Zbs)") |
| }; |
| |
| static void riscv_cpu_validate_misa_mxl(RISCVCPUClass *mcc) |
| { |
| CPUClass *cc = CPU_CLASS(mcc); |
| |
| /* Validate that MISA_MXL is set properly. */ |
| switch (mcc->misa_mxl_max) { |
| #ifdef TARGET_RISCV64 |
| case MXL_RV64: |
| case MXL_RV128: |
| cc->gdb_core_xml_file = "riscv-64bit-cpu.xml"; |
| break; |
| #endif |
| case MXL_RV32: |
| cc->gdb_core_xml_file = "riscv-32bit-cpu.xml"; |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } |
| |
| static int riscv_validate_misa_info_idx(uint32_t bit) |
| { |
| int idx; |
| |
| /* |
| * Our lowest valid input (RVA) is 1 and |
| * __builtin_ctz() is UB with zero. |
| */ |
| g_assert(bit != 0); |
| idx = MISA_INFO_IDX(bit); |
| |
| g_assert(idx < ARRAY_SIZE(misa_ext_info_arr)); |
| return idx; |
| } |
| |
| const char *riscv_get_misa_ext_name(uint32_t bit) |
| { |
| int idx = riscv_validate_misa_info_idx(bit); |
| const char *val = misa_ext_info_arr[idx].name; |
| |
| g_assert(val != NULL); |
| return val; |
| } |
| |
| const char *riscv_get_misa_ext_description(uint32_t bit) |
| { |
| int idx = riscv_validate_misa_info_idx(bit); |
| const char *val = misa_ext_info_arr[idx].description; |
| |
| g_assert(val != NULL); |
| return val; |
| } |
| |
| #define MULTI_EXT_CFG_BOOL(_name, _prop, _defval) \ |
| {.name = _name, .offset = CPU_CFG_OFFSET(_prop), \ |
| .enabled = _defval} |
| |
| const RISCVCPUMultiExtConfig riscv_cpu_extensions[] = { |
| /* Defaults for standard extensions */ |
| MULTI_EXT_CFG_BOOL("sscofpmf", ext_sscofpmf, false), |
| MULTI_EXT_CFG_BOOL("smcntrpmf", ext_smcntrpmf, false), |
| MULTI_EXT_CFG_BOOL("zifencei", ext_zifencei, true), |
| MULTI_EXT_CFG_BOOL("zicsr", ext_zicsr, true), |
| MULTI_EXT_CFG_BOOL("zihintntl", ext_zihintntl, true), |
| MULTI_EXT_CFG_BOOL("zihintpause", ext_zihintpause, true), |
| MULTI_EXT_CFG_BOOL("zimop", ext_zimop, false), |
| MULTI_EXT_CFG_BOOL("zcmop", ext_zcmop, false), |
| MULTI_EXT_CFG_BOOL("zacas", ext_zacas, false), |
| MULTI_EXT_CFG_BOOL("zama16b", ext_zama16b, false), |
| MULTI_EXT_CFG_BOOL("zabha", ext_zabha, false), |
| MULTI_EXT_CFG_BOOL("zaamo", ext_zaamo, false), |
| MULTI_EXT_CFG_BOOL("zalrsc", ext_zalrsc, false), |
| MULTI_EXT_CFG_BOOL("zawrs", ext_zawrs, true), |
| MULTI_EXT_CFG_BOOL("zfa", ext_zfa, true), |
| MULTI_EXT_CFG_BOOL("zfbfmin", ext_zfbfmin, false), |
| MULTI_EXT_CFG_BOOL("zfh", ext_zfh, false), |
| MULTI_EXT_CFG_BOOL("zfhmin", ext_zfhmin, false), |
| MULTI_EXT_CFG_BOOL("zve32f", ext_zve32f, false), |
| MULTI_EXT_CFG_BOOL("zve32x", ext_zve32x, false), |
| MULTI_EXT_CFG_BOOL("zve64f", ext_zve64f, false), |
| MULTI_EXT_CFG_BOOL("zve64d", ext_zve64d, false), |
| MULTI_EXT_CFG_BOOL("zve64x", ext_zve64x, false), |
| MULTI_EXT_CFG_BOOL("zvfbfmin", ext_zvfbfmin, false), |
| MULTI_EXT_CFG_BOOL("zvfbfwma", ext_zvfbfwma, false), |
| MULTI_EXT_CFG_BOOL("zvfh", ext_zvfh, false), |
| MULTI_EXT_CFG_BOOL("zvfhmin", ext_zvfhmin, false), |
| MULTI_EXT_CFG_BOOL("sstc", ext_sstc, true), |
| |
| MULTI_EXT_CFG_BOOL("smaia", ext_smaia, false), |
| MULTI_EXT_CFG_BOOL("smepmp", ext_smepmp, false), |
| MULTI_EXT_CFG_BOOL("smstateen", ext_smstateen, false), |
| MULTI_EXT_CFG_BOOL("ssaia", ext_ssaia, false), |
| MULTI_EXT_CFG_BOOL("svade", ext_svade, false), |
| MULTI_EXT_CFG_BOOL("svadu", ext_svadu, true), |
| MULTI_EXT_CFG_BOOL("svinval", ext_svinval, false), |
| MULTI_EXT_CFG_BOOL("svnapot", ext_svnapot, false), |
| MULTI_EXT_CFG_BOOL("svpbmt", ext_svpbmt, false), |
| |
| MULTI_EXT_CFG_BOOL("zicntr", ext_zicntr, true), |
| MULTI_EXT_CFG_BOOL("zihpm", ext_zihpm, true), |
| |
| MULTI_EXT_CFG_BOOL("zba", ext_zba, true), |
| MULTI_EXT_CFG_BOOL("zbb", ext_zbb, true), |
| MULTI_EXT_CFG_BOOL("zbc", ext_zbc, true), |
| MULTI_EXT_CFG_BOOL("zbkb", ext_zbkb, false), |
| MULTI_EXT_CFG_BOOL("zbkc", ext_zbkc, false), |
| MULTI_EXT_CFG_BOOL("zbkx", ext_zbkx, false), |
| MULTI_EXT_CFG_BOOL("zbs", ext_zbs, true), |
| MULTI_EXT_CFG_BOOL("zk", ext_zk, false), |
| MULTI_EXT_CFG_BOOL("zkn", ext_zkn, false), |
| MULTI_EXT_CFG_BOOL("zknd", ext_zknd, false), |
| MULTI_EXT_CFG_BOOL("zkne", ext_zkne, false), |
| MULTI_EXT_CFG_BOOL("zknh", ext_zknh, false), |
| MULTI_EXT_CFG_BOOL("zkr", ext_zkr, false), |
| MULTI_EXT_CFG_BOOL("zks", ext_zks, false), |
| MULTI_EXT_CFG_BOOL("zksed", ext_zksed, false), |
| MULTI_EXT_CFG_BOOL("zksh", ext_zksh, false), |
| MULTI_EXT_CFG_BOOL("zkt", ext_zkt, false), |
| MULTI_EXT_CFG_BOOL("ztso", ext_ztso, false), |
| |
| MULTI_EXT_CFG_BOOL("zdinx", ext_zdinx, false), |
| MULTI_EXT_CFG_BOOL("zfinx", ext_zfinx, false), |
| MULTI_EXT_CFG_BOOL("zhinx", ext_zhinx, false), |
| MULTI_EXT_CFG_BOOL("zhinxmin", ext_zhinxmin, false), |
| |
| MULTI_EXT_CFG_BOOL("zicbom", ext_zicbom, true), |
| MULTI_EXT_CFG_BOOL("zicbop", ext_zicbop, true), |
| MULTI_EXT_CFG_BOOL("zicboz", ext_zicboz, true), |
| |
| MULTI_EXT_CFG_BOOL("zmmul", ext_zmmul, false), |
| |
| MULTI_EXT_CFG_BOOL("zca", ext_zca, false), |
| MULTI_EXT_CFG_BOOL("zcb", ext_zcb, false), |
| MULTI_EXT_CFG_BOOL("zcd", ext_zcd, false), |
| MULTI_EXT_CFG_BOOL("zce", ext_zce, false), |
| MULTI_EXT_CFG_BOOL("zcf", ext_zcf, false), |
| MULTI_EXT_CFG_BOOL("zcmp", ext_zcmp, false), |
| MULTI_EXT_CFG_BOOL("zcmt", ext_zcmt, false), |
| MULTI_EXT_CFG_BOOL("zicond", ext_zicond, false), |
| |
| /* Vector cryptography extensions */ |
| MULTI_EXT_CFG_BOOL("zvbb", ext_zvbb, false), |
| MULTI_EXT_CFG_BOOL("zvbc", ext_zvbc, false), |
| MULTI_EXT_CFG_BOOL("zvkb", ext_zvkb, false), |
| MULTI_EXT_CFG_BOOL("zvkg", ext_zvkg, false), |
| MULTI_EXT_CFG_BOOL("zvkned", ext_zvkned, false), |
| MULTI_EXT_CFG_BOOL("zvknha", ext_zvknha, false), |
| MULTI_EXT_CFG_BOOL("zvknhb", ext_zvknhb, false), |
| MULTI_EXT_CFG_BOOL("zvksed", ext_zvksed, false), |
| MULTI_EXT_CFG_BOOL("zvksh", ext_zvksh, false), |
| MULTI_EXT_CFG_BOOL("zvkt", ext_zvkt, false), |
| MULTI_EXT_CFG_BOOL("zvkn", ext_zvkn, false), |
| MULTI_EXT_CFG_BOOL("zvknc", ext_zvknc, false), |
| MULTI_EXT_CFG_BOOL("zvkng", ext_zvkng, false), |
| MULTI_EXT_CFG_BOOL("zvks", ext_zvks, false), |
| MULTI_EXT_CFG_BOOL("zvksc", ext_zvksc, false), |
| MULTI_EXT_CFG_BOOL("zvksg", ext_zvksg, false), |
| |
| DEFINE_PROP_END_OF_LIST(), |
| }; |
| |
| const RISCVCPUMultiExtConfig riscv_cpu_vendor_exts[] = { |
| MULTI_EXT_CFG_BOOL("xtheadba", ext_xtheadba, false), |
| MULTI_EXT_CFG_BOOL("xtheadbb", ext_xtheadbb, false), |
| MULTI_EXT_CFG_BOOL("xtheadbs", ext_xtheadbs, false), |
| MULTI_EXT_CFG_BOOL("xtheadcmo", ext_xtheadcmo, false), |
| MULTI_EXT_CFG_BOOL("xtheadcondmov", ext_xtheadcondmov, false), |
| MULTI_EXT_CFG_BOOL("xtheadfmemidx", ext_xtheadfmemidx, false), |
| MULTI_EXT_CFG_BOOL("xtheadfmv", ext_xtheadfmv, false), |
| MULTI_EXT_CFG_BOOL("xtheadmac", ext_xtheadmac, false), |
| MULTI_EXT_CFG_BOOL("xtheadmemidx", ext_xtheadmemidx, false), |
| MULTI_EXT_CFG_BOOL("xtheadmempair", ext_xtheadmempair, false), |
| MULTI_EXT_CFG_BOOL("xtheadsync", ext_xtheadsync, false), |
| MULTI_EXT_CFG_BOOL("xventanacondops", ext_XVentanaCondOps, false), |
| |
| DEFINE_PROP_END_OF_LIST(), |
| }; |
| |
| /* These are experimental so mark with 'x-' */ |
| const RISCVCPUMultiExtConfig riscv_cpu_experimental_exts[] = { |
| DEFINE_PROP_END_OF_LIST(), |
| }; |
| |
| /* |
| * 'Named features' is the name we give to extensions that we |
| * don't want to expose to users. They are either immutable |
| * (always enabled/disable) or they'll vary depending on |
| * the resulting CPU state. They have riscv,isa strings |
| * and priv_ver like regular extensions. |
| */ |
| const RISCVCPUMultiExtConfig riscv_cpu_named_features[] = { |
| MULTI_EXT_CFG_BOOL("zic64b", ext_zic64b, true), |
| |
| DEFINE_PROP_END_OF_LIST(), |
| }; |
| |
| /* Deprecated entries marked for future removal */ |
| const RISCVCPUMultiExtConfig riscv_cpu_deprecated_exts[] = { |
| MULTI_EXT_CFG_BOOL("Zifencei", ext_zifencei, true), |
| MULTI_EXT_CFG_BOOL("Zicsr", ext_zicsr, true), |
| MULTI_EXT_CFG_BOOL("Zihintntl", ext_zihintntl, true), |
| MULTI_EXT_CFG_BOOL("Zihintpause", ext_zihintpause, true), |
| MULTI_EXT_CFG_BOOL("Zawrs", ext_zawrs, true), |
| MULTI_EXT_CFG_BOOL("Zfa", ext_zfa, true), |
| MULTI_EXT_CFG_BOOL("Zfh", ext_zfh, false), |
| MULTI_EXT_CFG_BOOL("Zfhmin", ext_zfhmin, false), |
| MULTI_EXT_CFG_BOOL("Zve32f", ext_zve32f, false), |
| MULTI_EXT_CFG_BOOL("Zve64f", ext_zve64f, false), |
| MULTI_EXT_CFG_BOOL("Zve64d", ext_zve64d, false), |
| |
| DEFINE_PROP_END_OF_LIST(), |
| }; |
| |
| static void cpu_set_prop_err(RISCVCPU *cpu, const char *propname, |
| Error **errp) |
| { |
| g_autofree char *cpuname = riscv_cpu_get_name(cpu); |
| error_setg(errp, "CPU '%s' does not allow changing the value of '%s'", |
| cpuname, propname); |
| } |
| |
| static void prop_pmu_num_set(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| uint8_t pmu_num, curr_pmu_num; |
| uint32_t pmu_mask; |
| |
| visit_type_uint8(v, name, &pmu_num, errp); |
| |
| curr_pmu_num = ctpop32(cpu->cfg.pmu_mask); |
| |
| if (pmu_num != curr_pmu_num && riscv_cpu_is_vendor(obj)) { |
| cpu_set_prop_err(cpu, name, errp); |
| error_append_hint(errp, "Current '%s' val: %u\n", |
| name, curr_pmu_num); |
| return; |
| } |
| |
| if (pmu_num > (RV_MAX_MHPMCOUNTERS - 3)) { |
| error_setg(errp, "Number of counters exceeds maximum available"); |
| return; |
| } |
| |
| if (pmu_num == 0) { |
| pmu_mask = 0; |
| } else { |
| pmu_mask = MAKE_64BIT_MASK(3, pmu_num); |
| } |
| |
| warn_report("\"pmu-num\" property is deprecated; use \"pmu-mask\""); |
| cpu->cfg.pmu_mask = pmu_mask; |
| cpu_option_add_user_setting("pmu-mask", pmu_mask); |
| } |
| |
| static void prop_pmu_num_get(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| uint8_t pmu_num = ctpop32(cpu->cfg.pmu_mask); |
| |
| visit_type_uint8(v, name, &pmu_num, errp); |
| } |
| |
| static const PropertyInfo prop_pmu_num = { |
| .name = "pmu-num", |
| .get = prop_pmu_num_get, |
| .set = prop_pmu_num_set, |
| }; |
| |
| static void prop_pmu_mask_set(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| uint32_t value; |
| uint8_t pmu_num; |
| |
| visit_type_uint32(v, name, &value, errp); |
| |
| if (value != cpu->cfg.pmu_mask && riscv_cpu_is_vendor(obj)) { |
| cpu_set_prop_err(cpu, name, errp); |
| error_append_hint(errp, "Current '%s' val: %x\n", |
| name, cpu->cfg.pmu_mask); |
| return; |
| } |
| |
| pmu_num = ctpop32(value); |
| |
| if (pmu_num > (RV_MAX_MHPMCOUNTERS - 3)) { |
| error_setg(errp, "Number of counters exceeds maximum available"); |
| return; |
| } |
| |
| cpu_option_add_user_setting(name, value); |
| cpu->cfg.pmu_mask = value; |
| } |
| |
| static void prop_pmu_mask_get(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| uint8_t pmu_mask = RISCV_CPU(obj)->cfg.pmu_mask; |
| |
| visit_type_uint8(v, name, &pmu_mask, errp); |
| } |
| |
| static const PropertyInfo prop_pmu_mask = { |
| .name = "pmu-mask", |
| .get = prop_pmu_mask_get, |
| .set = prop_pmu_mask_set, |
| }; |
| |
| static void prop_mmu_set(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| bool value; |
| |
| visit_type_bool(v, name, &value, errp); |
| |
| if (cpu->cfg.mmu != value && riscv_cpu_is_vendor(obj)) { |
| cpu_set_prop_err(cpu, "mmu", errp); |
| return; |
| } |
| |
| cpu_option_add_user_setting(name, value); |
| cpu->cfg.mmu = value; |
| } |
| |
| static void prop_mmu_get(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| bool value = RISCV_CPU(obj)->cfg.mmu; |
| |
| visit_type_bool(v, name, &value, errp); |
| } |
| |
| static const PropertyInfo prop_mmu = { |
| .name = "mmu", |
| .get = prop_mmu_get, |
| .set = prop_mmu_set, |
| }; |
| |
| static void prop_pmp_set(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| bool value; |
| |
| visit_type_bool(v, name, &value, errp); |
| |
| if (cpu->cfg.pmp != value && riscv_cpu_is_vendor(obj)) { |
| cpu_set_prop_err(cpu, name, errp); |
| return; |
| } |
| |
| cpu_option_add_user_setting(name, value); |
| cpu->cfg.pmp = value; |
| } |
| |
| static void prop_pmp_get(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| bool value = RISCV_CPU(obj)->cfg.pmp; |
| |
| visit_type_bool(v, name, &value, errp); |
| } |
| |
| static const PropertyInfo prop_pmp = { |
| .name = "pmp", |
| .get = prop_pmp_get, |
| .set = prop_pmp_set, |
| }; |
| |
| static int priv_spec_from_str(const char *priv_spec_str) |
| { |
| int priv_version = -1; |
| |
| if (!g_strcmp0(priv_spec_str, PRIV_VER_1_13_0_STR)) { |
| priv_version = PRIV_VERSION_1_13_0; |
| } else if (!g_strcmp0(priv_spec_str, PRIV_VER_1_12_0_STR)) { |
| priv_version = PRIV_VERSION_1_12_0; |
| } else if (!g_strcmp0(priv_spec_str, PRIV_VER_1_11_0_STR)) { |
| priv_version = PRIV_VERSION_1_11_0; |
| } else if (!g_strcmp0(priv_spec_str, PRIV_VER_1_10_0_STR)) { |
| priv_version = PRIV_VERSION_1_10_0; |
| } |
| |
| return priv_version; |
| } |
| |
| const char *priv_spec_to_str(int priv_version) |
| { |
| switch (priv_version) { |
| case PRIV_VERSION_1_10_0: |
| return PRIV_VER_1_10_0_STR; |
| case PRIV_VERSION_1_11_0: |
| return PRIV_VER_1_11_0_STR; |
| case PRIV_VERSION_1_12_0: |
| return PRIV_VER_1_12_0_STR; |
| case PRIV_VERSION_1_13_0: |
| return PRIV_VER_1_13_0_STR; |
| default: |
| return NULL; |
| } |
| } |
| |
| static void prop_priv_spec_set(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| g_autofree char *value = NULL; |
| int priv_version = -1; |
| |
| visit_type_str(v, name, &value, errp); |
| |
| priv_version = priv_spec_from_str(value); |
| if (priv_version < 0) { |
| error_setg(errp, "Unsupported privilege spec version '%s'", value); |
| return; |
| } |
| |
| if (priv_version != cpu->env.priv_ver && riscv_cpu_is_vendor(obj)) { |
| cpu_set_prop_err(cpu, name, errp); |
| error_append_hint(errp, "Current '%s' val: %s\n", name, |
| object_property_get_str(obj, name, NULL)); |
| return; |
| } |
| |
| cpu_option_add_user_setting(name, priv_version); |
| cpu->env.priv_ver = priv_version; |
| } |
| |
| static void prop_priv_spec_get(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| const char *value = priv_spec_to_str(cpu->env.priv_ver); |
| |
| visit_type_str(v, name, (char **)&value, errp); |
| } |
| |
| static const PropertyInfo prop_priv_spec = { |
| .name = "priv_spec", |
| .get = prop_priv_spec_get, |
| .set = prop_priv_spec_set, |
| }; |
| |
| static void prop_vext_spec_set(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| g_autofree char *value = NULL; |
| |
| visit_type_str(v, name, &value, errp); |
| |
| if (g_strcmp0(value, VEXT_VER_1_00_0_STR) != 0) { |
| error_setg(errp, "Unsupported vector spec version '%s'", value); |
| return; |
| } |
| |
| cpu_option_add_user_setting(name, VEXT_VERSION_1_00_0); |
| cpu->env.vext_ver = VEXT_VERSION_1_00_0; |
| } |
| |
| static void prop_vext_spec_get(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| const char *value = VEXT_VER_1_00_0_STR; |
| |
| visit_type_str(v, name, (char **)&value, errp); |
| } |
| |
| static const PropertyInfo prop_vext_spec = { |
| .name = "vext_spec", |
| .get = prop_vext_spec_get, |
| .set = prop_vext_spec_set, |
| }; |
| |
| static void prop_vlen_set(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| uint16_t value; |
| |
| if (!visit_type_uint16(v, name, &value, errp)) { |
| return; |
| } |
| |
| if (!is_power_of_2(value)) { |
| error_setg(errp, "Vector extension VLEN must be power of 2"); |
| return; |
| } |
| |
| if (value != cpu->cfg.vlenb && riscv_cpu_is_vendor(obj)) { |
| cpu_set_prop_err(cpu, name, errp); |
| error_append_hint(errp, "Current '%s' val: %u\n", |
| name, cpu->cfg.vlenb << 3); |
| return; |
| } |
| |
| cpu_option_add_user_setting(name, value); |
| cpu->cfg.vlenb = value >> 3; |
| } |
| |
| static void prop_vlen_get(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| uint16_t value = RISCV_CPU(obj)->cfg.vlenb << 3; |
| |
| visit_type_uint16(v, name, &value, errp); |
| } |
| |
| static const PropertyInfo prop_vlen = { |
| .name = "vlen", |
| .get = prop_vlen_get, |
| .set = prop_vlen_set, |
| }; |
| |
| static void prop_elen_set(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| uint16_t value; |
| |
| if (!visit_type_uint16(v, name, &value, errp)) { |
| return; |
| } |
| |
| if (!is_power_of_2(value)) { |
| error_setg(errp, "Vector extension ELEN must be power of 2"); |
| return; |
| } |
| |
| if (value != cpu->cfg.elen && riscv_cpu_is_vendor(obj)) { |
| cpu_set_prop_err(cpu, name, errp); |
| error_append_hint(errp, "Current '%s' val: %u\n", |
| name, cpu->cfg.elen); |
| return; |
| } |
| |
| cpu_option_add_user_setting(name, value); |
| cpu->cfg.elen = value; |
| } |
| |
| static void prop_elen_get(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| uint16_t value = RISCV_CPU(obj)->cfg.elen; |
| |
| visit_type_uint16(v, name, &value, errp); |
| } |
| |
| static const PropertyInfo prop_elen = { |
| .name = "elen", |
| .get = prop_elen_get, |
| .set = prop_elen_set, |
| }; |
| |
| static void prop_cbom_blksize_set(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| uint16_t value; |
| |
| if (!visit_type_uint16(v, name, &value, errp)) { |
| return; |
| } |
| |
| if (value != cpu->cfg.cbom_blocksize && riscv_cpu_is_vendor(obj)) { |
| cpu_set_prop_err(cpu, name, errp); |
| error_append_hint(errp, "Current '%s' val: %u\n", |
| name, cpu->cfg.cbom_blocksize); |
| return; |
| } |
| |
| cpu_option_add_user_setting(name, value); |
| cpu->cfg.cbom_blocksize = value; |
| } |
| |
| static void prop_cbom_blksize_get(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| uint16_t value = RISCV_CPU(obj)->cfg.cbom_blocksize; |
| |
| visit_type_uint16(v, name, &value, errp); |
| } |
| |
| static const PropertyInfo prop_cbom_blksize = { |
| .name = "cbom_blocksize", |
| .get = prop_cbom_blksize_get, |
| .set = prop_cbom_blksize_set, |
| }; |
| |
| static void prop_cbop_blksize_set(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| uint16_t value; |
| |
| if (!visit_type_uint16(v, name, &value, errp)) { |
| return; |
| } |
| |
| if (value != cpu->cfg.cbop_blocksize && riscv_cpu_is_vendor(obj)) { |
| cpu_set_prop_err(cpu, name, errp); |
| error_append_hint(errp, "Current '%s' val: %u\n", |
| name, cpu->cfg.cbop_blocksize); |
| return; |
| } |
| |
| cpu_option_add_user_setting(name, value); |
| cpu->cfg.cbop_blocksize = value; |
| } |
| |
| static void prop_cbop_blksize_get(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| uint16_t value = RISCV_CPU(obj)->cfg.cbop_blocksize; |
| |
| visit_type_uint16(v, name, &value, errp); |
| } |
| |
| static const PropertyInfo prop_cbop_blksize = { |
| .name = "cbop_blocksize", |
| .get = prop_cbop_blksize_get, |
| .set = prop_cbop_blksize_set, |
| }; |
| |
| static void prop_cboz_blksize_set(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| uint16_t value; |
| |
| if (!visit_type_uint16(v, name, &value, errp)) { |
| return; |
| } |
| |
| if (value != cpu->cfg.cboz_blocksize && riscv_cpu_is_vendor(obj)) { |
| cpu_set_prop_err(cpu, name, errp); |
| error_append_hint(errp, "Current '%s' val: %u\n", |
| name, cpu->cfg.cboz_blocksize); |
| return; |
| } |
| |
| cpu_option_add_user_setting(name, value); |
| cpu->cfg.cboz_blocksize = value; |
| } |
| |
| static void prop_cboz_blksize_get(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| uint16_t value = RISCV_CPU(obj)->cfg.cboz_blocksize; |
| |
| visit_type_uint16(v, name, &value, errp); |
| } |
| |
| static const PropertyInfo prop_cboz_blksize = { |
| .name = "cboz_blocksize", |
| .get = prop_cboz_blksize_get, |
| .set = prop_cboz_blksize_set, |
| }; |
| |
| static void prop_mvendorid_set(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| bool dynamic_cpu = riscv_cpu_is_dynamic(obj); |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| uint32_t prev_val = cpu->cfg.mvendorid; |
| uint32_t value; |
| |
| if (!visit_type_uint32(v, name, &value, errp)) { |
| return; |
| } |
| |
| if (!dynamic_cpu && prev_val != value) { |
| error_setg(errp, "Unable to change %s mvendorid (0x%x)", |
| object_get_typename(obj), prev_val); |
| return; |
| } |
| |
| cpu->cfg.mvendorid = value; |
| } |
| |
| static void prop_mvendorid_get(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| uint32_t value = RISCV_CPU(obj)->cfg.mvendorid; |
| |
| visit_type_uint32(v, name, &value, errp); |
| } |
| |
| static const PropertyInfo prop_mvendorid = { |
| .name = "mvendorid", |
| .get = prop_mvendorid_get, |
| .set = prop_mvendorid_set, |
| }; |
| |
| static void prop_mimpid_set(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| bool dynamic_cpu = riscv_cpu_is_dynamic(obj); |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| uint64_t prev_val = cpu->cfg.mimpid; |
| uint64_t value; |
| |
| if (!visit_type_uint64(v, name, &value, errp)) { |
| return; |
| } |
| |
| if (!dynamic_cpu && prev_val != value) { |
| error_setg(errp, "Unable to change %s mimpid (0x%" PRIu64 ")", |
| object_get_typename(obj), prev_val); |
| return; |
| } |
| |
| cpu->cfg.mimpid = value; |
| } |
| |
| static void prop_mimpid_get(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| uint64_t value = RISCV_CPU(obj)->cfg.mimpid; |
| |
| visit_type_uint64(v, name, &value, errp); |
| } |
| |
| static const PropertyInfo prop_mimpid = { |
| .name = "mimpid", |
| .get = prop_mimpid_get, |
| .set = prop_mimpid_set, |
| }; |
| |
| static void prop_marchid_set(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| bool dynamic_cpu = riscv_cpu_is_dynamic(obj); |
| RISCVCPU *cpu = RISCV_CPU(obj); |
| uint64_t prev_val = cpu->cfg.marchid; |
| uint64_t value, invalid_val; |
| uint32_t mxlen = 0; |
| |
| if (!visit_type_uint64(v, name, &value, errp)) { |
| return; |
| } |
| |
| if (!dynamic_cpu && prev_val != value) { |
| error_setg(errp, "Unable to change %s marchid (0x%" PRIu64 ")", |
| object_get_typename(obj), prev_val); |
| return; |
| } |
| |
| switch (riscv_cpu_mxl(&cpu->env)) { |
| case MXL_RV32: |
| mxlen = 32; |
| break; |
| case MXL_RV64: |
| case MXL_RV128: |
| mxlen = 64; |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| |
| invalid_val = 1LL << (mxlen - 1); |
| |
| if (value == invalid_val) { |
| error_setg(errp, "Unable to set marchid with MSB (%u) bit set " |
| "and the remaining bits zero", mxlen); |
| return; |
| } |
| |
| cpu->cfg.marchid = value; |
| } |
| |
| static void prop_marchid_get(Object *obj, Visitor *v, const char *name, |
| void *opaque, Error **errp) |
| { |
| uint64_t value = RISCV_CPU(obj)->cfg.marchid; |
| |
| visit_type_uint64(v, name, &value, errp); |
| } |
| |
| static const PropertyInfo prop_marchid = { |
| .name = "marchid", |
| .get = prop_marchid_get, |
| .set = prop_marchid_set, |
| }; |
| |
| /* |
| * RVA22U64 defines some 'named features' that are cache |
| * related: Za64rs, Zic64b, Ziccif, Ziccrse, Ziccamoa |
| * and Zicclsm. They are always implemented in TCG and |
| * doesn't need to be manually enabled by the profile. |
| */ |
| static RISCVCPUProfile RVA22U64 = { |
| .parent = NULL, |
| .name = "rva22u64", |
| .misa_ext = RVI | RVM | RVA | RVF | RVD | RVC | RVU, |
| .priv_spec = RISCV_PROFILE_ATTR_UNUSED, |
| .satp_mode = RISCV_PROFILE_ATTR_UNUSED, |
| .ext_offsets = { |
| CPU_CFG_OFFSET(ext_zicsr), CPU_CFG_OFFSET(ext_zihintpause), |
| CPU_CFG_OFFSET(ext_zba), CPU_CFG_OFFSET(ext_zbb), |
| CPU_CFG_OFFSET(ext_zbs), CPU_CFG_OFFSET(ext_zfhmin), |
| CPU_CFG_OFFSET(ext_zkt), CPU_CFG_OFFSET(ext_zicntr), |
| CPU_CFG_OFFSET(ext_zihpm), CPU_CFG_OFFSET(ext_zicbom), |
| CPU_CFG_OFFSET(ext_zicbop), CPU_CFG_OFFSET(ext_zicboz), |
| |
| /* mandatory named features for this profile */ |
| CPU_CFG_OFFSET(ext_zic64b), |
| |
| RISCV_PROFILE_EXT_LIST_END |
| } |
| }; |
| |
| /* |
| * As with RVA22U64, RVA22S64 also defines 'named features'. |
| * |
| * Cache related features that we consider enabled since we don't |
| * implement cache: Ssccptr |
| * |
| * Other named features that we already implement: Sstvecd, Sstvala, |
| * Sscounterenw |
| * |
| * The remaining features/extensions comes from RVA22U64. |
| */ |
| static RISCVCPUProfile RVA22S64 = { |
| .parent = &RVA22U64, |
| .name = "rva22s64", |
| .misa_ext = RVS, |
| .priv_spec = PRIV_VERSION_1_12_0, |
| .satp_mode = VM_1_10_SV39, |
| .ext_offsets = { |
| /* rva22s64 exts */ |
| CPU_CFG_OFFSET(ext_zifencei), CPU_CFG_OFFSET(ext_svpbmt), |
| CPU_CFG_OFFSET(ext_svinval), CPU_CFG_OFFSET(ext_svade), |
| |
| RISCV_PROFILE_EXT_LIST_END |
| } |
| }; |
| |
| RISCVCPUProfile *riscv_profiles[] = { |
| &RVA22U64, |
| &RVA22S64, |
| NULL, |
| }; |
| |
| static RISCVCPUImpliedExtsRule RVA_IMPLIED = { |
| .is_misa = true, |
| .ext = RVA, |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zalrsc), CPU_CFG_OFFSET(ext_zaamo), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule RVD_IMPLIED = { |
| .is_misa = true, |
| .ext = RVD, |
| .implied_misa_exts = RVF, |
| .implied_multi_exts = { RISCV_IMPLIED_EXTS_RULE_END }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule RVF_IMPLIED = { |
| .is_misa = true, |
| .ext = RVF, |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zicsr), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule RVM_IMPLIED = { |
| .is_misa = true, |
| .ext = RVM, |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zmmul), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule RVV_IMPLIED = { |
| .is_misa = true, |
| .ext = RVV, |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zve64d), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZCB_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zcb), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zca), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZCD_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zcd), |
| .implied_misa_exts = RVD, |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zca), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZCE_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zce), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zcb), CPU_CFG_OFFSET(ext_zcmp), |
| CPU_CFG_OFFSET(ext_zcmt), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZCF_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zcf), |
| .implied_misa_exts = RVF, |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zca), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZCMP_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zcmp), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zca), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZCMT_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zcmt), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zca), CPU_CFG_OFFSET(ext_zicsr), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZDINX_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zdinx), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zfinx), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZFA_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zfa), |
| .implied_misa_exts = RVF, |
| .implied_multi_exts = { RISCV_IMPLIED_EXTS_RULE_END }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZFBFMIN_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zfbfmin), |
| .implied_misa_exts = RVF, |
| .implied_multi_exts = { RISCV_IMPLIED_EXTS_RULE_END }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZFH_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zfh), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zfhmin), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZFHMIN_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zfhmin), |
| .implied_misa_exts = RVF, |
| .implied_multi_exts = { RISCV_IMPLIED_EXTS_RULE_END }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZFINX_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zfinx), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zicsr), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZHINX_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zhinx), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zhinxmin), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZHINXMIN_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zhinxmin), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zfinx), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZICNTR_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zicntr), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zicsr), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZIHPM_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zihpm), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zicsr), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZK_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zk), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zkn), CPU_CFG_OFFSET(ext_zkr), |
| CPU_CFG_OFFSET(ext_zkt), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZKN_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zkn), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zbkb), CPU_CFG_OFFSET(ext_zbkc), |
| CPU_CFG_OFFSET(ext_zbkx), CPU_CFG_OFFSET(ext_zkne), |
| CPU_CFG_OFFSET(ext_zknd), CPU_CFG_OFFSET(ext_zknh), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZKS_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zks), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zbkb), CPU_CFG_OFFSET(ext_zbkc), |
| CPU_CFG_OFFSET(ext_zbkx), CPU_CFG_OFFSET(ext_zksed), |
| CPU_CFG_OFFSET(ext_zksh), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVBB_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zvbb), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zvkb), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVE32F_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zve32f), |
| .implied_misa_exts = RVF, |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zve32x), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVE32X_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zve32x), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zicsr), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVE64D_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zve64d), |
| .implied_misa_exts = RVD, |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zve64f), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVE64F_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zve64f), |
| .implied_misa_exts = RVF, |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zve32f), CPU_CFG_OFFSET(ext_zve64x), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVE64X_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zve64x), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zve32x), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVFBFMIN_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zvfbfmin), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zve32f), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVFBFWMA_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zvfbfwma), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zvfbfmin), CPU_CFG_OFFSET(ext_zfbfmin), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVFH_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zvfh), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zvfhmin), CPU_CFG_OFFSET(ext_zfhmin), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVFHMIN_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zvfhmin), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zve32f), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVKN_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zvkn), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zvkned), CPU_CFG_OFFSET(ext_zvknhb), |
| CPU_CFG_OFFSET(ext_zvkb), CPU_CFG_OFFSET(ext_zvkt), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVKNC_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zvknc), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zvkn), CPU_CFG_OFFSET(ext_zvbc), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVKNG_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zvkng), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zvkn), CPU_CFG_OFFSET(ext_zvkg), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVKNHB_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zvknhb), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zve64x), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVKS_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zvks), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zvksed), CPU_CFG_OFFSET(ext_zvksh), |
| CPU_CFG_OFFSET(ext_zvkb), CPU_CFG_OFFSET(ext_zvkt), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVKSC_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zvksc), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zvks), CPU_CFG_OFFSET(ext_zvbc), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| static RISCVCPUImpliedExtsRule ZVKSG_IMPLIED = { |
| .ext = CPU_CFG_OFFSET(ext_zvksg), |
| .implied_multi_exts = { |
| CPU_CFG_OFFSET(ext_zvks), CPU_CFG_OFFSET(ext_zvkg), |
| |
| RISCV_IMPLIED_EXTS_RULE_END |
| }, |
| }; |
| |
| RISCVCPUImpliedExtsRule *riscv_misa_ext_implied_rules[] = { |
| &RVA_IMPLIED, &RVD_IMPLIED, &RVF_IMPLIED, |
| &RVM_IMPLIED, &RVV_IMPLIED, NULL |
| }; |
| |
| RISCVCPUImpliedExtsRule *riscv_multi_ext_implied_rules[] = { |
| &ZCB_IMPLIED, &ZCD_IMPLIED, &ZCE_IMPLIED, |
| &ZCF_IMPLIED, &ZCMP_IMPLIED, &ZCMT_IMPLIED, |
| &ZDINX_IMPLIED, &ZFA_IMPLIED, &ZFBFMIN_IMPLIED, |
| &ZFH_IMPLIED, &ZFHMIN_IMPLIED, &ZFINX_IMPLIED, |
| &ZHINX_IMPLIED, &ZHINXMIN_IMPLIED, &ZICNTR_IMPLIED, |
| &ZIHPM_IMPLIED, &ZK_IMPLIED, &ZKN_IMPLIED, |
| &ZKS_IMPLIED, &ZVBB_IMPLIED, &ZVE32F_IMPLIED, |
| &ZVE32X_IMPLIED, &ZVE64D_IMPLIED, &ZVE64F_IMPLIED, |
| &ZVE64X_IMPLIED, &ZVFBFMIN_IMPLIED, &ZVFBFWMA_IMPLIED, |
| &ZVFH_IMPLIED, &ZVFHMIN_IMPLIED, &ZVKN_IMPLIED, |
| &ZVKNC_IMPLIED, &ZVKNG_IMPLIED, &ZVKNHB_IMPLIED, |
| &ZVKS_IMPLIED, &ZVKSC_IMPLIED, &ZVKSG_IMPLIED, |
| NULL |
| }; |
| |
| static Property riscv_cpu_properties[] = { |
| DEFINE_PROP_BOOL("debug", RISCVCPU, cfg.debug, true), |
| |
| {.name = "pmu-mask", .info = &prop_pmu_mask}, |
| {.name = "pmu-num", .info = &prop_pmu_num}, /* Deprecated */ |
| |
| {.name = "mmu", .info = &prop_mmu}, |
| {.name = "pmp", .info = &prop_pmp}, |
| |
| {.name = "priv_spec", .info = &prop_priv_spec}, |
| {.name = "vext_spec", .info = &prop_vext_spec}, |
| |
| {.name = "vlen", .info = &prop_vlen}, |
| {.name = "elen", .info = &prop_elen}, |
| |
| {.name = "cbom_blocksize", .info = &prop_cbom_blksize}, |
| {.name = "cbop_blocksize", .info = &prop_cbop_blksize}, |
| {.name = "cboz_blocksize", .info = &prop_cboz_blksize}, |
| |
| {.name = "mvendorid", .info = &prop_mvendorid}, |
| {.name = "mimpid", .info = &prop_mimpid}, |
| {.name = "marchid", .info = &prop_marchid}, |
| |
| #ifndef CONFIG_USER_ONLY |
| DEFINE_PROP_UINT64("resetvec", RISCVCPU, env.resetvec, DEFAULT_RSTVEC), |
| #endif |
| |
| DEFINE_PROP_BOOL("short-isa-string", RISCVCPU, cfg.short_isa_string, false), |
| |
| DEFINE_PROP_BOOL("rvv_ta_all_1s", RISCVCPU, cfg.rvv_ta_all_1s, false), |
| DEFINE_PROP_BOOL("rvv_ma_all_1s", RISCVCPU, cfg.rvv_ma_all_1s, false), |
| DEFINE_PROP_BOOL("rvv_vl_half_avl", RISCVCPU, cfg.rvv_vl_half_avl, false), |
| |
| /* |
| * write_misa() is marked as experimental for now so mark |
| * it with -x and default to 'false'. |
| */ |
| DEFINE_PROP_BOOL("x-misa-w", RISCVCPU, cfg.misa_w, false), |
| DEFINE_PROP_END_OF_LIST(), |
| }; |
| |
| #if defined(TARGET_RISCV64) |
| static void rva22u64_profile_cpu_init(Object *obj) |
| { |
| rv64i_bare_cpu_init(obj); |
| |
| RVA22U64.enabled = true; |
| } |
| |
| static void rva22s64_profile_cpu_init(Object *obj) |
| { |
| rv64i_bare_cpu_init(obj); |
| |
| RVA22S64.enabled = true; |
| } |
| #endif |
| |
| static const gchar *riscv_gdb_arch_name(CPUState *cs) |
| { |
| RISCVCPU *cpu = RISCV_CPU(cs); |
| CPURISCVState *env = &cpu->env; |
| |
| switch (riscv_cpu_mxl(env)) { |
| case MXL_RV32: |
| return "riscv:rv32"; |
| case MXL_RV64: |
| case MXL_RV128: |
| return "riscv:rv64"; |
| default: |
| g_assert_not_reached(); |
| } |
| } |
| |
| #ifndef CONFIG_USER_ONLY |
| static int64_t riscv_get_arch_id(CPUState *cs) |
| { |
| RISCVCPU *cpu = RISCV_CPU(cs); |
| |
| return cpu->env.mhartid; |
| } |
| |
| #include "hw/core/sysemu-cpu-ops.h" |
| |
| static const struct SysemuCPUOps riscv_sysemu_ops = { |
| .get_phys_page_debug = riscv_cpu_get_phys_page_debug, |
| .write_elf64_note = riscv_cpu_write_elf64_note, |
| .write_elf32_note = riscv_cpu_write_elf32_note, |
| .legacy_vmsd = &vmstate_riscv_cpu, |
| }; |
| #endif |
| |
| static void riscv_cpu_common_class_init(ObjectClass *c, void *data) |
| { |
| RISCVCPUClass *mcc = RISCV_CPU_CLASS(c); |
| CPUClass *cc = CPU_CLASS(c); |
| DeviceClass *dc = DEVICE_CLASS(c); |
| ResettableClass *rc = RESETTABLE_CLASS(c); |
| |
| device_class_set_parent_realize(dc, riscv_cpu_realize, |
| &mcc->parent_realize); |
| |
| resettable_class_set_parent_phases(rc, NULL, riscv_cpu_reset_hold, NULL, |
| &mcc->parent_phases); |
| |
| cc->class_by_name = riscv_cpu_class_by_name; |
| cc->has_work = riscv_cpu_has_work; |
| cc->mmu_index = riscv_cpu_mmu_index; |
| cc->dump_state = riscv_cpu_dump_state; |
| cc->set_pc = riscv_cpu_set_pc; |
| cc->get_pc = riscv_cpu_get_pc; |
| cc->gdb_read_register = riscv_cpu_gdb_read_register; |
| cc->gdb_write_register = riscv_cpu_gdb_write_register; |
| cc->gdb_stop_before_watchpoint = true; |
| cc->disas_set_info = riscv_cpu_disas_set_info; |
| #ifndef CONFIG_USER_ONLY |
| cc->sysemu_ops = &riscv_sysemu_ops; |
| cc->get_arch_id = riscv_get_arch_id; |
| #endif |
| cc->gdb_arch_name = riscv_gdb_arch_name; |
| |
| device_class_set_props(dc, riscv_cpu_properties); |
| } |
| |
| static void riscv_cpu_class_init(ObjectClass *c, void *data) |
| { |
| RISCVCPUClass *mcc = RISCV_CPU_CLASS(c); |
| |
| mcc->misa_mxl_max = (uint32_t)(uintptr_t)data; |
| riscv_cpu_validate_misa_mxl(mcc); |
| } |
| |
| static void riscv_isa_string_ext(RISCVCPU *cpu, char **isa_str, |
| int max_str_len) |
| { |
| const RISCVIsaExtData *edata; |
| char *old = *isa_str; |
| char *new = *isa_str; |
| |
| for (edata = isa_edata_arr; edata && edata->name; edata++) { |
| if (isa_ext_is_enabled(cpu, edata->ext_enable_offset)) { |
| new = g_strconcat(old, "_", edata->name, NULL); |
| g_free(old); |
| old = new; |
| } |
| } |
| |
| *isa_str = new; |
| } |
| |
| char *riscv_isa_string(RISCVCPU *cpu) |
| { |
| RISCVCPUClass *mcc = RISCV_CPU_GET_CLASS(cpu); |
| int i; |
| const size_t maxlen = sizeof("rv128") + sizeof(riscv_single_letter_exts); |
| char *isa_str = g_new(char, maxlen); |
| int xlen = riscv_cpu_max_xlen(mcc); |
| char *p = isa_str + snprintf(isa_str, maxlen, "rv%d", xlen); |
| |
| for (i = 0; i < sizeof(riscv_single_letter_exts) - 1; i++) { |
| if (cpu->env.misa_ext & RV(riscv_single_letter_exts[i])) { |
| *p++ = qemu_tolower(riscv_single_letter_exts[i]); |
| } |
| } |
| *p = '\0'; |
| if (!cpu->cfg.short_isa_string) { |
| riscv_isa_string_ext(cpu, &isa_str, maxlen); |
| } |
| return isa_str; |
| } |
| |
| #ifndef CONFIG_USER_ONLY |
| static char **riscv_isa_extensions_list(RISCVCPU *cpu, int *count) |
| { |
| int maxlen = ARRAY_SIZE(riscv_single_letter_exts) + ARRAY_SIZE(isa_edata_arr); |
| char **extensions = g_new(char *, maxlen); |
| |
| for (int i = 0; i < sizeof(riscv_single_letter_exts) - 1; i++) { |
| if (cpu->env.misa_ext & RV(riscv_single_letter_exts[i])) { |
| extensions[*count] = g_new(char, 2); |
| snprintf(extensions[*count], 2, "%c", |
| qemu_tolower(riscv_single_letter_exts[i])); |
| (*count)++; |
| } |
| } |
| |
| for (const RISCVIsaExtData *edata = isa_edata_arr; edata->name; edata++) { |
| if (isa_ext_is_enabled(cpu, edata->ext_enable_offset)) { |
| extensions[*count] = g_strdup(edata->name); |
| (*count)++; |
| } |
| } |
| |
| return extensions; |
| } |
| |
| void riscv_isa_write_fdt(RISCVCPU *cpu, void *fdt, char *nodename) |
| { |
| RISCVCPUClass *mcc = RISCV_CPU_GET_CLASS(cpu); |
| const size_t maxlen = sizeof("rv128i"); |
| g_autofree char *isa_base = g_new(char, maxlen); |
| g_autofree char *riscv_isa; |
| char **isa_extensions; |
| int count = 0; |
| int xlen = riscv_cpu_max_xlen(mcc); |
| |
| riscv_isa = riscv_isa_string(cpu); |
| qemu_fdt_setprop_string(fdt, nodename, "riscv,isa", riscv_isa); |
| |
| snprintf(isa_base, maxlen, "rv%di", xlen); |
| qemu_fdt_setprop_string(fdt, nodename, "riscv,isa-base", isa_base); |
| |
| isa_extensions = riscv_isa_extensions_list(cpu, &count); |
| qemu_fdt_setprop_string_array(fdt, nodename, "riscv,isa-extensions", |
| isa_extensions, count); |
| |
| for (int i = 0; i < count; i++) { |
| g_free(isa_extensions[i]); |
| } |
| |
| g_free(isa_extensions); |
| } |
| #endif |
| |
| #define DEFINE_CPU(type_name, misa_mxl_max, initfn) \ |
| { \ |
| .name = (type_name), \ |
| .parent = TYPE_RISCV_CPU, \ |
| .instance_init = (initfn), \ |
| .class_init = riscv_cpu_class_init, \ |
| .class_data = (void *)(misa_mxl_max) \ |
| } |
| |
| #define DEFINE_DYNAMIC_CPU(type_name, misa_mxl_max, initfn) \ |
| { \ |
| .name = (type_name), \ |
| .parent = TYPE_RISCV_DYNAMIC_CPU, \ |
| .instance_init = (initfn), \ |
| .class_init = riscv_cpu_class_init, \ |
| .class_data = (void *)(misa_mxl_max) \ |
| } |
| |
| #define DEFINE_VENDOR_CPU(type_name, misa_mxl_max, initfn) \ |
| { \ |
| .name = (type_name), \ |
| .parent = TYPE_RISCV_VENDOR_CPU, \ |
| .instance_init = (initfn), \ |
| .class_init = riscv_cpu_class_init, \ |
| .class_data = (void *)(misa_mxl_max) \ |
| } |
| |
| #define DEFINE_BARE_CPU(type_name, misa_mxl_max, initfn) \ |
| { \ |
| .name = (type_name), \ |
| .parent = TYPE_RISCV_BARE_CPU, \ |
| .instance_init = (initfn), \ |
| .class_init = riscv_cpu_class_init, \ |
| .class_data = (void *)(misa_mxl_max) \ |
| } |
| |
| #define DEFINE_PROFILE_CPU(type_name, misa_mxl_max, initfn) \ |
| { \ |
| .name = (type_name), \ |
| .parent = TYPE_RISCV_BARE_CPU, \ |
| .instance_init = (initfn), \ |
| .class_init = riscv_cpu_class_init, \ |
| .class_data = (void *)(misa_mxl_max) \ |
| } |
| |
| static const TypeInfo riscv_cpu_type_infos[] = { |
| { |
| .name = TYPE_RISCV_CPU, |
| .parent = TYPE_CPU, |
| .instance_size = sizeof(RISCVCPU), |
| .instance_align = __alignof(RISCVCPU), |
| .instance_init = riscv_cpu_init, |
| .instance_post_init = riscv_cpu_post_init, |
| .abstract = true, |
| .class_size = sizeof(RISCVCPUClass), |
| .class_init = riscv_cpu_common_class_init, |
| }, |
| { |
| .name = TYPE_RISCV_DYNAMIC_CPU, |
| .parent = TYPE_RISCV_CPU, |
| .abstract = true, |
| }, |
| { |
| .name = TYPE_RISCV_VENDOR_CPU, |
| .parent = TYPE_RISCV_CPU, |
| .abstract = true, |
| }, |
| { |
| .name = TYPE_RISCV_BARE_CPU, |
| .parent = TYPE_RISCV_CPU, |
| .instance_init = riscv_bare_cpu_init, |
| .abstract = true, |
| }, |
| #if defined(TARGET_RISCV32) |
| DEFINE_DYNAMIC_CPU(TYPE_RISCV_CPU_MAX, MXL_RV32, riscv_max_cpu_init), |
| DEFINE_DYNAMIC_CPU(TYPE_RISCV_CPU_BASE32, MXL_RV32, rv32_base_cpu_init), |
| DEFINE_VENDOR_CPU(TYPE_RISCV_CPU_IBEX, MXL_RV32, rv32_ibex_cpu_init), |
| DEFINE_VENDOR_CPU(TYPE_RISCV_CPU_SIFIVE_E31, MXL_RV32, rv32_sifive_e_cpu_init), |
| DEFINE_VENDOR_CPU(TYPE_RISCV_CPU_SIFIVE_E34, MXL_RV32, rv32_imafcu_nommu_cpu_init), |
| DEFINE_VENDOR_CPU(TYPE_RISCV_CPU_SIFIVE_U34, MXL_RV32, rv32_sifive_u_cpu_init), |
| DEFINE_BARE_CPU(TYPE_RISCV_CPU_RV32I, MXL_RV32, rv32i_bare_cpu_init), |
| DEFINE_BARE_CPU(TYPE_RISCV_CPU_RV32E, MXL_RV32, rv32e_bare_cpu_init), |
| #elif defined(TARGET_RISCV64) |
| DEFINE_DYNAMIC_CPU(TYPE_RISCV_CPU_MAX, MXL_RV64, riscv_max_cpu_init), |
| DEFINE_DYNAMIC_CPU(TYPE_RISCV_CPU_BASE64, MXL_RV64, rv64_base_cpu_init), |
| DEFINE_VENDOR_CPU(TYPE_RISCV_CPU_SIFIVE_E51, MXL_RV64, rv64_sifive_e_cpu_init), |
| DEFINE_VENDOR_CPU(TYPE_RISCV_CPU_SIFIVE_U54, MXL_RV64, rv64_sifive_u_cpu_init), |
| DEFINE_VENDOR_CPU(TYPE_RISCV_CPU_SHAKTI_C, MXL_RV64, rv64_sifive_u_cpu_init), |
| DEFINE_VENDOR_CPU(TYPE_RISCV_CPU_THEAD_C906, MXL_RV64, rv64_thead_c906_cpu_init), |
| DEFINE_VENDOR_CPU(TYPE_RISCV_CPU_VEYRON_V1, MXL_RV64, rv64_veyron_v1_cpu_init), |
| #ifdef CONFIG_TCG |
| DEFINE_DYNAMIC_CPU(TYPE_RISCV_CPU_BASE128, MXL_RV128, rv128_base_cpu_init), |
| #endif /* CONFIG_TCG */ |
| DEFINE_BARE_CPU(TYPE_RISCV_CPU_RV64I, MXL_RV64, rv64i_bare_cpu_init), |
| DEFINE_BARE_CPU(TYPE_RISCV_CPU_RV64E, MXL_RV64, rv64e_bare_cpu_init), |
| DEFINE_PROFILE_CPU(TYPE_RISCV_CPU_RVA22U64, MXL_RV64, rva22u64_profile_cpu_init), |
| DEFINE_PROFILE_CPU(TYPE_RISCV_CPU_RVA22S64, MXL_RV64, rva22s64_profile_cpu_init), |
| #endif /* TARGET_RISCV64 */ |
| }; |
| |
| DEFINE_TYPES(riscv_cpu_type_infos) |