| /* |
| * RISC-V translation routines for the T-Head vendor extensions (xthead*). |
| * |
| * Copyright (c) 2022 VRULL GmbH. |
| * |
| * This program is free software; you can redistribute it and/or modify it |
| * under the terms and conditions of the GNU General Public License, |
| * version 2 or later, as published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope it will be useful, but WITHOUT |
| * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for |
| * more details. |
| * |
| * You should have received a copy of the GNU General Public License along with |
| * this program. If not, see <http://www.gnu.org/licenses/>. |
| */ |
| |
| #define REQUIRE_XTHEADBA(ctx) do { \ |
| if (!ctx->cfg_ptr->ext_xtheadba) { \ |
| return false; \ |
| } \ |
| } while (0) |
| |
| #define REQUIRE_XTHEADBB(ctx) do { \ |
| if (!ctx->cfg_ptr->ext_xtheadbb) { \ |
| return false; \ |
| } \ |
| } while (0) |
| |
| #define REQUIRE_XTHEADBS(ctx) do { \ |
| if (!ctx->cfg_ptr->ext_xtheadbs) { \ |
| return false; \ |
| } \ |
| } while (0) |
| |
| #define REQUIRE_XTHEADCMO(ctx) do { \ |
| if (!ctx->cfg_ptr->ext_xtheadcmo) { \ |
| return false; \ |
| } \ |
| } while (0) |
| |
| #define REQUIRE_XTHEADCONDMOV(ctx) do { \ |
| if (!ctx->cfg_ptr->ext_xtheadcondmov) { \ |
| return false; \ |
| } \ |
| } while (0) |
| |
| #define REQUIRE_XTHEADFMEMIDX(ctx) do { \ |
| if (!ctx->cfg_ptr->ext_xtheadfmemidx) { \ |
| return false; \ |
| } \ |
| } while (0) |
| |
| #define REQUIRE_XTHEADFMV(ctx) do { \ |
| if (!ctx->cfg_ptr->ext_xtheadfmv) { \ |
| return false; \ |
| } \ |
| } while (0) |
| |
| #define REQUIRE_XTHEADMAC(ctx) do { \ |
| if (!ctx->cfg_ptr->ext_xtheadmac) { \ |
| return false; \ |
| } \ |
| } while (0) |
| |
| #define REQUIRE_XTHEADMEMIDX(ctx) do { \ |
| if (!ctx->cfg_ptr->ext_xtheadmemidx) { \ |
| return false; \ |
| } \ |
| } while (0) |
| |
| #define REQUIRE_XTHEADMEMPAIR(ctx) do { \ |
| if (!ctx->cfg_ptr->ext_xtheadmempair) { \ |
| return false; \ |
| } \ |
| } while (0) |
| |
| #define REQUIRE_XTHEADSYNC(ctx) do { \ |
| if (!ctx->cfg_ptr->ext_xtheadsync) { \ |
| return false; \ |
| } \ |
| } while (0) |
| |
| /* |
| * Calculate and return the address for indexed mem operations: |
| * If !zext_offs, then the address is rs1 + (rs2 << imm2). |
| * If zext_offs, then the address is rs1 + (zext(rs2[31:0]) << imm2). |
| */ |
| static TCGv get_th_address_indexed(DisasContext *ctx, int rs1, int rs2, |
| int imm2, bool zext_offs) |
| { |
| TCGv src2 = get_gpr(ctx, rs2, EXT_NONE); |
| TCGv offs = tcg_temp_new(); |
| |
| if (zext_offs) { |
| tcg_gen_extract_tl(offs, src2, 0, 32); |
| tcg_gen_shli_tl(offs, offs, imm2); |
| } else { |
| tcg_gen_shli_tl(offs, src2, imm2); |
| } |
| |
| return get_address_indexed(ctx, rs1, offs); |
| } |
| |
| /* XTheadBa */ |
| |
| /* |
| * th.addsl is similar to sh[123]add (from Zba), but not an |
| * alternative encoding: while sh[123] applies the shift to rs1, |
| * th.addsl shifts rs2. |
| */ |
| |
| #define GEN_TH_ADDSL(SHAMT) \ |
| static void gen_th_addsl##SHAMT(TCGv ret, TCGv arg1, TCGv arg2) \ |
| { \ |
| TCGv t = tcg_temp_new(); \ |
| tcg_gen_shli_tl(t, arg2, SHAMT); \ |
| tcg_gen_add_tl(ret, t, arg1); \ |
| } |
| |
| GEN_TH_ADDSL(1) |
| GEN_TH_ADDSL(2) |
| GEN_TH_ADDSL(3) |
| |
| #define GEN_TRANS_TH_ADDSL(SHAMT) \ |
| static bool trans_th_addsl##SHAMT(DisasContext *ctx, \ |
| arg_th_addsl##SHAMT * a) \ |
| { \ |
| REQUIRE_XTHEADBA(ctx); \ |
| return gen_arith(ctx, a, EXT_NONE, gen_th_addsl##SHAMT, NULL); \ |
| } |
| |
| GEN_TRANS_TH_ADDSL(1) |
| GEN_TRANS_TH_ADDSL(2) |
| GEN_TRANS_TH_ADDSL(3) |
| |
| /* XTheadBb */ |
| |
| /* th.srri is an alternate encoding for rori (from Zbb) */ |
| static bool trans_th_srri(DisasContext *ctx, arg_th_srri * a) |
| { |
| REQUIRE_XTHEADBB(ctx); |
| return gen_shift_imm_fn_per_ol(ctx, a, EXT_NONE, |
| tcg_gen_rotri_tl, gen_roriw, NULL); |
| } |
| |
| /* th.srriw is an alternate encoding for roriw (from Zbb) */ |
| static bool trans_th_srriw(DisasContext *ctx, arg_th_srriw *a) |
| { |
| REQUIRE_XTHEADBB(ctx); |
| REQUIRE_64BIT(ctx); |
| ctx->ol = MXL_RV32; |
| return gen_shift_imm_fn(ctx, a, EXT_NONE, gen_roriw, NULL); |
| } |
| |
| /* th.ext and th.extu perform signed/unsigned bitfield extraction */ |
| static bool gen_th_bfextract(DisasContext *ctx, arg_th_bfext *a, |
| void (*f)(TCGv, TCGv, unsigned int, unsigned int)) |
| { |
| TCGv dest = dest_gpr(ctx, a->rd); |
| TCGv source = get_gpr(ctx, a->rs1, EXT_ZERO); |
| |
| if (a->lsb <= a->msb) { |
| f(dest, source, a->lsb, a->msb - a->lsb + 1); |
| gen_set_gpr(ctx, a->rd, dest); |
| } |
| return true; |
| } |
| |
| static bool trans_th_ext(DisasContext *ctx, arg_th_ext *a) |
| { |
| REQUIRE_XTHEADBB(ctx); |
| return gen_th_bfextract(ctx, a, tcg_gen_sextract_tl); |
| } |
| |
| static bool trans_th_extu(DisasContext *ctx, arg_th_extu *a) |
| { |
| REQUIRE_XTHEADBB(ctx); |
| return gen_th_bfextract(ctx, a, tcg_gen_extract_tl); |
| } |
| |
| /* th.ff0: find first zero (clz on an inverted input) */ |
| static bool gen_th_ff0(DisasContext *ctx, arg_th_ff0 *a, DisasExtend ext) |
| { |
| TCGv dest = dest_gpr(ctx, a->rd); |
| TCGv src1 = get_gpr(ctx, a->rs1, ext); |
| |
| int olen = get_olen(ctx); |
| TCGv t = tcg_temp_new(); |
| |
| tcg_gen_not_tl(t, src1); |
| if (olen != TARGET_LONG_BITS) { |
| if (olen == 32) { |
| gen_clzw(dest, t); |
| } else { |
| g_assert_not_reached(); |
| } |
| } else { |
| gen_clz(dest, t); |
| } |
| |
| gen_set_gpr(ctx, a->rd, dest); |
| |
| return true; |
| } |
| |
| static bool trans_th_ff0(DisasContext *ctx, arg_th_ff0 *a) |
| { |
| REQUIRE_XTHEADBB(ctx); |
| return gen_th_ff0(ctx, a, EXT_NONE); |
| } |
| |
| /* th.ff1 is an alternate encoding for clz (from Zbb) */ |
| static bool trans_th_ff1(DisasContext *ctx, arg_th_ff1 *a) |
| { |
| REQUIRE_XTHEADBB(ctx); |
| return gen_unary_per_ol(ctx, a, EXT_NONE, gen_clz, gen_clzw); |
| } |
| |
| static void gen_th_revw(TCGv ret, TCGv arg1) |
| { |
| tcg_gen_bswap32_tl(ret, arg1, TCG_BSWAP_OS); |
| } |
| |
| /* th.rev is an alternate encoding for the RV64 rev8 (from Zbb) */ |
| static bool trans_th_rev(DisasContext *ctx, arg_th_rev *a) |
| { |
| REQUIRE_XTHEADBB(ctx); |
| |
| return gen_unary_per_ol(ctx, a, EXT_NONE, tcg_gen_bswap_tl, gen_th_revw); |
| } |
| |
| /* th.revw is a sign-extended byte-swap of the lower word */ |
| static bool trans_th_revw(DisasContext *ctx, arg_th_revw *a) |
| { |
| REQUIRE_XTHEADBB(ctx); |
| REQUIRE_64BIT(ctx); |
| return gen_unary(ctx, a, EXT_NONE, gen_th_revw); |
| } |
| |
| /* th.tstnbz is equivalent to an orc.b (from Zbb) with inverted result */ |
| static void gen_th_tstnbz(TCGv ret, TCGv source1) |
| { |
| gen_orc_b(ret, source1); |
| tcg_gen_not_tl(ret, ret); |
| } |
| |
| static bool trans_th_tstnbz(DisasContext *ctx, arg_th_tstnbz *a) |
| { |
| REQUIRE_XTHEADBB(ctx); |
| return gen_unary(ctx, a, EXT_ZERO, gen_th_tstnbz); |
| } |
| |
| /* XTheadBs */ |
| |
| /* th.tst is an alternate encoding for bexti (from Zbs) */ |
| static bool trans_th_tst(DisasContext *ctx, arg_th_tst *a) |
| { |
| REQUIRE_XTHEADBS(ctx); |
| return gen_shift_imm_tl(ctx, a, EXT_NONE, gen_bext); |
| } |
| |
| /* XTheadCmo */ |
| |
| static inline int priv_level(DisasContext *ctx) |
| { |
| #ifdef CONFIG_USER_ONLY |
| return PRV_U; |
| #else |
| /* Priv level is part of mem_idx. */ |
| return ctx->mem_idx & TB_FLAGS_PRIV_MMU_MASK; |
| #endif |
| } |
| |
| /* Test if priv level is M, S, or U (cannot fail). */ |
| #define REQUIRE_PRIV_MSU(ctx) |
| |
| /* Test if priv level is M or S. */ |
| #define REQUIRE_PRIV_MS(ctx) \ |
| do { \ |
| int priv = priv_level(ctx); \ |
| if (!(priv == PRV_M || \ |
| priv == PRV_S)) { \ |
| return false; \ |
| } \ |
| } while (0) |
| |
| #define NOP_PRIVCHECK(insn, extcheck, privcheck) \ |
| static bool trans_ ## insn(DisasContext *ctx, arg_ ## insn * a) \ |
| { \ |
| (void) a; \ |
| extcheck(ctx); \ |
| privcheck(ctx); \ |
| return true; \ |
| } |
| |
| NOP_PRIVCHECK(th_dcache_call, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| NOP_PRIVCHECK(th_dcache_ciall, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| NOP_PRIVCHECK(th_dcache_iall, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| NOP_PRIVCHECK(th_dcache_cpa, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| NOP_PRIVCHECK(th_dcache_cipa, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| NOP_PRIVCHECK(th_dcache_ipa, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| NOP_PRIVCHECK(th_dcache_cva, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MSU) |
| NOP_PRIVCHECK(th_dcache_civa, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MSU) |
| NOP_PRIVCHECK(th_dcache_iva, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MSU) |
| NOP_PRIVCHECK(th_dcache_csw, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| NOP_PRIVCHECK(th_dcache_cisw, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| NOP_PRIVCHECK(th_dcache_isw, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| NOP_PRIVCHECK(th_dcache_cpal1, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| NOP_PRIVCHECK(th_dcache_cval1, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| |
| NOP_PRIVCHECK(th_icache_iall, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| NOP_PRIVCHECK(th_icache_ialls, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| NOP_PRIVCHECK(th_icache_ipa, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| NOP_PRIVCHECK(th_icache_iva, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MSU) |
| |
| NOP_PRIVCHECK(th_l2cache_call, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| NOP_PRIVCHECK(th_l2cache_ciall, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| NOP_PRIVCHECK(th_l2cache_iall, REQUIRE_XTHEADCMO, REQUIRE_PRIV_MS) |
| |
| /* XTheadCondMov */ |
| |
| static bool gen_th_condmove(DisasContext *ctx, arg_r *a, TCGCond cond) |
| { |
| TCGv src1 = get_gpr(ctx, a->rs1, EXT_NONE); |
| TCGv src2 = get_gpr(ctx, a->rs2, EXT_NONE); |
| TCGv old = get_gpr(ctx, a->rd, EXT_NONE); |
| TCGv dest = dest_gpr(ctx, a->rd); |
| |
| tcg_gen_movcond_tl(cond, dest, src2, ctx->zero, src1, old); |
| |
| gen_set_gpr(ctx, a->rd, dest); |
| return true; |
| } |
| |
| /* th.mveqz: "if (rs2 == 0) rd = rs1;" */ |
| static bool trans_th_mveqz(DisasContext *ctx, arg_th_mveqz *a) |
| { |
| REQUIRE_XTHEADCONDMOV(ctx); |
| return gen_th_condmove(ctx, a, TCG_COND_EQ); |
| } |
| |
| /* th.mvnez: "if (rs2 != 0) rd = rs1;" */ |
| static bool trans_th_mvnez(DisasContext *ctx, arg_th_mveqz *a) |
| { |
| REQUIRE_XTHEADCONDMOV(ctx); |
| return gen_th_condmove(ctx, a, TCG_COND_NE); |
| } |
| |
| /* XTheadFMem */ |
| |
| /* |
| * Load 64-bit float from indexed address. |
| * If !zext_offs, then address is rs1 + (rs2 << imm2). |
| * If zext_offs, then address is rs1 + (zext(rs2[31:0]) << imm2). |
| */ |
| static bool gen_fload_idx(DisasContext *ctx, arg_th_memidx *a, MemOp memop, |
| bool zext_offs) |
| { |
| TCGv_i64 rd = cpu_fpr[a->rd]; |
| TCGv addr = get_th_address_indexed(ctx, a->rs1, a->rs2, a->imm2, zext_offs); |
| |
| tcg_gen_qemu_ld_i64(rd, addr, ctx->mem_idx, memop); |
| if ((memop & MO_SIZE) == MO_32) { |
| gen_nanbox_s(rd, rd); |
| } |
| |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| /* |
| * Store 64-bit float to indexed address. |
| * If !zext_offs, then address is rs1 + (rs2 << imm2). |
| * If zext_offs, then address is rs1 + (zext(rs2[31:0]) << imm2). |
| */ |
| static bool gen_fstore_idx(DisasContext *ctx, arg_th_memidx *a, MemOp memop, |
| bool zext_offs) |
| { |
| TCGv_i64 rd = cpu_fpr[a->rd]; |
| TCGv addr = get_th_address_indexed(ctx, a->rs1, a->rs2, a->imm2, zext_offs); |
| |
| tcg_gen_qemu_st_i64(rd, addr, ctx->mem_idx, memop); |
| |
| return true; |
| } |
| |
| static bool trans_th_flrd(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADFMEMIDX(ctx); |
| REQUIRE_FPU; |
| REQUIRE_EXT(ctx, RVD); |
| return gen_fload_idx(ctx, a, MO_TEUQ, false); |
| } |
| |
| static bool trans_th_flrw(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADFMEMIDX(ctx); |
| REQUIRE_FPU; |
| REQUIRE_EXT(ctx, RVF); |
| return gen_fload_idx(ctx, a, MO_TEUL, false); |
| } |
| |
| static bool trans_th_flurd(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADFMEMIDX(ctx); |
| REQUIRE_FPU; |
| REQUIRE_EXT(ctx, RVD); |
| return gen_fload_idx(ctx, a, MO_TEUQ, true); |
| } |
| |
| static bool trans_th_flurw(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADFMEMIDX(ctx); |
| REQUIRE_FPU; |
| REQUIRE_EXT(ctx, RVF); |
| return gen_fload_idx(ctx, a, MO_TEUL, true); |
| } |
| |
| static bool trans_th_fsrd(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADFMEMIDX(ctx); |
| REQUIRE_FPU; |
| REQUIRE_EXT(ctx, RVD); |
| return gen_fstore_idx(ctx, a, MO_TEUQ, false); |
| } |
| |
| static bool trans_th_fsrw(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADFMEMIDX(ctx); |
| REQUIRE_FPU; |
| REQUIRE_EXT(ctx, RVF); |
| return gen_fstore_idx(ctx, a, MO_TEUL, false); |
| } |
| |
| static bool trans_th_fsurd(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADFMEMIDX(ctx); |
| REQUIRE_FPU; |
| REQUIRE_EXT(ctx, RVD); |
| return gen_fstore_idx(ctx, a, MO_TEUQ, true); |
| } |
| |
| static bool trans_th_fsurw(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADFMEMIDX(ctx); |
| REQUIRE_FPU; |
| REQUIRE_EXT(ctx, RVF); |
| return gen_fstore_idx(ctx, a, MO_TEUL, true); |
| } |
| |
| /* XTheadFmv */ |
| |
| static bool trans_th_fmv_hw_x(DisasContext *ctx, arg_th_fmv_hw_x *a) |
| { |
| REQUIRE_XTHEADFMV(ctx); |
| REQUIRE_32BIT(ctx); |
| REQUIRE_FPU; |
| REQUIRE_EXT(ctx, RVD); |
| |
| TCGv src1 = get_gpr(ctx, a->rs1, EXT_ZERO); |
| TCGv_i64 t1 = tcg_temp_new_i64(); |
| |
| tcg_gen_extu_tl_i64(t1, src1); |
| tcg_gen_deposit_i64(cpu_fpr[a->rd], cpu_fpr[a->rd], t1, 32, 32); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_th_fmv_x_hw(DisasContext *ctx, arg_th_fmv_x_hw *a) |
| { |
| REQUIRE_XTHEADFMV(ctx); |
| REQUIRE_32BIT(ctx); |
| REQUIRE_FPU; |
| REQUIRE_EXT(ctx, RVD); |
| TCGv dst; |
| TCGv_i64 t1; |
| |
| dst = dest_gpr(ctx, a->rd); |
| t1 = tcg_temp_new_i64(); |
| |
| tcg_gen_extract_i64(t1, cpu_fpr[a->rs1], 32, 32); |
| tcg_gen_trunc_i64_tl(dst, t1); |
| gen_set_gpr(ctx, a->rd, dst); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| /* XTheadMac */ |
| |
| static bool gen_th_mac(DisasContext *ctx, arg_r *a, |
| void (*accumulate_func)(TCGv, TCGv, TCGv), |
| void (*extend_operand_func)(TCGv, TCGv)) |
| { |
| TCGv dest = dest_gpr(ctx, a->rd); |
| TCGv src0 = get_gpr(ctx, a->rd, EXT_NONE); |
| TCGv src1 = get_gpr(ctx, a->rs1, EXT_NONE); |
| TCGv src2 = get_gpr(ctx, a->rs2, EXT_NONE); |
| TCGv tmp = tcg_temp_new(); |
| |
| if (extend_operand_func) { |
| TCGv tmp2 = tcg_temp_new(); |
| extend_operand_func(tmp, src1); |
| extend_operand_func(tmp2, src2); |
| tcg_gen_mul_tl(tmp, tmp, tmp2); |
| } else { |
| tcg_gen_mul_tl(tmp, src1, src2); |
| } |
| |
| accumulate_func(dest, src0, tmp); |
| gen_set_gpr(ctx, a->rd, dest); |
| return true; |
| } |
| |
| /* th.mula: "rd = rd + rs1 * rs2" */ |
| static bool trans_th_mula(DisasContext *ctx, arg_th_mula *a) |
| { |
| REQUIRE_XTHEADMAC(ctx); |
| return gen_th_mac(ctx, a, tcg_gen_add_tl, NULL); |
| } |
| |
| /* th.mulah: "rd = sext.w(rd + sext.w(rs1[15:0]) * sext.w(rs2[15:0]))" */ |
| static bool trans_th_mulah(DisasContext *ctx, arg_th_mulah *a) |
| { |
| REQUIRE_XTHEADMAC(ctx); |
| ctx->ol = MXL_RV32; |
| return gen_th_mac(ctx, a, tcg_gen_add_tl, tcg_gen_ext16s_tl); |
| } |
| |
| /* th.mulaw: "rd = sext.w(rd + rs1 * rs2)" */ |
| static bool trans_th_mulaw(DisasContext *ctx, arg_th_mulaw *a) |
| { |
| REQUIRE_XTHEADMAC(ctx); |
| REQUIRE_64BIT(ctx); |
| ctx->ol = MXL_RV32; |
| return gen_th_mac(ctx, a, tcg_gen_add_tl, NULL); |
| } |
| |
| /* th.muls: "rd = rd - rs1 * rs2" */ |
| static bool trans_th_muls(DisasContext *ctx, arg_th_muls *a) |
| { |
| REQUIRE_XTHEADMAC(ctx); |
| return gen_th_mac(ctx, a, tcg_gen_sub_tl, NULL); |
| } |
| |
| /* th.mulsh: "rd = sext.w(rd - sext.w(rs1[15:0]) * sext.w(rs2[15:0]))" */ |
| static bool trans_th_mulsh(DisasContext *ctx, arg_th_mulsh *a) |
| { |
| REQUIRE_XTHEADMAC(ctx); |
| ctx->ol = MXL_RV32; |
| return gen_th_mac(ctx, a, tcg_gen_sub_tl, tcg_gen_ext16s_tl); |
| } |
| |
| /* th.mulsw: "rd = sext.w(rd - rs1 * rs2)" */ |
| static bool trans_th_mulsw(DisasContext *ctx, arg_th_mulsw *a) |
| { |
| REQUIRE_XTHEADMAC(ctx); |
| REQUIRE_64BIT(ctx); |
| ctx->ol = MXL_RV32; |
| return gen_th_mac(ctx, a, tcg_gen_sub_tl, NULL); |
| } |
| |
| /* XTheadMemIdx */ |
| |
| /* |
| * Load with memop from indexed address and add (imm5 << imm2) to rs1. |
| * If !preinc, then the load address is rs1. |
| * If preinc, then the load address is rs1 + (imm5) << imm2). |
| */ |
| static bool gen_load_inc(DisasContext *ctx, arg_th_meminc *a, MemOp memop, |
| bool preinc) |
| { |
| if (a->rs1 == a->rd) { |
| return false; |
| } |
| |
| int imm = a->imm5 << a->imm2; |
| TCGv addr = get_address(ctx, a->rs1, preinc ? imm : 0); |
| TCGv rd = dest_gpr(ctx, a->rd); |
| TCGv rs1 = get_gpr(ctx, a->rs1, EXT_NONE); |
| |
| tcg_gen_qemu_ld_tl(rd, addr, ctx->mem_idx, memop); |
| tcg_gen_addi_tl(rs1, rs1, imm); |
| gen_set_gpr(ctx, a->rd, rd); |
| gen_set_gpr(ctx, a->rs1, rs1); |
| return true; |
| } |
| |
| /* |
| * Store with memop to indexed address and add (imm5 << imm2) to rs1. |
| * If !preinc, then the store address is rs1. |
| * If preinc, then the store address is rs1 + (imm5) << imm2). |
| */ |
| static bool gen_store_inc(DisasContext *ctx, arg_th_meminc *a, MemOp memop, |
| bool preinc) |
| { |
| int imm = a->imm5 << a->imm2; |
| TCGv addr = get_address(ctx, a->rs1, preinc ? imm : 0); |
| TCGv data = get_gpr(ctx, a->rd, EXT_NONE); |
| TCGv rs1 = get_gpr(ctx, a->rs1, EXT_NONE); |
| |
| tcg_gen_qemu_st_tl(data, addr, ctx->mem_idx, memop); |
| tcg_gen_addi_tl(rs1, rs1, imm); |
| gen_set_gpr(ctx, a->rs1, rs1); |
| return true; |
| } |
| |
| static bool trans_th_ldia(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| REQUIRE_64BIT(ctx); |
| return gen_load_inc(ctx, a, MO_TESQ, false); |
| } |
| |
| static bool trans_th_ldib(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| REQUIRE_64BIT(ctx); |
| return gen_load_inc(ctx, a, MO_TESQ, true); |
| } |
| |
| static bool trans_th_lwia(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_inc(ctx, a, MO_TESL, false); |
| } |
| |
| static bool trans_th_lwib(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_inc(ctx, a, MO_TESL, true); |
| } |
| |
| static bool trans_th_lwuia(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| REQUIRE_64BIT(ctx); |
| return gen_load_inc(ctx, a, MO_TEUL, false); |
| } |
| |
| static bool trans_th_lwuib(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| REQUIRE_64BIT(ctx); |
| return gen_load_inc(ctx, a, MO_TEUL, true); |
| } |
| |
| static bool trans_th_lhia(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_inc(ctx, a, MO_TESW, false); |
| } |
| |
| static bool trans_th_lhib(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_inc(ctx, a, MO_TESW, true); |
| } |
| |
| static bool trans_th_lhuia(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_inc(ctx, a, MO_TEUW, false); |
| } |
| |
| static bool trans_th_lhuib(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_inc(ctx, a, MO_TEUW, true); |
| } |
| |
| static bool trans_th_lbia(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_inc(ctx, a, MO_SB, false); |
| } |
| |
| static bool trans_th_lbib(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_inc(ctx, a, MO_SB, true); |
| } |
| |
| static bool trans_th_lbuia(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_inc(ctx, a, MO_UB, false); |
| } |
| |
| static bool trans_th_lbuib(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_inc(ctx, a, MO_UB, true); |
| } |
| |
| static bool trans_th_sdia(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| REQUIRE_64BIT(ctx); |
| return gen_store_inc(ctx, a, MO_TESQ, false); |
| } |
| |
| static bool trans_th_sdib(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| REQUIRE_64BIT(ctx); |
| return gen_store_inc(ctx, a, MO_TESQ, true); |
| } |
| |
| static bool trans_th_swia(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_store_inc(ctx, a, MO_TESL, false); |
| } |
| |
| static bool trans_th_swib(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_store_inc(ctx, a, MO_TESL, true); |
| } |
| |
| static bool trans_th_shia(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_store_inc(ctx, a, MO_TESW, false); |
| } |
| |
| static bool trans_th_shib(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_store_inc(ctx, a, MO_TESW, true); |
| } |
| |
| static bool trans_th_sbia(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_store_inc(ctx, a, MO_SB, false); |
| } |
| |
| static bool trans_th_sbib(DisasContext *ctx, arg_th_meminc *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_store_inc(ctx, a, MO_SB, true); |
| } |
| |
| /* |
| * Load with memop from indexed address. |
| * If !zext_offs, then address is rs1 + (rs2 << imm2). |
| * If zext_offs, then address is rs1 + (zext(rs2[31:0]) << imm2). |
| */ |
| static bool gen_load_idx(DisasContext *ctx, arg_th_memidx *a, MemOp memop, |
| bool zext_offs) |
| { |
| TCGv rd = dest_gpr(ctx, a->rd); |
| TCGv addr = get_th_address_indexed(ctx, a->rs1, a->rs2, a->imm2, zext_offs); |
| |
| tcg_gen_qemu_ld_tl(rd, addr, ctx->mem_idx, memop); |
| gen_set_gpr(ctx, a->rd, rd); |
| |
| return true; |
| } |
| |
| /* |
| * Store with memop to indexed address. |
| * If !zext_offs, then address is rs1 + (rs2 << imm2). |
| * If zext_offs, then address is rs1 + (zext(rs2[31:0]) << imm2). |
| */ |
| static bool gen_store_idx(DisasContext *ctx, arg_th_memidx *a, MemOp memop, |
| bool zext_offs) |
| { |
| TCGv data = get_gpr(ctx, a->rd, EXT_NONE); |
| TCGv addr = get_th_address_indexed(ctx, a->rs1, a->rs2, a->imm2, zext_offs); |
| |
| tcg_gen_qemu_st_tl(data, addr, ctx->mem_idx, memop); |
| |
| return true; |
| } |
| |
| static bool trans_th_lrd(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| REQUIRE_64BIT(ctx); |
| return gen_load_idx(ctx, a, MO_TESQ, false); |
| } |
| |
| static bool trans_th_lrw(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_idx(ctx, a, MO_TESL, false); |
| } |
| |
| static bool trans_th_lrwu(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| REQUIRE_64BIT(ctx); |
| return gen_load_idx(ctx, a, MO_TEUL, false); |
| } |
| |
| static bool trans_th_lrh(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_idx(ctx, a, MO_TESW, false); |
| } |
| |
| static bool trans_th_lrhu(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_idx(ctx, a, MO_TEUW, false); |
| } |
| |
| static bool trans_th_lrb(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_idx(ctx, a, MO_SB, false); |
| } |
| |
| static bool trans_th_lrbu(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_idx(ctx, a, MO_UB, false); |
| } |
| |
| static bool trans_th_srd(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| REQUIRE_64BIT(ctx); |
| return gen_store_idx(ctx, a, MO_TESQ, false); |
| } |
| |
| static bool trans_th_srw(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_store_idx(ctx, a, MO_TESL, false); |
| } |
| |
| static bool trans_th_srh(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_store_idx(ctx, a, MO_TESW, false); |
| } |
| |
| static bool trans_th_srb(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_store_idx(ctx, a, MO_SB, false); |
| } |
| static bool trans_th_lurd(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| REQUIRE_64BIT(ctx); |
| return gen_load_idx(ctx, a, MO_TESQ, true); |
| } |
| |
| static bool trans_th_lurw(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_idx(ctx, a, MO_TESL, true); |
| } |
| |
| static bool trans_th_lurwu(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| REQUIRE_64BIT(ctx); |
| return gen_load_idx(ctx, a, MO_TEUL, true); |
| } |
| |
| static bool trans_th_lurh(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_idx(ctx, a, MO_TESW, true); |
| } |
| |
| static bool trans_th_lurhu(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_idx(ctx, a, MO_TEUW, true); |
| } |
| |
| static bool trans_th_lurb(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_idx(ctx, a, MO_SB, true); |
| } |
| |
| static bool trans_th_lurbu(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_load_idx(ctx, a, MO_UB, true); |
| } |
| |
| static bool trans_th_surd(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| REQUIRE_64BIT(ctx); |
| return gen_store_idx(ctx, a, MO_TESQ, true); |
| } |
| |
| static bool trans_th_surw(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_store_idx(ctx, a, MO_TESL, true); |
| } |
| |
| static bool trans_th_surh(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_store_idx(ctx, a, MO_TESW, true); |
| } |
| |
| static bool trans_th_surb(DisasContext *ctx, arg_th_memidx *a) |
| { |
| REQUIRE_XTHEADMEMIDX(ctx); |
| return gen_store_idx(ctx, a, MO_SB, true); |
| } |
| |
| /* XTheadMemPair */ |
| |
| static bool gen_loadpair_tl(DisasContext *ctx, arg_th_pair *a, MemOp memop, |
| int shamt) |
| { |
| if (a->rs == a->rd1 || a->rs == a->rd2 || a->rd1 == a->rd2) { |
| return false; |
| } |
| |
| TCGv t1 = tcg_temp_new(); |
| TCGv t2 = tcg_temp_new(); |
| TCGv addr1 = tcg_temp_new(); |
| TCGv addr2 = tcg_temp_new(); |
| int imm = a->sh2 << shamt; |
| |
| addr1 = get_address(ctx, a->rs, imm); |
| addr2 = get_address(ctx, a->rs, memop_size(memop) + imm); |
| |
| tcg_gen_qemu_ld_tl(t1, addr1, ctx->mem_idx, memop); |
| tcg_gen_qemu_ld_tl(t2, addr2, ctx->mem_idx, memop); |
| gen_set_gpr(ctx, a->rd1, t1); |
| gen_set_gpr(ctx, a->rd2, t2); |
| return true; |
| } |
| |
| static bool trans_th_ldd(DisasContext *ctx, arg_th_pair *a) |
| { |
| REQUIRE_XTHEADMEMPAIR(ctx); |
| REQUIRE_64BIT(ctx); |
| return gen_loadpair_tl(ctx, a, MO_TESQ, 4); |
| } |
| |
| static bool trans_th_lwd(DisasContext *ctx, arg_th_pair *a) |
| { |
| REQUIRE_XTHEADMEMPAIR(ctx); |
| return gen_loadpair_tl(ctx, a, MO_TESL, 3); |
| } |
| |
| static bool trans_th_lwud(DisasContext *ctx, arg_th_pair *a) |
| { |
| REQUIRE_XTHEADMEMPAIR(ctx); |
| return gen_loadpair_tl(ctx, a, MO_TEUL, 3); |
| } |
| |
| static bool gen_storepair_tl(DisasContext *ctx, arg_th_pair *a, MemOp memop, |
| int shamt) |
| { |
| TCGv data1 = get_gpr(ctx, a->rd1, EXT_NONE); |
| TCGv data2 = get_gpr(ctx, a->rd2, EXT_NONE); |
| TCGv addr1 = tcg_temp_new(); |
| TCGv addr2 = tcg_temp_new(); |
| int imm = a->sh2 << shamt; |
| |
| addr1 = get_address(ctx, a->rs, imm); |
| addr2 = get_address(ctx, a->rs, memop_size(memop) + imm); |
| |
| tcg_gen_qemu_st_tl(data1, addr1, ctx->mem_idx, memop); |
| tcg_gen_qemu_st_tl(data2, addr2, ctx->mem_idx, memop); |
| return true; |
| } |
| |
| static bool trans_th_sdd(DisasContext *ctx, arg_th_pair *a) |
| { |
| REQUIRE_XTHEADMEMPAIR(ctx); |
| REQUIRE_64BIT(ctx); |
| return gen_storepair_tl(ctx, a, MO_TESQ, 4); |
| } |
| |
| static bool trans_th_swd(DisasContext *ctx, arg_th_pair *a) |
| { |
| REQUIRE_XTHEADMEMPAIR(ctx); |
| return gen_storepair_tl(ctx, a, MO_TESL, 3); |
| } |
| |
| /* XTheadSync */ |
| |
| static bool trans_th_sfence_vmas(DisasContext *ctx, arg_th_sfence_vmas *a) |
| { |
| (void) a; |
| REQUIRE_XTHEADSYNC(ctx); |
| |
| #ifndef CONFIG_USER_ONLY |
| REQUIRE_PRIV_MS(ctx); |
| gen_helper_tlb_flush_all(cpu_env); |
| return true; |
| #else |
| return false; |
| #endif |
| } |
| |
| #ifndef CONFIG_USER_ONLY |
| static void gen_th_sync_local(DisasContext *ctx) |
| { |
| /* |
| * Emulate out-of-order barriers with pipeline flush |
| * by exiting the translation block. |
| */ |
| gen_set_pc_imm(ctx, ctx->pc_succ_insn); |
| tcg_gen_exit_tb(NULL, 0); |
| ctx->base.is_jmp = DISAS_NORETURN; |
| } |
| #endif |
| |
| static bool trans_th_sync(DisasContext *ctx, arg_th_sync *a) |
| { |
| (void) a; |
| REQUIRE_XTHEADSYNC(ctx); |
| |
| #ifndef CONFIG_USER_ONLY |
| REQUIRE_PRIV_MSU(ctx); |
| |
| /* |
| * th.sync is an out-of-order barrier. |
| */ |
| gen_th_sync_local(ctx); |
| |
| return true; |
| #else |
| return false; |
| #endif |
| } |
| |
| static bool trans_th_sync_i(DisasContext *ctx, arg_th_sync_i *a) |
| { |
| (void) a; |
| REQUIRE_XTHEADSYNC(ctx); |
| |
| #ifndef CONFIG_USER_ONLY |
| REQUIRE_PRIV_MSU(ctx); |
| |
| /* |
| * th.sync.i is th.sync plus pipeline flush. |
| */ |
| gen_th_sync_local(ctx); |
| |
| return true; |
| #else |
| return false; |
| #endif |
| } |
| |
| static bool trans_th_sync_is(DisasContext *ctx, arg_th_sync_is *a) |
| { |
| /* This instruction has the same behaviour like th.sync.i. */ |
| return trans_th_sync_i(ctx, a); |
| } |
| |
| static bool trans_th_sync_s(DisasContext *ctx, arg_th_sync_s *a) |
| { |
| /* This instruction has the same behaviour like th.sync. */ |
| return trans_th_sync(ctx, a); |
| } |