| /* |
| * RISC-V translation routines for the RV64D Standard Extension. |
| * |
| * Copyright (c) 2016-2017 Sagar Karandikar, sagark@eecs.berkeley.edu |
| * Copyright (c) 2018 Peer Adelt, peer.adelt@hni.uni-paderborn.de |
| * Bastian Koppelmann, kbastian@mail.uni-paderborn.de |
| * |
| * This program is free software; you can redistribute it and/or modify it |
| * under the terms and conditions of the GNU General Public License, |
| * version 2 or later, as published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope it will be useful, but WITHOUT |
| * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for |
| * more details. |
| * |
| * You should have received a copy of the GNU General Public License along with |
| * this program. If not, see <http://www.gnu.org/licenses/>. |
| */ |
| |
| #define REQUIRE_ZDINX_OR_D(ctx) do { \ |
| if (!ctx->cfg_ptr->ext_zdinx) { \ |
| REQUIRE_EXT(ctx, RVD); \ |
| } \ |
| } while (0) |
| |
| #define REQUIRE_EVEN(ctx, reg) do { \ |
| if (ctx->cfg_ptr->ext_zdinx && (get_xl(ctx) == MXL_RV32) && \ |
| ((reg) & 0x1)) { \ |
| return false; \ |
| } \ |
| } while (0) |
| |
| #define REQUIRE_ZCD_OR_DC(ctx) do { \ |
| if (!ctx->cfg_ptr->ext_zcd) { \ |
| if (!has_ext(ctx, RVD) || !has_ext(ctx, RVC)) { \ |
| return false; \ |
| } \ |
| } \ |
| } while (0) |
| |
| static bool trans_fld(DisasContext *ctx, arg_fld *a) |
| { |
| TCGv addr; |
| MemOp memop = MO_TEUQ; |
| |
| REQUIRE_FPU; |
| REQUIRE_EXT(ctx, RVD); |
| |
| /* |
| * FLD and FSD are only guaranteed to execute atomically if the effective |
| * address is naturally aligned and XLEN≥64. Also, zama16b applies to |
| * loads and stores of no more than MXLEN bits defined in the F, D, and |
| * Q extensions. |
| */ |
| if (get_xl_max(ctx) == MXL_RV32) { |
| memop |= MO_ATOM_NONE; |
| } else if (ctx->cfg_ptr->ext_zama16b) { |
| memop |= MO_ATOM_WITHIN16; |
| } else { |
| memop |= MO_ATOM_IFALIGN; |
| } |
| |
| decode_save_opc(ctx, 0); |
| addr = get_address(ctx, a->rs1, a->imm); |
| tcg_gen_qemu_ld_i64(cpu_fpr[a->rd], addr, ctx->mem_idx, memop); |
| |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fsd(DisasContext *ctx, arg_fsd *a) |
| { |
| TCGv addr; |
| MemOp memop = MO_TEUQ; |
| |
| REQUIRE_FPU; |
| REQUIRE_EXT(ctx, RVD); |
| |
| if (get_xl_max(ctx) == MXL_RV32) { |
| memop |= MO_ATOM_NONE; |
| } else if (ctx->cfg_ptr->ext_zama16b) { |
| memop |= MO_ATOM_WITHIN16; |
| } else { |
| memop |= MO_ATOM_IFALIGN; |
| } |
| |
| decode_save_opc(ctx, 0); |
| addr = get_address(ctx, a->rs1, a->imm); |
| tcg_gen_qemu_st_i64(cpu_fpr[a->rs2], addr, ctx->mem_idx, memop); |
| return true; |
| } |
| |
| static bool trans_c_fld(DisasContext *ctx, arg_fld *a) |
| { |
| REQUIRE_ZCD_OR_DC(ctx); |
| return trans_fld(ctx, a); |
| } |
| |
| static bool trans_c_fsd(DisasContext *ctx, arg_fsd *a) |
| { |
| REQUIRE_ZCD_OR_DC(ctx); |
| return trans_fsd(ctx, a); |
| } |
| |
| static bool trans_fmadd_d(DisasContext *ctx, arg_fmadd_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd | a->rs1 | a->rs2 | a->rs3); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| TCGv_i64 src3 = get_fpr_d(ctx, a->rs3); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fmadd_d(dest, tcg_env, src1, src2, src3); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fmsub_d(DisasContext *ctx, arg_fmsub_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd | a->rs1 | a->rs2 | a->rs3); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| TCGv_i64 src3 = get_fpr_d(ctx, a->rs3); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fmsub_d(dest, tcg_env, src1, src2, src3); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fnmsub_d(DisasContext *ctx, arg_fnmsub_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd | a->rs1 | a->rs2 | a->rs3); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| TCGv_i64 src3 = get_fpr_d(ctx, a->rs3); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fnmsub_d(dest, tcg_env, src1, src2, src3); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fnmadd_d(DisasContext *ctx, arg_fnmadd_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd | a->rs1 | a->rs2 | a->rs3); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| TCGv_i64 src3 = get_fpr_d(ctx, a->rs3); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fnmadd_d(dest, tcg_env, src1, src2, src3); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fadd_d(DisasContext *ctx, arg_fadd_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd | a->rs1 | a->rs2); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fadd_d(dest, tcg_env, src1, src2); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fsub_d(DisasContext *ctx, arg_fsub_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd | a->rs1 | a->rs2); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fsub_d(dest, tcg_env, src1, src2); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fmul_d(DisasContext *ctx, arg_fmul_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd | a->rs1 | a->rs2); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fmul_d(dest, tcg_env, src1, src2); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fdiv_d(DisasContext *ctx, arg_fdiv_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd | a->rs1 | a->rs2); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fdiv_d(dest, tcg_env, src1, src2); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fsqrt_d(DisasContext *ctx, arg_fsqrt_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd | a->rs1); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fsqrt_d(dest, tcg_env, src1); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fsgnj_d(DisasContext *ctx, arg_fsgnj_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd | a->rs1 | a->rs2); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| if (a->rs1 == a->rs2) { /* FMOV */ |
| dest = get_fpr_d(ctx, a->rs1); |
| } else { |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| tcg_gen_deposit_i64(dest, src2, src1, 0, 63); |
| } |
| gen_set_fpr_d(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fsgnjn_d(DisasContext *ctx, arg_fsgnjn_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd | a->rs1 | a->rs2); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| |
| if (a->rs1 == a->rs2) { /* FNEG */ |
| tcg_gen_xori_i64(dest, src1, INT64_MIN); |
| } else { |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| TCGv_i64 t0 = tcg_temp_new_i64(); |
| tcg_gen_not_i64(t0, src2); |
| tcg_gen_deposit_i64(dest, t0, src1, 0, 63); |
| } |
| gen_set_fpr_d(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fsgnjx_d(DisasContext *ctx, arg_fsgnjx_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd | a->rs1 | a->rs2); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| |
| if (a->rs1 == a->rs2) { /* FABS */ |
| tcg_gen_andi_i64(dest, src1, ~INT64_MIN); |
| } else { |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| TCGv_i64 t0 = tcg_temp_new_i64(); |
| tcg_gen_andi_i64(t0, src2, INT64_MIN); |
| tcg_gen_xor_i64(dest, src1, t0); |
| } |
| gen_set_fpr_d(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fmin_d(DisasContext *ctx, arg_fmin_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd | a->rs1 | a->rs2); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| |
| gen_helper_fmin_d(dest, tcg_env, src1, src2); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fmax_d(DisasContext *ctx, arg_fmax_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd | a->rs1 | a->rs2); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| |
| gen_helper_fmax_d(dest, tcg_env, src1, src2); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fcvt_s_d(DisasContext *ctx, arg_fcvt_s_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rs1); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fcvt_s_d(dest, tcg_env, src1); |
| gen_set_fpr_hs(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fcvt_d_s(DisasContext *ctx, arg_fcvt_d_s *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_hs(ctx, a->rs1); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fcvt_d_s(dest, tcg_env, src1); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_feq_d(DisasContext *ctx, arg_feq_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rs1 | a->rs2); |
| |
| TCGv dest = dest_gpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| |
| gen_helper_feq_d(dest, tcg_env, src1, src2); |
| gen_set_gpr(ctx, a->rd, dest); |
| return true; |
| } |
| |
| static bool trans_flt_d(DisasContext *ctx, arg_flt_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rs1 | a->rs2); |
| |
| TCGv dest = dest_gpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| |
| gen_helper_flt_d(dest, tcg_env, src1, src2); |
| gen_set_gpr(ctx, a->rd, dest); |
| return true; |
| } |
| |
| static bool trans_fle_d(DisasContext *ctx, arg_fle_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rs1 | a->rs2); |
| |
| TCGv dest = dest_gpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| TCGv_i64 src2 = get_fpr_d(ctx, a->rs2); |
| |
| gen_helper_fle_d(dest, tcg_env, src1, src2); |
| gen_set_gpr(ctx, a->rd, dest); |
| return true; |
| } |
| |
| static bool trans_fclass_d(DisasContext *ctx, arg_fclass_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rs1); |
| |
| TCGv dest = dest_gpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| |
| gen_helper_fclass_d(dest, src1); |
| gen_set_gpr(ctx, a->rd, dest); |
| return true; |
| } |
| |
| static bool trans_fcvt_w_d(DisasContext *ctx, arg_fcvt_w_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rs1); |
| |
| TCGv dest = dest_gpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fcvt_w_d(dest, tcg_env, src1); |
| gen_set_gpr(ctx, a->rd, dest); |
| return true; |
| } |
| |
| static bool trans_fcvt_wu_d(DisasContext *ctx, arg_fcvt_wu_d *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rs1); |
| |
| TCGv dest = dest_gpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fcvt_wu_d(dest, tcg_env, src1); |
| gen_set_gpr(ctx, a->rd, dest); |
| return true; |
| } |
| |
| static bool trans_fcvt_d_w(DisasContext *ctx, arg_fcvt_d_w *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv src = get_gpr(ctx, a->rs1, EXT_SIGN); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fcvt_d_w(dest, tcg_env, src); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fcvt_d_wu(DisasContext *ctx, arg_fcvt_d_wu *a) |
| { |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv src = get_gpr(ctx, a->rs1, EXT_ZERO); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fcvt_d_wu(dest, tcg_env, src); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fcvt_l_d(DisasContext *ctx, arg_fcvt_l_d *a) |
| { |
| REQUIRE_64BIT(ctx); |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rs1); |
| |
| TCGv dest = dest_gpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fcvt_l_d(dest, tcg_env, src1); |
| gen_set_gpr(ctx, a->rd, dest); |
| return true; |
| } |
| |
| static bool trans_fcvt_lu_d(DisasContext *ctx, arg_fcvt_lu_d *a) |
| { |
| REQUIRE_64BIT(ctx); |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rs1); |
| |
| TCGv dest = dest_gpr(ctx, a->rd); |
| TCGv_i64 src1 = get_fpr_d(ctx, a->rs1); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fcvt_lu_d(dest, tcg_env, src1); |
| gen_set_gpr(ctx, a->rd, dest); |
| return true; |
| } |
| |
| static bool trans_fmv_x_d(DisasContext *ctx, arg_fmv_x_d *a) |
| { |
| REQUIRE_64BIT(ctx); |
| REQUIRE_FPU; |
| REQUIRE_EXT(ctx, RVD); |
| |
| #ifdef TARGET_RISCV64 |
| gen_set_gpr(ctx, a->rd, cpu_fpr[a->rs1]); |
| return true; |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |
| |
| static bool trans_fcvt_d_l(DisasContext *ctx, arg_fcvt_d_l *a) |
| { |
| REQUIRE_64BIT(ctx); |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv src = get_gpr(ctx, a->rs1, EXT_SIGN); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fcvt_d_l(dest, tcg_env, src); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fcvt_d_lu(DisasContext *ctx, arg_fcvt_d_lu *a) |
| { |
| REQUIRE_64BIT(ctx); |
| REQUIRE_FPU; |
| REQUIRE_ZDINX_OR_D(ctx); |
| REQUIRE_EVEN(ctx, a->rd); |
| |
| TCGv_i64 dest = dest_fpr(ctx, a->rd); |
| TCGv src = get_gpr(ctx, a->rs1, EXT_ZERO); |
| |
| gen_set_rm(ctx, a->rm); |
| gen_helper_fcvt_d_lu(dest, tcg_env, src); |
| gen_set_fpr_d(ctx, a->rd, dest); |
| |
| mark_fs_dirty(ctx); |
| return true; |
| } |
| |
| static bool trans_fmv_d_x(DisasContext *ctx, arg_fmv_d_x *a) |
| { |
| REQUIRE_64BIT(ctx); |
| REQUIRE_FPU; |
| REQUIRE_EXT(ctx, RVD); |
| |
| #ifdef TARGET_RISCV64 |
| tcg_gen_mov_tl(cpu_fpr[a->rd], get_gpr(ctx, a->rs1, EXT_NONE)); |
| mark_fs_dirty(ctx); |
| return true; |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |